[
{
    "branch": "main",
    "index": "0",
    "phase": "pre",
    "playbook": "github.com/osism/zuul-config/playbooks/base/pre.yaml",
    "plays": [
        {
            "play": {
                "duration": {
                    "end": "2026-02-24T00:01:09.090551Z",
                    "start": "2026-02-24T00:00:14.594682Z"
                },
                "id": "0242ac1f-6606-c045-40cc-000000000002",
                "name": "Base pre"
            },
            "tasks": [
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "include_role",
                            "changed": false,
                            "include_args": {
                                "name": "set-zuul-log-path-fact"
                            }
                        }
                    },
                    "role": {
                        "id": "0242ac1f-6606-c045-40cc-00000000000a",
                        "name": "emit-job-header",
                        "path": "/var/lib/zuul/builds/7573dac089bc49b5b355588f880c7711/trusted/project_1/github.com/osism/openinfra-zuul-jobs/roles/emit-job-header"
                    },
                    "task": {
                        "duration": {
                            "end": "2026-02-24T00:00:14.707511Z",
                            "start": "2026-02-24T00:00:14.644625Z"
                        },
                        "id": "0242ac1f-6606-c045-40cc-00000000000c",
                        "name": "Setup log path fact"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "set_fact",
                            "ansible_facts": {
                                "zuul_log_path": "757/osism/7573dac089bc49b5b355588f880c7711"
                            },
                            "changed": false
                        }
                    },
                    "role": {
                        "id": "0242ac1f-6606-c045-40cc-000000000087",
                        "name": "set-zuul-log-path-fact",
                        "path": "/var/lib/zuul/builds/7573dac089bc49b5b355588f880c7711/trusted/project_1/github.com/osism/openinfra-zuul-jobs/roles/set-zuul-log-path-fact"
                    },
                    "task": {
                        "duration": {
                            "end": "2026-02-24T00:00:14.837539Z",
                            "start": "2026-02-24T00:00:14.757258Z"
                        },
                        "id": "0242ac1f-6606-c045-40cc-000000000089",
                        "name": "Set log path for a build"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "debug",
                            "changed": false,
                            "msg": "# Job Information\nAnsible Version: 2.16.14\nJob: testbed-deploy-current-in-a-nutshell-with-tempest-ubuntu-24.04\nPipeline: periodic-midnight\nExecutor: 521e9411259a\nTriggered by: https://github.com/osism/testbed\nEvent ID: 3832fb26ed904f699369d8dbb7c6eb38\n"
                        }
                    },
                    "role": {
                        "id": "0242ac1f-6606-c045-40cc-00000000000a",
                        "name": "emit-job-header",
                        "path": "/var/lib/zuul/builds/7573dac089bc49b5b355588f880c7711/trusted/project_1/github.com/osism/openinfra-zuul-jobs/roles/emit-job-header"
                    },
                    "task": {
                        "duration": {
                            "end": "2026-02-24T00:00:14.891155Z",
                            "start": "2026-02-24T00:00:14.850129Z"
                        },
                        "id": "0242ac1f-6606-c045-40cc-00000000000d",
                        "name": "Print job information"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "debug",
                            "changed": false,
                            "msg": "All items completed",
                            "results": [
                                {
                                    "ansible_loop_var": "zj_item",
                                    "changed": false,
                                    "failed": false,
                                    "msg": "# Node Information\nInventory Hostname: orchestrator\nHostname: zuul-static-regiocloud-infra-1\nUsername: zuul-testbed02\nDistro: Debian 12.13\nProvider: static-testbed\nRegion: \nLabel: testbed-orchestrator\nProduct Name: OpenStack Nova\nInterface IP: 81.163.193.140\n",
                                    "zj_item": "orchestrator"
                                }
                            ]
                        }
                    },
                    "role": {
                        "id": "0242ac1f-6606-c045-40cc-00000000000a",
                        "name": "emit-job-header",
                        "path": "/var/lib/zuul/builds/7573dac089bc49b5b355588f880c7711/trusted/project_1/github.com/osism/openinfra-zuul-jobs/roles/emit-job-header"
                    },
                    "task": {
                        "duration": {
                            "end": "2026-02-24T00:00:15.229223Z",
                            "start": "2026-02-24T00:00:14.898383Z"
                        },
                        "id": "0242ac1f-6606-c045-40cc-00000000000e",
                        "name": "Print node information"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "file",
                            "changed": true,
                            "diff": {
                                "after": {
                                    "path": "/var/lib/zuul/builds/7573dac089bc49b5b355588f880c7711/work/logs/zuul-info",
                                    "state": "directory"
                                },
                                "before": {
                                    "path": "/var/lib/zuul/builds/7573dac089bc49b5b355588f880c7711/work/logs/zuul-info",
                                    "state": "absent"
                                }
                            },
                            "gid": 0,
                            "group": "root",
                            "invocation": {
                                "module_args": {
                                    "_diff_peek": null,
                                    "_original_basename": null,
                                    "access_time": null,
                                    "access_time_format": "%Y%m%d%H%M.%S",
                                    "attributes": null,
                                    "follow": true,
                                    "force": false,
                                    "group": null,
                                    "mode": 493,
                                    "modification_time": null,
                                    "modification_time_format": "%Y%m%d%H%M.%S",
                                    "owner": null,
                                    "path": "/var/lib/zuul/builds/7573dac089bc49b5b355588f880c7711/work/logs/zuul-info",
                                    "recurse": false,
                                    "selevel": null,
                                    "serole": null,
                                    "setype": null,
                                    "seuser": null,
                                    "src": null,
                                    "state": "directory",
                                    "unsafe_writes": false
                                }
                            },
                            "mode": "0755",
                            "owner": "root",
                            "path": "/var/lib/zuul/builds/7573dac089bc49b5b355588f880c7711/work/logs/zuul-info",
                            "size": 4096,
                            "state": "directory",
                            "uid": 0
                        }
                    },
                    "role": {
                        "id": "0242ac1f-6606-c045-40cc-000000000011",
                        "name": "log-inventory",
                        "path": "/var/lib/zuul/builds/7573dac089bc49b5b355588f880c7711/trusted/project_1/github.com/osism/openinfra-zuul-jobs/roles/log-inventory"
                    },
                    "task": {
                        "duration": {
                            "end": "2026-02-24T00:00:16.454405Z",
                            "start": "2026-02-24T00:00:15.240706Z"
                        },
                        "id": "0242ac1f-6606-c045-40cc-000000000013",
                        "name": "Ensure Zuul Ansible directory exists"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "copy",
                            "changed": true,
                            "checksum": "f8e1c69b3bbd56635031dc4c802cb95a1e5518dc",
                            "dest": "/var/lib/zuul/builds/7573dac089bc49b5b355588f880c7711/work/logs/zuul-info/inventory.yaml",
                            "diff": [],
                            "gid": 0,
                            "group": "root",
                            "invocation": {
                                "module_args": {
                                    "_original_basename": "inventory.yaml",
                                    "attributes": null,
                                    "backup": false,
                                    "checksum": "f8e1c69b3bbd56635031dc4c802cb95a1e5518dc",
                                    "content": null,
                                    "dest": "/var/lib/zuul/builds/7573dac089bc49b5b355588f880c7711/work/logs/zuul-info",
                                    "directory_mode": null,
                                    "follow": false,
                                    "force": true,
                                    "group": null,
                                    "local_follow": null,
                                    "mode": 420,
                                    "owner": null,
                                    "remote_src": null,
                                    "selevel": null,
                                    "serole": null,
                                    "setype": null,
                                    "seuser": null,
                                    "src": "/var/lib/zuul/builds/7573dac089bc49b5b355588f880c7711/work/.ansible/tmp/ansible-tmp-1771891216.6136699-34-160333281001898/source",
                                    "unsafe_writes": false,
                                    "validate": null
                                }
                            },
                            "md5sum": "510082a860dcb0c8dfed8245d559148a",
                            "mode": "0644",
                            "owner": "root",
                            "size": 8206,
                            "src": "/var/lib/zuul/builds/7573dac089bc49b5b355588f880c7711/work/.ansible/tmp/ansible-tmp-1771891216.6136699-34-160333281001898/source",
                            "state": "file",
                            "uid": 0
                        }
                    },
                    "role": {
                        "id": "0242ac1f-6606-c045-40cc-000000000011",
                        "name": "log-inventory",
                        "path": "/var/lib/zuul/builds/7573dac089bc49b5b355588f880c7711/trusted/project_1/github.com/osism/openinfra-zuul-jobs/roles/log-inventory"
                    },
                    "task": {
                        "duration": {
                            "end": "2026-02-24T00:00:18.750912Z",
                            "start": "2026-02-24T00:00:16.467993Z"
                        },
                        "id": "0242ac1f-6606-c045-40cc-000000000014",
                        "name": "Copy ansible inventory to logs dir"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "stat",
                            "changed": false,
                            "failed_when_result": false,
                            "invocation": {
                                "module_args": {
                                    "checksum_algorithm": "sha1",
                                    "follow": false,
                                    "get_attributes": true,
                                    "get_checksum": true,
                                    "get_mime": true,
                                    "path": "/var/lib/zuul/builds/7573dac089bc49b5b355588f880c7711/work/7573dac089bc49b5b355588f880c7711_id_rsa"
                                }
                            },
                            "stat": {
                                "exists": false
                            }
                        }
                    },
                    "role": {
                        "id": "0242ac1f-6606-c045-40cc-000000000016",
                        "name": "add-build-sshkey",
                        "path": "/var/lib/zuul/builds/7573dac089bc49b5b355588f880c7711/trusted/project_1/github.com/osism/openinfra-zuul-jobs/roles/add-build-sshkey"
                    },
                    "task": {
                        "duration": {
                            "end": "2026-02-24T00:00:19.672531Z",
                            "start": "2026-02-24T00:00:18.761860Z"
                        },
                        "id": "0242ac1f-6606-c045-40cc-000000000018",
                        "name": "Check to see if ssh key was already created for this build"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "include_tasks",
                            "changed": false,
                            "include": "create-key-and-replace.yaml",
                            "include_args": {}
                        }
                    },
                    "role": {
                        "id": "0242ac1f-6606-c045-40cc-000000000016",
                        "name": "add-build-sshkey",
                        "path": "/var/lib/zuul/builds/7573dac089bc49b5b355588f880c7711/trusted/project_1/github.com/osism/openinfra-zuul-jobs/roles/add-build-sshkey"
                    },
                    "task": {
                        "duration": {
                            "end": "2026-02-24T00:00:19.716462Z",
                            "start": "2026-02-24T00:00:19.678789Z"
                        },
                        "id": "0242ac1f-6606-c045-40cc-000000000019",
                        "name": "Create a new key in workspace based on build UUID"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "command",
                            "changed": true,
                            "cmd": [
                                "ssh-keygen",
                                "-t",
                                "rsa",
                                "-N",
                                "",
                                "-C",
                                "zuul-build-sshkey",
                                "-f",
                                "/var/lib/zuul/builds/7573dac089bc49b5b355588f880c7711/work/7573dac089bc49b5b355588f880c7711_id_rsa",
                                "-b",
                                "3072"
                            ],
                            "delta": "0:00:02.584983",
                            "end": "2026-02-24 00:00:23.487331",
                            "invocation": {
                                "module_args": {
                                    "_raw_params": "ssh-keygen -t rsa -N '' -C 'zuul-build-sshkey' -f /var/lib/zuul/builds/7573dac089bc49b5b355588f880c7711/work/7573dac089bc49b5b355588f880c7711_id_rsa -b 3072",
                                    "_uses_shell": false,
                                    "argv": null,
                                    "chdir": null,
                                    "creates": null,
                                    "executable": null,
                                    "expand_argument_vars": true,
                                    "removes": null,
                                    "stdin": null,
                                    "stdin_add_newline": true,
                                    "strip_empty_ends": true,
                                    "zuul_ansible_split_streams": false,
                                    "zuul_log_id": "0242ac1f-6606-c045-40cc-0000000000aa-0-orchestrator",
                                    "zuul_output_max_bytes": 1073741824
                                }
                            },
                            "msg": "",
                            "rc": 0,
                            "start": "2026-02-24 00:00:20.902348",
                            "stderr": "",
                            "stderr_lines": [],
                            "stdout": "Generating public/private rsa key pair.\nYour identification has been saved in /var/lib/zuul/builds/7573dac089bc49b5b355588f880c7711/work/7573dac089bc49b5b355588f880c7711_id_rsa\nYour public key has been saved in /var/lib/zuul/builds/7573dac089bc49b5b355588f880c7711/work/7573dac089bc49b5b355588f880c7711_id_rsa.pub\nThe key fingerprint is:\nSHA256:PLer2VHwNc7KJO94S0erWaWXATigzw7cbgKu7jwUB/s zuul-build-sshkey\nThe key's randomart image is:\n+---[RSA 3072]----+\n|         .       |\n|   .    . . .    |\n|    o  .  .o .o  |\n|   o ...+  o.+.. |\n|    +. oS+o + +..|\n|   ..E. +o B o +o|\n|  .  . . +o = =..|\n|  ...   oo *.= . |\n|  o=.   o.+.=.   |\n+----[SHA256]-----+",
                            "stdout_lines": [
                                "Generating public/private rsa key pair.",
                                "Your identification has been saved in /var/lib/zuul/builds/7573dac089bc49b5b355588f880c7711/work/7573dac089bc49b5b355588f880c7711_id_rsa",
                                "Your public key has been saved in /var/lib/zuul/builds/7573dac089bc49b5b355588f880c7711/work/7573dac089bc49b5b355588f880c7711_id_rsa.pub",
                                "The key fingerprint is:",
                                "SHA256:PLer2VHwNc7KJO94S0erWaWXATigzw7cbgKu7jwUB/s zuul-build-sshkey",
                                "The key's randomart image is:",
                                "+---[RSA 3072]----+",
                                "|         .       |",
                                "|   .    . . .    |",
                                "|    o  .  .o .o  |",
                                "|   o ...+  o.+.. |",
                                "|    +. oS+o + +..|",
                                "|   ..E. +o B o +o|",
                                "|  .  . . +o = =..|",
                                "|  ...   oo *.= . |",
                                "|  o=.   o.+.=.   |",
                                "+----[SHA256]-----+"
                            ],
                            "zuul_log_id": "0242ac1f-6606-c045-40cc-0000000000aa-0-orchestrator"
                        }
                    },
                    "role": {
                        "id": "0242ac1f-6606-c045-40cc-000000000016",
                        "name": "add-build-sshkey",
                        "path": "/var/lib/zuul/builds/7573dac089bc49b5b355588f880c7711/trusted/project_1/github.com/osism/openinfra-zuul-jobs/roles/add-build-sshkey"
                    },
                    "task": {
                        "duration": {
                            "end": "2026-02-24T00:00:23.602397Z",
                            "start": "2026-02-24T00:00:19.756785Z"
                        },
                        "id": "0242ac1f-6606-c045-40cc-0000000000aa",
                        "name": "Create Temp SSH key"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "include_tasks",
                            "changed": false,
                            "include": "remote-linux.yaml",
                            "include_args": {}
                        }
                    },
                    "role": {
                        "id": "0242ac1f-6606-c045-40cc-000000000016",
                        "name": "add-build-sshkey",
                        "path": "/var/lib/zuul/builds/7573dac089bc49b5b355588f880c7711/trusted/project_1/github.com/osism/openinfra-zuul-jobs/roles/add-build-sshkey"
                    },
                    "task": {
                        "duration": {
                            "end": "2026-02-24T00:00:23.637533Z",
                            "start": "2026-02-24T00:00:23.609219Z"
                        },
                        "id": "0242ac1f-6606-c045-40cc-0000000000ab",
                        "name": "Remote setup ssh keys (linux)"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "lineinfile",
                            "changed": false,
                            "false_condition": "zuul_build_sshkey_cleanup",
                            "skip_reason": "Conditional result was False",
                            "skipped": true
                        }
                    },
                    "role": {
                        "id": "0242ac1f-6606-c045-40cc-000000000016",
                        "name": "add-build-sshkey",
                        "path": "/var/lib/zuul/builds/7573dac089bc49b5b355588f880c7711/trusted/project_1/github.com/osism/openinfra-zuul-jobs/roles/add-build-sshkey"
                    },
                    "task": {
                        "duration": {
                            "end": "2026-02-24T00:00:23.743631Z",
                            "start": "2026-02-24T00:00:23.684828Z"
                        },
                        "id": "0242ac1f-6606-c045-40cc-0000000000ca",
                        "name": "Remove previously added zuul-build-sshkey"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "authorized_key",
                            "changed": true,
                            "comment": null,
                            "exclusive": false,
                            "follow": false,
                            "invocation": {
                                "module_args": {
                                    "changed": true,
                                    "comment": null,
                                    "exclusive": false,
                                    "follow": false,
                                    "key": "ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABgQDJ+0S2SO1gCkr7fCftNF31nAe7WVxUgGUxJF48Xyi6ePW6/DHF3iqRVS/363OM7JwGtrYwwa6lUw9TvKCfs6kERvEV/scmx6xhtS6HR3g2+dcUCzdp4MhDCTkEn/MRojg/TkmQ1P1rrJ/bwYBKuS6hf0TD7JdvpVuHIO1KA4v4V3/jLLeUw66yYIs+xouIj54PwovhrVz3OGSTsytjZVfmgNtu8VQ3bRxxThZ+hkEJia3JJ/TiwbXkL0k7PVDEZFvTnFHpjN8T1a3Kkv2sJ9sDQpbeeeTZFmxon++pifXN8/AawZnaIXWdN5lEP5BmHMzSTM2KVor591v9kFRpNeBuQ/Gr3jE5ZIggAdzRINLXV1dT3dTMRe1TKtQm6P6XXnCjXZV/OAZmB145nAfNREpRJWx5UQRXRks9S3kS07N5MRWP3XhMVBB6gA6m76tbQq+5cbF9zSlZ7ABi6Wy3/LIUnHt8TYjWgKY23KPHyqMqy+nxd4mHcKScBp29NZClNtU= zuul-build-sshkey",
                                    "key_options": null,
                                    "keyfile": "/home/zuul-testbed02/.ssh/authorized_keys",
                                    "manage_dir": true,
                                    "path": null,
                                    "state": "present",
                                    "user": "zuul-testbed02",
                                    "validate_certs": true
                                }
                            },
                            "key": "ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABgQDJ+0S2SO1gCkr7fCftNF31nAe7WVxUgGUxJF48Xyi6ePW6/DHF3iqRVS/363OM7JwGtrYwwa6lUw9TvKCfs6kERvEV/scmx6xhtS6HR3g2+dcUCzdp4MhDCTkEn/MRojg/TkmQ1P1rrJ/bwYBKuS6hf0TD7JdvpVuHIO1KA4v4V3/jLLeUw66yYIs+xouIj54PwovhrVz3OGSTsytjZVfmgNtu8VQ3bRxxThZ+hkEJia3JJ/TiwbXkL0k7PVDEZFvTnFHpjN8T1a3Kkv2sJ9sDQpbeeeTZFmxon++pifXN8/AawZnaIXWdN5lEP5BmHMzSTM2KVor591v9kFRpNeBuQ/Gr3jE5ZIggAdzRINLXV1dT3dTMRe1TKtQm6P6XXnCjXZV/OAZmB145nAfNREpRJWx5UQRXRks9S3kS07N5MRWP3XhMVBB6gA6m76tbQq+5cbF9zSlZ7ABi6Wy3/LIUnHt8TYjWgKY23KPHyqMqy+nxd4mHcKScBp29NZClNtU= zuul-build-sshkey",
                            "key_options": null,
                            "keyfile": "/home/zuul-testbed02/.ssh/authorized_keys",
                            "manage_dir": true,
                            "path": null,
                            "state": "present",
                            "user": "zuul-testbed02",
                            "validate_certs": true
                        }
                    },
                    "role": {
                        "id": "0242ac1f-6606-c045-40cc-000000000016",
                        "name": "add-build-sshkey",
                        "path": "/var/lib/zuul/builds/7573dac089bc49b5b355588f880c7711/trusted/project_1/github.com/osism/openinfra-zuul-jobs/roles/add-build-sshkey"
                    },
                    "task": {
                        "duration": {
                            "end": "2026-02-24T00:00:24.761674Z",
                            "start": "2026-02-24T00:00:23.753651Z"
                        },
                        "id": "0242ac1f-6606-c045-40cc-0000000000cb",
                        "name": "Enable access via build key on all nodes"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "file",
                            "changed": false,
                            "diff": {
                                "after": {
                                    "path": "/home/zuul-testbed02/.ssh"
                                },
                                "before": {
                                    "path": "/home/zuul-testbed02/.ssh"
                                }
                            },
                            "gid": 1003,
                            "group": "zuul-testbed02",
                            "invocation": {
                                "module_args": {
                                    "_diff_peek": null,
                                    "_original_basename": null,
                                    "access_time": null,
                                    "access_time_format": "%Y%m%d%H%M.%S",
                                    "attributes": null,
                                    "follow": true,
                                    "force": false,
                                    "group": null,
                                    "mode": 448,
                                    "modification_time": null,
                                    "modification_time_format": "%Y%m%d%H%M.%S",
                                    "owner": null,
                                    "path": "/home/zuul-testbed02/.ssh",
                                    "recurse": false,
                                    "selevel": null,
                                    "serole": null,
                                    "setype": null,
                                    "seuser": null,
                                    "src": null,
                                    "state": "directory",
                                    "unsafe_writes": false
                                }
                            },
                            "mode": "0700",
                            "owner": "zuul-testbed02",
                            "path": "/home/zuul-testbed02/.ssh",
                            "size": 4096,
                            "state": "directory",
                            "uid": 1002
                        }
                    },
                    "role": {
                        "id": "0242ac1f-6606-c045-40cc-000000000016",
                        "name": "add-build-sshkey",
                        "path": "/var/lib/zuul/builds/7573dac089bc49b5b355588f880c7711/trusted/project_1/github.com/osism/openinfra-zuul-jobs/roles/add-build-sshkey"
                    },
                    "task": {
                        "duration": {
                            "end": "2026-02-24T00:00:25.122815Z",
                            "start": "2026-02-24T00:00:24.766858Z"
                        },
                        "id": "0242ac1f-6606-c045-40cc-0000000000cc",
                        "name": "Make sure user has a .ssh"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "copy",
                            "changed": false,
                            "dest": "/home/zuul-testbed02/.ssh/id_rsa",
                            "invocation": {
                                "dest": "~/.ssh/id_rsa",
                                "force": false,
                                "mode": 384,
                                "module_args": {
                                    "dest": "~/.ssh/id_rsa",
                                    "force": false,
                                    "mode": 384,
                                    "src": "/var/lib/zuul/builds/7573dac089bc49b5b355588f880c7711/work/7573dac089bc49b5b355588f880c7711_id_rsa"
                                },
                                "src": "/var/lib/zuul/builds/7573dac089bc49b5b355588f880c7711/work/7573dac089bc49b5b355588f880c7711_id_rsa"
                            },
                            "src": "/var/lib/zuul/builds/7573dac089bc49b5b355588f880c7711/work/7573dac089bc49b5b355588f880c7711_id_rsa"
                        }
                    },
                    "role": {
                        "id": "0242ac1f-6606-c045-40cc-000000000016",
                        "name": "add-build-sshkey",
                        "path": "/var/lib/zuul/builds/7573dac089bc49b5b355588f880c7711/trusted/project_1/github.com/osism/openinfra-zuul-jobs/roles/add-build-sshkey"
                    },
                    "task": {
                        "duration": {
                            "end": "2026-02-24T00:00:25.575231Z",
                            "start": "2026-02-24T00:00:25.128472Z"
                        },
                        "id": "0242ac1f-6606-c045-40cc-0000000000cd",
                        "name": "Install build private key as SSH key on all nodes"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "copy",
                            "changed": false,
                            "dest": "/home/zuul-testbed02/.ssh/id_rsa.pub",
                            "invocation": {
                                "dest": "~/.ssh/id_rsa.pub",
                                "force": false,
                                "mode": 420,
                                "module_args": {
                                    "dest": "~/.ssh/id_rsa.pub",
                                    "force": false,
                                    "mode": 420,
                                    "src": "/var/lib/zuul/builds/7573dac089bc49b5b355588f880c7711/work/7573dac089bc49b5b355588f880c7711_id_rsa.pub"
                                },
                                "src": "/var/lib/zuul/builds/7573dac089bc49b5b355588f880c7711/work/7573dac089bc49b5b355588f880c7711_id_rsa.pub"
                            },
                            "src": "/var/lib/zuul/builds/7573dac089bc49b5b355588f880c7711/work/7573dac089bc49b5b355588f880c7711_id_rsa.pub"
                        }
                    },
                    "role": {
                        "id": "0242ac1f-6606-c045-40cc-000000000016",
                        "name": "add-build-sshkey",
                        "path": "/var/lib/zuul/builds/7573dac089bc49b5b355588f880c7711/trusted/project_1/github.com/osism/openinfra-zuul-jobs/roles/add-build-sshkey"
                    },
                    "task": {
                        "duration": {
                            "end": "2026-02-24T00:00:26.124104Z",
                            "start": "2026-02-24T00:00:25.579878Z"
                        },
                        "id": "0242ac1f-6606-c045-40cc-0000000000ce",
                        "name": "Install build public key as SSH key on all nodes"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "include_tasks",
                            "changed": false,
                            "false_condition": "ansible_os_family == \"Windows\"",
                            "skip_reason": "Conditional result was False",
                            "skipped": true
                        }
                    },
                    "role": {
                        "id": "0242ac1f-6606-c045-40cc-000000000016",
                        "name": "add-build-sshkey",
                        "path": "/var/lib/zuul/builds/7573dac089bc49b5b355588f880c7711/trusted/project_1/github.com/osism/openinfra-zuul-jobs/roles/add-build-sshkey"
                    },
                    "task": {
                        "duration": {
                            "end": "2026-02-24T00:00:26.185651Z",
                            "start": "2026-02-24T00:00:26.128939Z"
                        },
                        "id": "0242ac1f-6606-c045-40cc-0000000000ac",
                        "name": "Remote setup ssh keys (windows)"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "sshagent_remove_keys",
                            "changed": true,
                            "invocation": {
                                "module_args": {
                                    "remove": "^(?!\\(stdin\\)).*"
                                }
                            },
                            "removed": [
                                "/var/ssh/nodepool"
                            ]
                        }
                    },
                    "role": {
                        "id": "0242ac1f-6606-c045-40cc-0000000000b1",
                        "name": "remove-zuul-sshkey",
                        "path": "/var/lib/zuul/builds/7573dac089bc49b5b355588f880c7711/trusted/project_1/github.com/osism/openinfra-zuul-jobs/roles/remove-zuul-sshkey"
                    },
                    "task": {
                        "duration": {
                            "end": "2026-02-24T00:00:27.585564Z",
                            "start": "2026-02-24T00:00:26.192026Z"
                        },
                        "id": "0242ac1f-6606-c045-40cc-0000000000b3",
                        "name": "Remove master key from local agent"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "command",
                            "changed": true,
                            "cmd": [
                                "ssh-add",
                                "/var/lib/zuul/builds/7573dac089bc49b5b355588f880c7711/work/7573dac089bc49b5b355588f880c7711_id_rsa"
                            ],
                            "delta": "0:00:00.032935",
                            "end": "2026-02-24 00:00:28.724965",
                            "invocation": {
                                "module_args": {
                                    "_raw_params": "ssh-add /var/lib/zuul/builds/7573dac089bc49b5b355588f880c7711/work/7573dac089bc49b5b355588f880c7711_id_rsa",
                                    "_uses_shell": false,
                                    "argv": null,
                                    "chdir": null,
                                    "creates": null,
                                    "executable": null,
                                    "expand_argument_vars": true,
                                    "removes": null,
                                    "stdin": null,
                                    "stdin_add_newline": true,
                                    "strip_empty_ends": true,
                                    "zuul_ansible_split_streams": false,
                                    "zuul_log_id": "0242ac1f-6606-c045-40cc-0000000000b9-0-orchestrator",
                                    "zuul_output_max_bytes": 1073741824
                                }
                            },
                            "msg": "",
                            "rc": 0,
                            "start": "2026-02-24 00:00:28.692030",
                            "stderr": "",
                            "stderr_lines": [],
                            "stdout": "Identity added: /var/lib/zuul/builds/7573dac089bc49b5b355588f880c7711/work/7573dac089bc49b5b355588f880c7711_id_rsa (zuul-build-sshkey)",
                            "stdout_lines": [
                                "Identity added: /var/lib/zuul/builds/7573dac089bc49b5b355588f880c7711/work/7573dac089bc49b5b355588f880c7711_id_rsa (zuul-build-sshkey)"
                            ],
                            "zuul_log_id": "0242ac1f-6606-c045-40cc-0000000000b9-0-orchestrator"
                        }
                    },
                    "role": {
                        "id": "0242ac1f-6606-c045-40cc-000000000016",
                        "name": "add-build-sshkey",
                        "path": "/var/lib/zuul/builds/7573dac089bc49b5b355588f880c7711/trusted/project_1/github.com/osism/openinfra-zuul-jobs/roles/add-build-sshkey"
                    },
                    "task": {
                        "duration": {
                            "end": "2026-02-24T00:00:28.851592Z",
                            "start": "2026-02-24T00:00:27.606734Z"
                        },
                        "id": "0242ac1f-6606-c045-40cc-0000000000b9",
                        "name": "Add back temp key"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "ping",
                            "changed": false,
                            "invocation": {
                                "module_args": {
                                    "data": "pong"
                                }
                            },
                            "ping": "pong"
                        }
                    },
                    "role": {
                        "id": "0242ac1f-6606-c045-40cc-000000000016",
                        "name": "add-build-sshkey",
                        "path": "/var/lib/zuul/builds/7573dac089bc49b5b355588f880c7711/trusted/project_1/github.com/osism/openinfra-zuul-jobs/roles/add-build-sshkey"
                    },
                    "task": {
                        "duration": {
                            "end": "2026-02-24T00:00:29.501452Z",
                            "start": "2026-02-24T00:00:28.858044Z"
                        },
                        "id": "0242ac1f-6606-c045-40cc-0000000000ba",
                        "name": "Verify we can still SSH to all nodes"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "command",
                            "changed": false,
                            "false_condition": "ansible_os_family == \"Windows\"",
                            "skip_reason": "Conditional result was False",
                            "skipped": true
                        }
                    },
                    "role": {
                        "id": "0242ac1f-6606-c045-40cc-000000000016",
                        "name": "add-build-sshkey",
                        "path": "/var/lib/zuul/builds/7573dac089bc49b5b355588f880c7711/trusted/project_1/github.com/osism/openinfra-zuul-jobs/roles/add-build-sshkey"
                    },
                    "task": {
                        "duration": {
                            "end": "2026-02-24T00:00:29.549280Z",
                            "start": "2026-02-24T00:00:29.506361Z"
                        },
                        "id": "0242ac1f-6606-c045-40cc-0000000000bb",
                        "name": "Verify we can still SSH to all nodes (windows)"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "zuul_console",
                            "changed": false,
                            "invocation": {
                                "module_args": {
                                    "path": "/tmp/console-{log_uuid}.log",
                                    "port": 19885,
                                    "state": "present"
                                }
                            }
                        }
                    },
                    "role": {
                        "id": "0242ac1f-6606-c045-40cc-00000000001b",
                        "name": "start-zuul-console",
                        "path": "/var/lib/zuul/builds/7573dac089bc49b5b355588f880c7711/trusted/project_1/github.com/osism/openinfra-zuul-jobs/roles/start-zuul-console"
                    },
                    "task": {
                        "duration": {
                            "end": "2026-02-24T00:00:30.169749Z",
                            "start": "2026-02-24T00:00:29.589276Z"
                        },
                        "id": "0242ac1f-6606-c045-40cc-00000000001d",
                        "name": "Start zuul_console daemon."
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "set_fact",
                            "ansible_facts": {
                                "zuul_info_dir": "/var/lib/zuul/builds/7573dac089bc49b5b355588f880c7711/work/logs/zuul-info"
                            },
                            "changed": false
                        }
                    },
                    "role": {
                        "id": "0242ac1f-6606-c045-40cc-00000000001f",
                        "name": "validate-host",
                        "path": "/var/lib/zuul/builds/7573dac089bc49b5b355588f880c7711/trusted/project_1/github.com/osism/openinfra-zuul-jobs/roles/validate-host"
                    },
                    "task": {
                        "duration": {
                            "end": "2026-02-24T00:00:30.236201Z",
                            "start": "2026-02-24T00:00:30.185656Z"
                        },
                        "id": "0242ac1f-6606-c045-40cc-000000000021",
                        "name": "Define zuul_info_dir fact"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "file",
                            "changed": false,
                            "diff": {
                                "after": {
                                    "path": "/var/lib/zuul/builds/7573dac089bc49b5b355588f880c7711/work/logs/zuul-info"
                                },
                                "before": {
                                    "path": "/var/lib/zuul/builds/7573dac089bc49b5b355588f880c7711/work/logs/zuul-info"
                                }
                            },
                            "gid": 0,
                            "group": "root",
                            "invocation": {
                                "module_args": {
                                    "_diff_peek": null,
                                    "_original_basename": null,
                                    "access_time": null,
                                    "access_time_format": "%Y%m%d%H%M.%S",
                                    "attributes": null,
                                    "follow": true,
                                    "force": false,
                                    "group": null,
                                    "mode": 493,
                                    "modification_time": null,
                                    "modification_time_format": "%Y%m%d%H%M.%S",
                                    "owner": null,
                                    "path": "/var/lib/zuul/builds/7573dac089bc49b5b355588f880c7711/work/logs/zuul-info",
                                    "recurse": false,
                                    "selevel": null,
                                    "serole": null,
                                    "setype": null,
                                    "seuser": null,
                                    "src": null,
                                    "state": "directory",
                                    "unsafe_writes": false
                                }
                            },
                            "mode": "0755",
                            "owner": "root",
                            "path": "/var/lib/zuul/builds/7573dac089bc49b5b355588f880c7711/work/logs/zuul-info",
                            "size": 4096,
                            "state": "directory",
                            "uid": 0
                        }
                    },
                    "role": {
                        "id": "0242ac1f-6606-c045-40cc-00000000001f",
                        "name": "validate-host",
                        "path": "/var/lib/zuul/builds/7573dac089bc49b5b355588f880c7711/trusted/project_1/github.com/osism/openinfra-zuul-jobs/roles/validate-host"
                    },
                    "task": {
                        "duration": {
                            "end": "2026-02-24T00:00:31.033377Z",
                            "start": "2026-02-24T00:00:30.242580Z"
                        },
                        "id": "0242ac1f-6606-c045-40cc-000000000022",
                        "name": "Ensure Zuul Ansible directory exists"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "setup",
                            "ansible_facts": {
                                "ansible_all_ipv4_addresses": [
                                    "10.128.0.21"
                                ],
                                "ansible_all_ipv6_addresses": [
                                    "2a13:1a81:8000:2::33d",
                                    "fe80::f816:3eff:febf:60f0"
                                ],
                                "ansible_apparmor": {
                                    "status": "enabled"
                                },
                                "ansible_architecture": "x86_64",
                                "ansible_bios_date": "04/01/2014",
                                "ansible_bios_vendor": "SeaBIOS",
                                "ansible_bios_version": "1.15.0-1",
                                "ansible_board_asset_tag": "NA",
                                "ansible_board_name": "NA",
                                "ansible_board_serial": "NA",
                                "ansible_board_vendor": "NA",
                                "ansible_board_version": "NA",
                                "ansible_chassis_asset_tag": "NA",
                                "ansible_chassis_serial": "NA",
                                "ansible_chassis_vendor": "QEMU",
                                "ansible_chassis_version": "pc-i440fx-6.2",
                                "ansible_cmdline": {
                                    "BOOT_IMAGE": "/boot/vmlinuz-6.1.0-23-cloud-amd64",
                                    "console": "ttyS0,115200",
                                    "consoleblank": "0",
                                    "earlyprintk": "ttyS0,115200",
                                    "ro": true,
                                    "root": "UUID=97ecbbc5-b8a6-45c0-878b-05c164bcc1ed"
                                },
                                "ansible_date_time": {
                                    "date": "2026-02-24",
                                    "day": "24",
                                    "epoch": "1771891232",
                                    "epoch_int": "1771891232",
                                    "hour": "00",
                                    "iso8601": "2026-02-24T00:00:32Z",
                                    "iso8601_basic": "20260224T000032126520",
                                    "iso8601_basic_short": "20260224T000032",
                                    "iso8601_micro": "2026-02-24T00:00:32.126520Z",
                                    "minute": "00",
                                    "month": "02",
                                    "second": "32",
                                    "time": "00:00:32",
                                    "tz": "UTC",
                                    "tz_dst": "UTC",
                                    "tz_offset": "+0000",
                                    "weekday": "Tuesday",
                                    "weekday_number": "2",
                                    "weeknumber": "08",
                                    "year": "2026"
                                },
                                "ansible_default_ipv4": {
                                    "address": "10.128.0.21",
                                    "alias": "ens3",
                                    "broadcast": "",
                                    "gateway": "10.128.0.17",
                                    "interface": "ens3",
                                    "macaddress": "fa:16:3e:bf:60:f0",
                                    "mtu": 1500,
                                    "netmask": "255.255.255.240",
                                    "network": "10.128.0.16",
                                    "prefix": "28",
                                    "type": "ether"
                                },
                                "ansible_default_ipv6": {
                                    "address": "2a13:1a81:8000:2::33d",
                                    "gateway": "fe80::f816:3eff:fe98:d9f",
                                    "interface": "ens3",
                                    "macaddress": "fa:16:3e:bf:60:f0",
                                    "mtu": 1500,
                                    "prefix": "128",
                                    "scope": "global",
                                    "type": "ether"
                                },
                                "ansible_device_links": {
                                    "ids": {
                                        "sda": [
                                            "scsi-0QEMU_QEMU_HARDDISK_c056a867-3435-47ac-9224-3c26d34e1b7c"
                                        ],
                                        "sda1": [
                                            "scsi-0QEMU_QEMU_HARDDISK_c056a867-3435-47ac-9224-3c26d34e1b7c-part1"
                                        ],
                                        "sda14": [
                                            "scsi-0QEMU_QEMU_HARDDISK_c056a867-3435-47ac-9224-3c26d34e1b7c-part14"
                                        ],
                                        "sda15": [
                                            "scsi-0QEMU_QEMU_HARDDISK_c056a867-3435-47ac-9224-3c26d34e1b7c-part15"
                                        ]
                                    },
                                    "labels": {},
                                    "masters": {},
                                    "uuids": {
                                        "sda1": [
                                            "97ecbbc5-b8a6-45c0-878b-05c164bcc1ed"
                                        ],
                                        "sda15": [
                                            "1E24-44C9"
                                        ]
                                    }
                                },
                                "ansible_devices": {
                                    "loop0": {
                                        "holders": [],
                                        "host": "",
                                        "links": {
                                            "ids": [],
                                            "labels": [],
                                            "masters": [],
                                            "uuids": []
                                        },
                                        "model": null,
                                        "partitions": {},
                                        "removable": "0",
                                        "rotational": "1",
                                        "sas_address": null,
                                        "sas_device_handle": null,
                                        "scheduler_mode": "none",
                                        "sectors": "0",
                                        "sectorsize": "512",
                                        "size": "0.00 Bytes",
                                        "support_discard": "0",
                                        "vendor": null,
                                        "virtual": 1
                                    },
                                    "loop1": {
                                        "holders": [],
                                        "host": "",
                                        "links": {
                                            "ids": [],
                                            "labels": [],
                                            "masters": [],
                                            "uuids": []
                                        },
                                        "model": null,
                                        "partitions": {},
                                        "removable": "0",
                                        "rotational": "1",
                                        "sas_address": null,
                                        "sas_device_handle": null,
                                        "scheduler_mode": "none",
                                        "sectors": "0",
                                        "sectorsize": "512",
                                        "size": "0.00 Bytes",
                                        "support_discard": "0",
                                        "vendor": null,
                                        "virtual": 1
                                    },
                                    "loop2": {
                                        "holders": [],
                                        "host": "",
                                        "links": {
                                            "ids": [],
                                            "labels": [],
                                            "masters": [],
                                            "uuids": []
                                        },
                                        "model": null,
                                        "partitions": {},
                                        "removable": "0",
                                        "rotational": "1",
                                        "sas_address": null,
                                        "sas_device_handle": null,
                                        "scheduler_mode": "none",
                                        "sectors": "0",
                                        "sectorsize": "512",
                                        "size": "0.00 Bytes",
                                        "support_discard": "0",
                                        "vendor": null,
                                        "virtual": 1
                                    },
                                    "loop3": {
                                        "holders": [],
                                        "host": "",
                                        "links": {
                                            "ids": [],
                                            "labels": [],
                                            "masters": [],
                                            "uuids": []
                                        },
                                        "model": null,
                                        "partitions": {},
                                        "removable": "0",
                                        "rotational": "1",
                                        "sas_address": null,
                                        "sas_device_handle": null,
                                        "scheduler_mode": "none",
                                        "sectors": "0",
                                        "sectorsize": "512",
                                        "size": "0.00 Bytes",
                                        "support_discard": "0",
                                        "vendor": null,
                                        "virtual": 1
                                    },
                                    "loop4": {
                                        "holders": [],
                                        "host": "",
                                        "links": {
                                            "ids": [],
                                            "labels": [],
                                            "masters": [],
                                            "uuids": []
                                        },
                                        "model": null,
                                        "partitions": {},
                                        "removable": "0",
                                        "rotational": "1",
                                        "sas_address": null,
                                        "sas_device_handle": null,
                                        "scheduler_mode": "none",
                                        "sectors": "0",
                                        "sectorsize": "512",
                                        "size": "0.00 Bytes",
                                        "support_discard": "0",
                                        "vendor": null,
                                        "virtual": 1
                                    },
                                    "loop5": {
                                        "holders": [],
                                        "host": "",
                                        "links": {
                                            "ids": [],
                                            "labels": [],
                                            "masters": [],
                                            "uuids": []
                                        },
                                        "model": null,
                                        "partitions": {},
                                        "removable": "0",
                                        "rotational": "1",
                                        "sas_address": null,
                                        "sas_device_handle": null,
                                        "scheduler_mode": "none",
                                        "sectors": "0",
                                        "sectorsize": "512",
                                        "size": "0.00 Bytes",
                                        "support_discard": "0",
                                        "vendor": null,
                                        "virtual": 1
                                    },
                                    "loop6": {
                                        "holders": [],
                                        "host": "",
                                        "links": {
                                            "ids": [],
                                            "labels": [],
                                            "masters": [],
                                            "uuids": []
                                        },
                                        "model": null,
                                        "partitions": {},
                                        "removable": "0",
                                        "rotational": "1",
                                        "sas_address": null,
                                        "sas_device_handle": null,
                                        "scheduler_mode": "none",
                                        "sectors": "0",
                                        "sectorsize": "512",
                                        "size": "0.00 Bytes",
                                        "support_discard": "0",
                                        "vendor": null,
                                        "virtual": 1
                                    },
                                    "loop7": {
                                        "holders": [],
                                        "host": "",
                                        "links": {
                                            "ids": [],
                                            "labels": [],
                                            "masters": [],
                                            "uuids": []
                                        },
                                        "model": null,
                                        "partitions": {},
                                        "removable": "0",
                                        "rotational": "1",
                                        "sas_address": null,
                                        "sas_device_handle": null,
                                        "scheduler_mode": "none",
                                        "sectors": "0",
                                        "sectorsize": "512",
                                        "size": "0.00 Bytes",
                                        "support_discard": "0",
                                        "vendor": null,
                                        "virtual": 1
                                    },
                                    "sda": {
                                        "holders": [],
                                        "host": "SCSI storage controller: Red Hat, Inc. Virtio SCSI",
                                        "links": {
                                            "ids": [
                                                "scsi-0QEMU_QEMU_HARDDISK_c056a867-3435-47ac-9224-3c26d34e1b7c"
                                            ],
                                            "labels": [],
                                            "masters": [],
                                            "uuids": []
                                        },
                                        "model": "QEMU HARDDISK",
                                        "partitions": {
                                            "sda1": {
                                                "holders": [],
                                                "links": {
                                                    "ids": [
                                                        "scsi-0QEMU_QEMU_HARDDISK_c056a867-3435-47ac-9224-3c26d34e1b7c-part1"
                                                    ],
                                                    "labels": [],
                                                    "masters": [],
                                                    "uuids": [
                                                        "97ecbbc5-b8a6-45c0-878b-05c164bcc1ed"
                                                    ]
                                                },
                                                "sectors": "209453023",
                                                "sectorsize": 512,
                                                "size": "99.87 GB",
                                                "start": "262144",
                                                "uuid": "97ecbbc5-b8a6-45c0-878b-05c164bcc1ed"
                                            },
                                            "sda14": {
                                                "holders": [],
                                                "links": {
                                                    "ids": [
                                                        "scsi-0QEMU_QEMU_HARDDISK_c056a867-3435-47ac-9224-3c26d34e1b7c-part14"
                                                    ],
                                                    "labels": [],
                                                    "masters": [],
                                                    "uuids": []
                                                },
                                                "sectors": "6144",
                                                "sectorsize": 512,
                                                "size": "3.00 MB",
                                                "start": "2048",
                                                "uuid": null
                                            },
                                            "sda15": {
                                                "holders": [],
                                                "links": {
                                                    "ids": [
                                                        "scsi-0QEMU_QEMU_HARDDISK_c056a867-3435-47ac-9224-3c26d34e1b7c-part15"
                                                    ],
                                                    "labels": [],
                                                    "masters": [],
                                                    "uuids": [
                                                        "1E24-44C9"
                                                    ]
                                                },
                                                "sectors": "253952",
                                                "sectorsize": 512,
                                                "size": "124.00 MB",
                                                "start": "8192",
                                                "uuid": "1E24-44C9"
                                            }
                                        },
                                        "removable": "0",
                                        "rotational": "1",
                                        "sas_address": null,
                                        "sas_device_handle": null,
                                        "scheduler_mode": "none",
                                        "sectors": "209715200",
                                        "sectorsize": "512",
                                        "size": "100.00 GB",
                                        "support_discard": "4096",
                                        "vendor": "QEMU",
                                        "virtual": 1
                                    }
                                },
                                "ansible_distribution": "Debian",
                                "ansible_distribution_file_parsed": true,
                                "ansible_distribution_file_path": "/etc/os-release",
                                "ansible_distribution_file_variety": "Debian",
                                "ansible_distribution_major_version": "12",
                                "ansible_distribution_minor_version": "13",
                                "ansible_distribution_release": "bookworm",
                                "ansible_distribution_version": "12.13",
                                "ansible_dns": {
                                    "nameservers": [
                                        "81.163.194.9",
                                        "81.163.194.10",
                                        "2a13:1a81::11:53",
                                        "2a13:1a81::12:53"
                                    ],
                                    "search": [
                                        "."
                                    ]
                                },
                                "ansible_domain": "",
                                "ansible_effective_group_id": 1003,
                                "ansible_effective_user_id": 1002,
                                "ansible_ens3": {
                                    "active": true,
                                    "device": "ens3",
                                    "features": {
                                        "esp_hw_offload": "off [fixed]",
                                        "esp_tx_csum_hw_offload": "off [fixed]",
                                        "fcoe_mtu": "off [fixed]",
                                        "generic_receive_offload": "on",
                                        "generic_segmentation_offload": "on",
                                        "highdma": "on [fixed]",
                                        "hsr_dup_offload": "off [fixed]",
                                        "hsr_fwd_offload": "off [fixed]",
                                        "hsr_tag_ins_offload": "off [fixed]",
                                        "hsr_tag_rm_offload": "off [fixed]",
                                        "hw_tc_offload": "off [fixed]",
                                        "l2_fwd_offload": "off [fixed]",
                                        "large_receive_offload": "off [fixed]",
                                        "loopback": "off [fixed]",
                                        "macsec_hw_offload": "off [fixed]",
                                        "netns_local": "off [fixed]",
                                        "ntuple_filters": "off [fixed]",
                                        "receive_hashing": "off [fixed]",
                                        "rx_all": "off [fixed]",
                                        "rx_checksumming": "on [fixed]",
                                        "rx_fcs": "off [fixed]",
                                        "rx_gro_hw": "on",
                                        "rx_gro_list": "off",
                                        "rx_udp_gro_forwarding": "off",
                                        "rx_udp_tunnel_port_offload": "off [fixed]",
                                        "rx_vlan_filter": "on [fixed]",
                                        "rx_vlan_offload": "off [fixed]",
                                        "rx_vlan_stag_filter": "off [fixed]",
                                        "rx_vlan_stag_hw_parse": "off [fixed]",
                                        "scatter_gather": "on",
                                        "tcp_segmentation_offload": "on",
                                        "tls_hw_record": "off [fixed]",
                                        "tls_hw_rx_offload": "off [fixed]",
                                        "tls_hw_tx_offload": "off [fixed]",
                                        "tx_checksum_fcoe_crc": "off [fixed]",
                                        "tx_checksum_ip_generic": "on",
                                        "tx_checksum_ipv4": "off [fixed]",
                                        "tx_checksum_ipv6": "off [fixed]",
                                        "tx_checksum_sctp": "off [fixed]",
                                        "tx_checksumming": "on",
                                        "tx_esp_segmentation": "off [fixed]",
                                        "tx_fcoe_segmentation": "off [fixed]",
                                        "tx_gre_csum_segmentation": "off [fixed]",
                                        "tx_gre_segmentation": "off [fixed]",
                                        "tx_gso_list": "off [fixed]",
                                        "tx_gso_partial": "off [fixed]",
                                        "tx_gso_robust": "on [fixed]",
                                        "tx_ipxip4_segmentation": "off [fixed]",
                                        "tx_ipxip6_segmentation": "off [fixed]",
                                        "tx_lockless": "off [fixed]",
                                        "tx_nocache_copy": "off",
                                        "tx_scatter_gather": "on",
                                        "tx_scatter_gather_fraglist": "off [fixed]",
                                        "tx_sctp_segmentation": "off [fixed]",
                                        "tx_tcp6_segmentation": "on",
                                        "tx_tcp_ecn_segmentation": "on",
                                        "tx_tcp_mangleid_segmentation": "off",
                                        "tx_tcp_segmentation": "on",
                                        "tx_tunnel_remcsum_segmentation": "off [fixed]",
                                        "tx_udp_segmentation": "off [fixed]",
                                        "tx_udp_tnl_csum_segmentation": "off [fixed]",
                                        "tx_udp_tnl_segmentation": "off [fixed]",
                                        "tx_vlan_offload": "off [fixed]",
                                        "tx_vlan_stag_hw_insert": "off [fixed]",
                                        "vlan_challenged": "off [fixed]"
                                    },
                                    "hw_timestamp_filters": [],
                                    "ipv4": {
                                        "address": "10.128.0.21",
                                        "broadcast": "",
                                        "netmask": "255.255.255.240",
                                        "network": "10.128.0.16",
                                        "prefix": "28"
                                    },
                                    "ipv6": [
                                        {
                                            "address": "2a13:1a81:8000:2::33d",
                                            "prefix": "128",
                                            "scope": "global"
                                        },
                                        {
                                            "address": "fe80::f816:3eff:febf:60f0",
                                            "prefix": "64",
                                            "scope": "link"
                                        }
                                    ],
                                    "macaddress": "fa:16:3e:bf:60:f0",
                                    "module": "virtio_net",
                                    "mtu": 1500,
                                    "pciid": "virtio1",
                                    "promisc": false,
                                    "speed": -1,
                                    "timestamping": [],
                                    "type": "ether"
                                },
                                "ansible_env": {
                                    "HOME": "/home/zuul-testbed02",
                                    "LANG": "C.UTF-8",
                                    "LOGNAME": "zuul-testbed02",
                                    "MOTD_SHOWN": "pam",
                                    "PATH": "/usr/local/bin:/usr/bin:/bin:/usr/games",
                                    "PWD": "/home/zuul-testbed02",
                                    "SHELL": "/bin/sh",
                                    "SSH_CLIENT": "81.163.193.69 34106 22",
                                    "SSH_CONNECTION": "81.163.193.69 34106 10.128.0.21 22",
                                    "USER": "zuul-testbed02",
                                    "XDG_RUNTIME_DIR": "/run/user/1002",
                                    "XDG_SESSION_CLASS": "user",
                                    "XDG_SESSION_ID": "20981",
                                    "XDG_SESSION_TYPE": "tty"
                                },
                                "ansible_fibre_channel_wwn": [],
                                "ansible_fips": false,
                                "ansible_form_factor": "Other",
                                "ansible_fqdn": "zuul-static-regiocloud-infra-1",
                                "ansible_hostname": "zuul-static-regiocloud-infra-1",
                                "ansible_hostnqn": "",
                                "ansible_interfaces": [
                                    "ens3",
                                    "lo"
                                ],
                                "ansible_is_chroot": false,
                                "ansible_iscsi_iqn": "",
                                "ansible_kernel": "6.1.0-23-cloud-amd64",
                                "ansible_kernel_version": "#1 SMP PREEMPT_DYNAMIC Debian 6.1.99-1 (2024-07-15)",
                                "ansible_lo": {
                                    "active": true,
                                    "device": "lo",
                                    "features": {
                                        "esp_hw_offload": "off [fixed]",
                                        "esp_tx_csum_hw_offload": "off [fixed]",
                                        "fcoe_mtu": "off [fixed]",
                                        "generic_receive_offload": "on",
                                        "generic_segmentation_offload": "on",
                                        "highdma": "on [fixed]",
                                        "hsr_dup_offload": "off [fixed]",
                                        "hsr_fwd_offload": "off [fixed]",
                                        "hsr_tag_ins_offload": "off [fixed]",
                                        "hsr_tag_rm_offload": "off [fixed]",
                                        "hw_tc_offload": "off [fixed]",
                                        "l2_fwd_offload": "off [fixed]",
                                        "large_receive_offload": "off [fixed]",
                                        "loopback": "on [fixed]",
                                        "macsec_hw_offload": "off [fixed]",
                                        "netns_local": "on [fixed]",
                                        "ntuple_filters": "off [fixed]",
                                        "receive_hashing": "off [fixed]",
                                        "rx_all": "off [fixed]",
                                        "rx_checksumming": "on [fixed]",
                                        "rx_fcs": "off [fixed]",
                                        "rx_gro_hw": "off [fixed]",
                                        "rx_gro_list": "off",
                                        "rx_udp_gro_forwarding": "off",
                                        "rx_udp_tunnel_port_offload": "off [fixed]",
                                        "rx_vlan_filter": "off [fixed]",
                                        "rx_vlan_offload": "off [fixed]",
                                        "rx_vlan_stag_filter": "off [fixed]",
                                        "rx_vlan_stag_hw_parse": "off [fixed]",
                                        "scatter_gather": "on",
                                        "tcp_segmentation_offload": "on",
                                        "tls_hw_record": "off [fixed]",
                                        "tls_hw_rx_offload": "off [fixed]",
                                        "tls_hw_tx_offload": "off [fixed]",
                                        "tx_checksum_fcoe_crc": "off [fixed]",
                                        "tx_checksum_ip_generic": "on [fixed]",
                                        "tx_checksum_ipv4": "off [fixed]",
                                        "tx_checksum_ipv6": "off [fixed]",
                                        "tx_checksum_sctp": "on [fixed]",
                                        "tx_checksumming": "on",
                                        "tx_esp_segmentation": "off [fixed]",
                                        "tx_fcoe_segmentation": "off [fixed]",
                                        "tx_gre_csum_segmentation": "off [fixed]",
                                        "tx_gre_segmentation": "off [fixed]",
                                        "tx_gso_list": "on",
                                        "tx_gso_partial": "off [fixed]",
                                        "tx_gso_robust": "off [fixed]",
                                        "tx_ipxip4_segmentation": "off [fixed]",
                                        "tx_ipxip6_segmentation": "off [fixed]",
                                        "tx_lockless": "on [fixed]",
                                        "tx_nocache_copy": "off [fixed]",
                                        "tx_scatter_gather": "on [fixed]",
                                        "tx_scatter_gather_fraglist": "on [fixed]",
                                        "tx_sctp_segmentation": "on",
                                        "tx_tcp6_segmentation": "on",
                                        "tx_tcp_ecn_segmentation": "on",
                                        "tx_tcp_mangleid_segmentation": "on",
                                        "tx_tcp_segmentation": "on",
                                        "tx_tunnel_remcsum_segmentation": "off [fixed]",
                                        "tx_udp_segmentation": "on",
                                        "tx_udp_tnl_csum_segmentation": "off [fixed]",
                                        "tx_udp_tnl_segmentation": "off [fixed]",
                                        "tx_vlan_offload": "off [fixed]",
                                        "tx_vlan_stag_hw_insert": "off [fixed]",
                                        "vlan_challenged": "on [fixed]"
                                    },
                                    "hw_timestamp_filters": [],
                                    "ipv4": {
                                        "address": "127.0.0.1",
                                        "broadcast": "",
                                        "netmask": "255.0.0.0",
                                        "network": "127.0.0.0",
                                        "prefix": "8"
                                    },
                                    "ipv6": [
                                        {
                                            "address": "::1",
                                            "prefix": "128",
                                            "scope": "host"
                                        }
                                    ],
                                    "mtu": 65536,
                                    "promisc": false,
                                    "timestamping": [],
                                    "type": "loopback"
                                },
                                "ansible_loadavg": {
                                    "15m": 0.00927734375,
                                    "1m": 0.13037109375,
                                    "5m": 0.04296875
                                },
                                "ansible_local": {},
                                "ansible_locally_reachable_ips": {
                                    "ipv4": [
                                        "10.128.0.21",
                                        "127.0.0.0/8",
                                        "127.0.0.1"
                                    ],
                                    "ipv6": [
                                        "::1",
                                        "2a13:1a81:8000:2::33d",
                                        "fe80::f816:3eff:febf:60f0"
                                    ]
                                },
                                "ansible_lsb": {
                                    "codename": "bookworm",
                                    "description": "Debian GNU/Linux 12 (bookworm)",
                                    "id": "Debian",
                                    "major_release": "12",
                                    "release": "12"
                                },
                                "ansible_lvm": "N/A",
                                "ansible_machine": "x86_64",
                                "ansible_machine_id": "28f0fb0ddd3c48c38b3aa31603cb444d",
                                "ansible_memfree_mb": 460,
                                "ansible_memory_mb": {
                                    "nocache": {
                                        "free": 7212,
                                        "used": 740
                                    },
                                    "real": {
                                        "free": 460,
                                        "total": 7952,
                                        "used": 7492
                                    },
                                    "swap": {
                                        "cached": 0,
                                        "free": 0,
                                        "total": 0,
                                        "used": 0
                                    }
                                },
                                "ansible_memtotal_mb": 7952,
                                "ansible_mounts": [
                                    {
                                        "block_available": 17325954,
                                        "block_size": 4096,
                                        "block_total": 25759134,
                                        "block_used": 8433180,
                                        "device": "/dev/sda1",
                                        "fstype": "ext4",
                                        "inode_available": 5917786,
                                        "inode_total": 6545408,
                                        "inode_used": 627622,
                                        "mount": "/",
                                        "options": "rw,relatime,discard,errors=remount-ro",
                                        "size_available": 70967107584,
                                        "size_total": 105509412864,
                                        "uuid": "97ecbbc5-b8a6-45c0-878b-05c164bcc1ed"
                                    },
                                    {
                                        "block_available": 57419,
                                        "block_size": 2048,
                                        "block_total": 63339,
                                        "block_used": 5920,
                                        "device": "/dev/sda15",
                                        "fstype": "vfat",
                                        "inode_available": 0,
                                        "inode_total": 0,
                                        "inode_used": 0,
                                        "mount": "/boot/efi",
                                        "options": "rw,relatime,fmask=0022,dmask=0022,codepage=437,iocharset=ascii,shortname=mixed,utf8,errors=remount-ro",
                                        "size_available": 117594112,
                                        "size_total": 129718272,
                                        "uuid": "1E24-44C9"
                                    }
                                ],
                                "ansible_nodename": "zuul-static-regiocloud-infra-1",
                                "ansible_os_family": "Debian",
                                "ansible_pkg_mgr": "apt",
                                "ansible_proc_cmdline": {
                                    "BOOT_IMAGE": "/boot/vmlinuz-6.1.0-23-cloud-amd64",
                                    "console": [
                                        "tty0",
                                        "ttyS0,115200"
                                    ],
                                    "consoleblank": "0",
                                    "earlyprintk": "ttyS0,115200",
                                    "ro": true,
                                    "root": "UUID=97ecbbc5-b8a6-45c0-878b-05c164bcc1ed"
                                },
                                "ansible_processor": [
                                    "0",
                                    "GenuineIntel",
                                    "Intel Xeon Processor (Icelake)",
                                    "1",
                                    "GenuineIntel",
                                    "Intel Xeon Processor (Icelake)",
                                    "2",
                                    "GenuineIntel",
                                    "Intel Xeon Processor (Icelake)",
                                    "3",
                                    "GenuineIntel",
                                    "Intel Xeon Processor (Icelake)"
                                ],
                                "ansible_processor_cores": 1,
                                "ansible_processor_count": 4,
                                "ansible_processor_nproc": 4,
                                "ansible_processor_threads_per_core": 1,
                                "ansible_processor_vcpus": 4,
                                "ansible_product_name": "OpenStack Nova",
                                "ansible_product_serial": "NA",
                                "ansible_product_uuid": "NA",
                                "ansible_product_version": "26.2.1",
                                "ansible_python": {
                                    "executable": "/usr/bin/python3",
                                    "has_sslcontext": true,
                                    "type": "cpython",
                                    "version": {
                                        "major": 3,
                                        "micro": 2,
                                        "minor": 11,
                                        "releaselevel": "final",
                                        "serial": 0
                                    },
                                    "version_info": [
                                        3,
                                        11,
                                        2,
                                        "final",
                                        0
                                    ]
                                },
                                "ansible_python_version": "3.11.2",
                                "ansible_real_group_id": 1003,
                                "ansible_real_user_id": 1002,
                                "ansible_selinux": {
                                    "status": "disabled"
                                },
                                "ansible_selinux_python_present": true,
                                "ansible_service_mgr": "systemd",
                                "ansible_ssh_host_key_dsa_public": "AAAAB3NzaC1kc3MAAACBAOVugYBPOvEOPI9DA30PvhAEr88mJeaFmKN4pkfnBHV/1A1OlDKtX7lKWCBfEUilcd+tQFVKXm4KY+K9q4J5x5tpd7cL7LRDCFKN+py3+jc6U5J9EZ0o5cVG84z0c3XLEL6cCljG9Etx/0pXjjELRh1SGuUZweEx0143OSiKs/9DAAAAFQCT7R5MDTPEdFXQcZxA/fPRmtWFswAAAIBp8Iz2Njfcxfnh/mdxMztzkdNXulDt679+ghvl5588eCIDPa9xMtojivk53JaV6V4R+wsBN+MaKufzMOoVxuDkZcmw940H/1kcr3yOsJcZIkU1GRkJiNnSEIWJ6dVaXMh+wwbjw+8xNCA3+n3fARqUK906ICT5RsRrBe7W2K7dIAAAAIAL9qFr0vTidcXXhyWKN8I2iBsHSKfb6cjbseShw3ZeZE+kJ3IXz3z9oe0OwKp52GhvJrOXEkxST9YbtutbsVz5iS7lDpisvIWfFtvRRYaOtyvE8b4PWpHrk4s5tssqGDQ7y7zSqmRrj1sf5w0pwokPGmoSnPWUi0Vp0yhlysI/1w==",
                                "ansible_ssh_host_key_dsa_public_keytype": "ssh-dss",
                                "ansible_ssh_host_key_ecdsa_public": "AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBAuyO5R2SR4XT/Y65sItV7OQrUJXmGJ+//L9L/VugDohOWInFrkaZ6yFJMccVS7fmApekshaCGYnkIVdXV72hzg=",
                                "ansible_ssh_host_key_ecdsa_public_keytype": "ecdsa-sha2-nistp256",
                                "ansible_ssh_host_key_ed25519_public": "AAAAC3NzaC1lZDI1NTE5AAAAILxLVCAcH99lPquKCT3PwHsvi+569sV28I73LxchEjeg",
                                "ansible_ssh_host_key_ed25519_public_keytype": "ssh-ed25519",
                                "ansible_ssh_host_key_rsa_public": "AAAAB3NzaC1yc2EAAAADAQABAAABgQCZ3YX5Qhzq+HT8jTGGcX7rLsF3GCwObHFepfMmBZuD+gTfvdY5sYFSoXiT70rLPBjqETBTS0SQwmo8sd5AJI1vdgIqI5Sry8p52ng8tDg6C5CONt+xAaYOmE+ZDD/6naNX56b/gmgqBRQqPQHPmuptNxTlimdAcpoi+p9/UFBLTObMYmNL+5DTqw1+XECJ5j5kj5Ez40NT5w/1XnelEMcdw06WJZrzNEFt4Qpr2Zf+ybh0s+Xcn9u7d/gCbwXIPJvJnT0iSkiOzFvXJgOzwpvPtoi/5cV7g1LmgWsMLadT6gHof2poQaofC+1Bi0KiB2ANARMumvVBULx03qKvsqulsnlQAix2ffMraa3aMdUzK5aZo3hBx+H1uyEznUlyG9TUT5zhLIhNObetbpJgv6OJHiknBW6v/80FGD9t1pDiAwwBc/tJ8TyKpwyAdGMiEBIIVxTzx8QfPuApPJBJZXhMo2FBwRAEDEuuHbZu0znjhKKz4zlFT8TDrW+W7iThTvc=",
                                "ansible_ssh_host_key_rsa_public_keytype": "ssh-rsa",
                                "ansible_swapfree_mb": 0,
                                "ansible_swaptotal_mb": 0,
                                "ansible_system": "Linux",
                                "ansible_system_capabilities": [
                                    ""
                                ],
                                "ansible_system_capabilities_enforced": "True",
                                "ansible_system_vendor": "OpenStack Foundation",
                                "ansible_uptime_seconds": 47221207,
                                "ansible_user_dir": "/home/zuul-testbed02",
                                "ansible_user_gecos": "",
                                "ansible_user_gid": 1003,
                                "ansible_user_id": "zuul-testbed02",
                                "ansible_user_shell": "/bin/sh",
                                "ansible_user_uid": 1002,
                                "ansible_userspace_architecture": "x86_64",
                                "ansible_userspace_bits": "64",
                                "ansible_virtualization_role": "guest",
                                "ansible_virtualization_tech_guest": [
                                    "openstack"
                                ],
                                "ansible_virtualization_tech_host": [
                                    "kvm"
                                ],
                                "ansible_virtualization_type": "openstack",
                                "gather_subset": [
                                    "all"
                                ],
                                "module_setup": true
                            },
                            "changed": false,
                            "invocation": {
                                "module_args": {
                                    "fact_path": "/etc/ansible/facts.d",
                                    "filter": [],
                                    "gather_subset": [
                                        "all"
                                    ],
                                    "gather_timeout": 10
                                }
                            }
                        }
                    },
                    "role": {
                        "id": "0242ac1f-6606-c045-40cc-00000000001f",
                        "name": "validate-host",
                        "path": "/var/lib/zuul/builds/7573dac089bc49b5b355588f880c7711/trusted/project_1/github.com/osism/openinfra-zuul-jobs/roles/validate-host"
                    },
                    "task": {
                        "duration": {
                            "end": "2026-02-24T00:00:32.709948Z",
                            "start": "2026-02-24T00:00:31.038812Z"
                        },
                        "id": "0242ac1f-6606-c045-40cc-000000000023",
                        "name": "Collect information about the host"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "set_fact",
                            "ansible_facts": {
                                "validate_host_hostname_sanitized": "orchestrator"
                            },
                            "changed": false
                        }
                    },
                    "role": {
                        "id": "0242ac1f-6606-c045-40cc-00000000001f",
                        "name": "validate-host",
                        "path": "/var/lib/zuul/builds/7573dac089bc49b5b355588f880c7711/trusted/project_1/github.com/osism/openinfra-zuul-jobs/roles/validate-host"
                    },
                    "task": {
                        "duration": {
                            "end": "2026-02-24T00:00:32.935951Z",
                            "start": "2026-02-24T00:00:32.747462Z"
                        },
                        "id": "0242ac1f-6606-c045-40cc-000000000024",
                        "name": "Sanitize hostname"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "template",
                            "changed": true,
                            "checksum": "31284934f16941ef394f58b8c646dab6e479d33e",
                            "dest": "/var/lib/zuul/builds/7573dac089bc49b5b355588f880c7711/work/logs/zuul-info/host-info.orchestrator.yaml",
                            "diff": [],
                            "gid": 0,
                            "group": "root",
                            "invocation": {
                                "module_args": {
                                    "_original_basename": "host-info.j2",
                                    "attributes": null,
                                    "backup": false,
                                    "checksum": "31284934f16941ef394f58b8c646dab6e479d33e",
                                    "content": null,
                                    "dest": "/var/lib/zuul/builds/7573dac089bc49b5b355588f880c7711/work/logs/zuul-info/host-info.orchestrator.yaml",
                                    "directory_mode": null,
                                    "follow": false,
                                    "force": true,
                                    "group": null,
                                    "local_follow": null,
                                    "mode": 420,
                                    "owner": null,
                                    "remote_src": null,
                                    "selevel": null,
                                    "serole": null,
                                    "setype": null,
                                    "seuser": null,
                                    "src": "/var/lib/zuul/builds/7573dac089bc49b5b355588f880c7711/work/.ansible/tmp/ansible-tmp-1771891232.9758115-238-99675684833445/source",
                                    "unsafe_writes": false,
                                    "validate": null
                                }
                            },
                            "md5sum": "19521c67cd108f90332b281b5ac6893d",
                            "mode": "0644",
                            "owner": "root",
                            "size": 19728,
                            "src": "/var/lib/zuul/builds/7573dac089bc49b5b355588f880c7711/work/.ansible/tmp/ansible-tmp-1771891232.9758115-238-99675684833445/source",
                            "state": "file",
                            "uid": 0
                        }
                    },
                    "role": {
                        "id": "0242ac1f-6606-c045-40cc-00000000001f",
                        "name": "validate-host",
                        "path": "/var/lib/zuul/builds/7573dac089bc49b5b355588f880c7711/trusted/project_1/github.com/osism/openinfra-zuul-jobs/roles/validate-host"
                    },
                    "task": {
                        "duration": {
                            "end": "2026-02-24T00:00:34.122324Z",
                            "start": "2026-02-24T00:00:32.943350Z"
                        },
                        "id": "0242ac1f-6606-c045-40cc-000000000025",
                        "name": "Write out all ansible variables/facts known for each host"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "_zuul_nolog_return": true,
                            "action": "zuul_debug_info",
                            "changed": false,
                            "df_i": "Filesystem      Inodes  IUsed   IFree IUse% Mounted on\nudev           1015018    312 1014706    1% /dev\ntmpfs          1017896    542 1017354    1% /run\n/dev/sda1      6545408 627622 5917786   10% /\ntmpfs          1017896      1 1017895    1% /dev/shm\ntmpfs          1017896      2 1017894    1% /run/lock\n/dev/sda15           0      0       0     - /boot/efi\ntmpfs           203579     17  203562    1% /run/user/1001\ntmpfs           203579     17  203562    1% /run/user/1005\ntmpfs           203579     17  203562    1% /run/user/1004\ntmpfs           203579     17  203562    1% /run/user/1006\ntmpfs           203579     17  203562    1% /run/user/1002\n",
                            "df_m": "Filesystem     1M-blocks  Used Available Use% Mounted on\nudev                3965     0      3965   0% /dev\ntmpfs                796     1       795   1% /run\n/dev/sda1         100622 28817     67680  30% /\ntmpfs               3977     0      3977   0% /dev/shm\ntmpfs                  5     0         5   0% /run/lock\n/dev/sda15           124    12       113  10% /boot/efi\ntmpfs                796     0       796   0% /run/user/1001\ntmpfs                796     0       796   0% /run/user/1005\ntmpfs                796     0       796   0% /run/user/1004\ntmpfs                796     0       796   0% /run/user/1006\ntmpfs                796     0       796   0% /run/user/1002\n",
                            "image_manifest_files": [],
                            "invocation": {
                                "module_args": {
                                    "image_manifest": null,
                                    "image_manifest_files": [
                                        "/etc/dib-builddate.txt",
                                        "/etc/image-hostname.txt"
                                    ],
                                    "ipv4_route_required": false,
                                    "ipv6_route_required": false,
                                    "traceroute_host": null
                                }
                            },
                            "network_interfaces": "1: lo: <LOOPBACK,UP,LOWER_UP> mtu 65536 qdisc noqueue state UNKNOWN group default qlen 1000\n    link/loopback 00:00:00:00:00:00 brd 00:00:00:00:00:00\n    inet 127.0.0.1/8 scope host lo\n       valid_lft forever preferred_lft forever\n    inet6 ::1/128 scope host noprefixroute \n       valid_lft forever preferred_lft forever\n2: ens3: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc fq_codel state UP group default qlen 1000\n    link/ether fa:16:3e:bf:60:f0 brd ff:ff:ff:ff:ff:ff\n    altname enp0s3\n    inet 10.128.0.21/28 metric 100 brd 10.128.0.31 scope global dynamic ens3\n       valid_lft 24829sec preferred_lft 24829sec\n    inet6 2a13:1a81:8000:2::33d/128 scope global noprefixroute \n       valid_lft forever preferred_lft forever\n    inet6 fe80::f816:3eff:febf:60f0/64 scope link \n       valid_lft forever preferred_lft forever\n",
                            "network_neighbors": "10.128.0.17 dev ens3 lladdr fa:16:3e:e3:0c:52 REACHABLE \nfe80::f816:3eff:fe98:d9f dev ens3 lladdr fa:16:3e:98:0d:9f router STALE \n",
                            "network_routing_v4": "default via 10.128.0.17 dev ens3 proto dhcp src 10.128.0.21 metric 100 \n10.128.0.16/28 dev ens3 proto kernel scope link src 10.128.0.21 metric 100 \n10.128.0.17 dev ens3 proto dhcp scope link src 10.128.0.21 metric 100 \n10.128.0.18 dev ens3 proto dhcp scope link src 10.128.0.21 metric 100 \n81.163.194.9 via 10.128.0.17 dev ens3 proto dhcp src 10.128.0.21 metric 100 \n81.163.194.10 via 10.128.0.17 dev ens3 proto dhcp src 10.128.0.21 metric 100 \n169.254.169.254 via 10.128.0.18 dev ens3 proto dhcp src 10.128.0.21 metric 100 \n",
                            "network_routing_v6": "2a13:1a81:8000:2::300/120 dev ens3 proto ra metric 100 pref medium\nfe80::/64 dev ens3 proto kernel metric 256 pref medium\ndefault via fe80::f816:3eff:fe98:d9f dev ens3 proto ra metric 100 mtu 1500 pref medium\n",
                            "proc_cpuinfo": "processor\t: 0\nvendor_id\t: GenuineIntel\ncpu family\t: 6\nmodel\t\t: 134\nmodel name\t: Intel Xeon Processor (Icelake)\nstepping\t: 0\nmicrocode\t: 0x1\ncpu MHz\t\t: 2299.998\ncache size\t: 16384 KB\nphysical id\t: 0\nsiblings\t: 1\ncore id\t\t: 0\ncpu cores\t: 1\napicid\t\t: 0\ninitial apicid\t: 0\nfpu\t\t: yes\nfpu_exception\t: yes\ncpuid level\t: 13\nwp\t\t: yes\nflags\t\t: fpu vme de pse tsc msr pae mce cx8 apic sep mtrr pge mca cmov pat pse36 clflush mmx fxsr sse sse2 ss syscall nx pdpe1gb rdtscp lm constant_tsc rep_good nopl xtopology cpuid tsc_known_freq pni pclmulqdq vmx ssse3 fma cx16 pcid sse4_1 sse4_2 x2apic movbe popcnt tsc_deadline_timer aes xsave avx f16c rdrand hypervisor lahf_lm abm 3dnowprefetch cpuid_fault invpcid_single ssbd ibrs ibpb stibp ibrs_enhanced tpr_shadow vnmi flexpriority ept vpid ept_ad fsgsbase tsc_adjust bmi1 avx2 smep bmi2 erms invpcid avx512f avx512dq rdseed adx smap avx512ifma clflushopt clwb avx512cd sha_ni avx512bw avx512vl xsaveopt xsavec xgetbv1 xsaves wbnoinvd arat avx512vbmi umip pku ospke avx512_vbmi2 gfni vaes vpclmulqdq avx512_vnni avx512_bitalg avx512_vpopcntdq la57 rdpid fsrm md_clear arch_capabilities\nvmx flags\t: vnmi preemption_timer posted_intr invvpid ept_x_only ept_ad ept_1gb flexpriority apicv tsc_offset vtpr mtf vapic ept vpid unrestricted_guest vapic_reg vid shadow_vmcs\nbugs\t\t: spectre_v1 spectre_v2 spec_store_bypass swapgs taa mmio_stale_data rfds bhi\nbogomips\t: 4599.99\nclflush size\t: 64\ncache_alignment\t: 64\naddress sizes\t: 40 bits physical, 57 bits virtual\npower management:\n\nprocessor\t: 1\nvendor_id\t: GenuineIntel\ncpu family\t: 6\nmodel\t\t: 134\nmodel name\t: Intel Xeon Processor (Icelake)\nstepping\t: 0\nmicrocode\t: 0x1\ncpu MHz\t\t: 2299.998\ncache size\t: 16384 KB\nphysical id\t: 1\nsiblings\t: 1\ncore id\t\t: 0\ncpu cores\t: 1\napicid\t\t: 1\ninitial apicid\t: 1\nfpu\t\t: yes\nfpu_exception\t: yes\ncpuid level\t: 13\nwp\t\t: yes\nflags\t\t: fpu vme de pse tsc msr pae mce cx8 apic sep mtrr pge mca cmov pat pse36 clflush mmx fxsr sse sse2 ss syscall nx pdpe1gb rdtscp lm constant_tsc rep_good nopl xtopology cpuid tsc_known_freq pni pclmulqdq vmx ssse3 fma cx16 pcid sse4_1 sse4_2 x2apic movbe popcnt tsc_deadline_timer aes xsave avx f16c rdrand hypervisor lahf_lm abm 3dnowprefetch cpuid_fault invpcid_single ssbd ibrs ibpb stibp ibrs_enhanced tpr_shadow vnmi flexpriority ept vpid ept_ad fsgsbase tsc_adjust bmi1 avx2 smep bmi2 erms invpcid avx512f avx512dq rdseed adx smap avx512ifma clflushopt clwb avx512cd sha_ni avx512bw avx512vl xsaveopt xsavec xgetbv1 xsaves wbnoinvd arat avx512vbmi umip pku ospke avx512_vbmi2 gfni vaes vpclmulqdq avx512_vnni avx512_bitalg avx512_vpopcntdq la57 rdpid fsrm md_clear arch_capabilities\nvmx flags\t: vnmi preemption_timer posted_intr invvpid ept_x_only ept_ad ept_1gb flexpriority apicv tsc_offset vtpr mtf vapic ept vpid unrestricted_guest vapic_reg vid shadow_vmcs\nbugs\t\t: spectre_v1 spectre_v2 spec_store_bypass swapgs taa mmio_stale_data rfds bhi\nbogomips\t: 4599.99\nclflush size\t: 64\ncache_alignment\t: 64\naddress sizes\t: 40 bits physical, 57 bits virtual\npower management:\n\nprocessor\t: 2\nvendor_id\t: GenuineIntel\ncpu family\t: 6\nmodel\t\t: 134\nmodel name\t: Intel Xeon Processor (Icelake)\nstepping\t: 0\nmicrocode\t: 0x1\ncpu MHz\t\t: 2299.998\ncache size\t: 16384 KB\nphysical id\t: 2\nsiblings\t: 1\ncore id\t\t: 0\ncpu cores\t: 1\napicid\t\t: 2\ninitial apicid\t: 2\nfpu\t\t: yes\nfpu_exception\t: yes\ncpuid level\t: 13\nwp\t\t: yes\nflags\t\t: fpu vme de pse tsc msr pae mce cx8 apic sep mtrr pge mca cmov pat pse36 clflush mmx fxsr sse sse2 ss syscall nx pdpe1gb rdtscp lm constant_tsc rep_good nopl xtopology cpuid tsc_known_freq pni pclmulqdq vmx ssse3 fma cx16 pcid sse4_1 sse4_2 x2apic movbe popcnt tsc_deadline_timer aes xsave avx f16c rdrand hypervisor lahf_lm abm 3dnowprefetch cpuid_fault invpcid_single ssbd ibrs ibpb stibp ibrs_enhanced tpr_shadow vnmi flexpriority ept vpid ept_ad fsgsbase tsc_adjust bmi1 avx2 smep bmi2 erms invpcid avx512f avx512dq rdseed adx smap avx512ifma clflushopt clwb avx512cd sha_ni avx512bw avx512vl xsaveopt xsavec xgetbv1 xsaves wbnoinvd arat avx512vbmi umip pku ospke avx512_vbmi2 gfni vaes vpclmulqdq avx512_vnni avx512_bitalg avx512_vpopcntdq la57 rdpid fsrm md_clear arch_capabilities\nvmx flags\t: vnmi preemption_timer posted_intr invvpid ept_x_only ept_ad ept_1gb flexpriority apicv tsc_offset vtpr mtf vapic ept vpid unrestricted_guest vapic_reg vid shadow_vmcs\nbugs\t\t: spectre_v1 spectre_v2 spec_store_bypass swapgs taa mmio_stale_data rfds bhi\nbogomips\t: 4599.99\nclflush size\t: 64\ncache_alignment\t: 64\naddress sizes\t: 40 bits physical, 57 bits virtual\npower management:\n\nprocessor\t: 3\nvendor_id\t: GenuineIntel\ncpu family\t: 6\nmodel\t\t: 134\nmodel name\t: Intel Xeon Processor (Icelake)\nstepping\t: 0\nmicrocode\t: 0x1\ncpu MHz\t\t: 2299.998\ncache size\t: 16384 KB\nphysical id\t: 3\nsiblings\t: 1\ncore id\t\t: 0\ncpu cores\t: 1\napicid\t\t: 3\ninitial apicid\t: 3\nfpu\t\t: yes\nfpu_exception\t: yes\ncpuid level\t: 13\nwp\t\t: yes\nflags\t\t: fpu vme de pse tsc msr pae mce cx8 apic sep mtrr pge mca cmov pat pse36 clflush mmx fxsr sse sse2 ss syscall nx pdpe1gb rdtscp lm constant_tsc rep_good nopl xtopology cpuid tsc_known_freq pni pclmulqdq vmx ssse3 fma cx16 pcid sse4_1 sse4_2 x2apic movbe popcnt tsc_deadline_timer aes xsave avx f16c rdrand hypervisor lahf_lm abm 3dnowprefetch cpuid_fault invpcid_single ssbd ibrs ibpb stibp ibrs_enhanced tpr_shadow vnmi flexpriority ept vpid ept_ad fsgsbase tsc_adjust bmi1 avx2 smep bmi2 erms invpcid avx512f avx512dq rdseed adx smap avx512ifma clflushopt clwb avx512cd sha_ni avx512bw avx512vl xsaveopt xsavec xgetbv1 xsaves wbnoinvd arat avx512vbmi umip pku ospke avx512_vbmi2 gfni vaes vpclmulqdq avx512_vnni avx512_bitalg avx512_vpopcntdq la57 rdpid fsrm md_clear arch_capabilities\nvmx flags\t: vnmi preemption_timer posted_intr invvpid ept_x_only ept_ad ept_1gb flexpriority apicv tsc_offset vtpr mtf vapic ept vpid unrestricted_guest vapic_reg vid shadow_vmcs\nbugs\t\t: spectre_v1 spectre_v2 spec_store_bypass swapgs taa mmio_stale_data rfds bhi\nbogomips\t: 4599.99\nclflush size\t: 64\ncache_alignment\t: 64\naddress sizes\t: 40 bits physical, 57 bits virtual\npower management:\n\n",
                            "traceroute": null,
                            "uname": "Linux zuul-static-regiocloud-infra-1 6.1.0-23-cloud-amd64 #1 SMP PREEMPT_DYNAMIC Debian 6.1.99-1 (2024-07-15) x86_64 GNU/Linux\n"
                        }
                    },
                    "role": {
                        "id": "0242ac1f-6606-c045-40cc-00000000001f",
                        "name": "validate-host",
                        "path": "/var/lib/zuul/builds/7573dac089bc49b5b355588f880c7711/trusted/project_1/github.com/osism/openinfra-zuul-jobs/roles/validate-host"
                    },
                    "task": {
                        "duration": {
                            "end": "2026-02-24T00:00:35.228498Z",
                            "start": "2026-02-24T00:00:34.128789Z"
                        },
                        "id": "0242ac1f-6606-c045-40cc-000000000027",
                        "name": "Collect information about zuul worker"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "template",
                            "changed": true,
                            "checksum": "94d6e7bfbd7750ff134bf8736dc1f612e3313c52",
                            "dest": "/var/lib/zuul/builds/7573dac089bc49b5b355588f880c7711/work/logs/zuul-info/zuul-info.orchestrator.txt",
                            "diff": [],
                            "gid": 0,
                            "group": "root",
                            "invocation": {
                                "module_args": {
                                    "_original_basename": "zuul-info.j2",
                                    "attributes": null,
                                    "backup": false,
                                    "checksum": "94d6e7bfbd7750ff134bf8736dc1f612e3313c52",
                                    "content": null,
                                    "dest": "/var/lib/zuul/builds/7573dac089bc49b5b355588f880c7711/work/logs/zuul-info/zuul-info.orchestrator.txt",
                                    "directory_mode": null,
                                    "follow": false,
                                    "force": true,
                                    "group": null,
                                    "local_follow": null,
                                    "mode": 420,
                                    "owner": null,
                                    "remote_src": null,
                                    "selevel": null,
                                    "serole": null,
                                    "setype": null,
                                    "seuser": null,
                                    "src": "/var/lib/zuul/builds/7573dac089bc49b5b355588f880c7711/work/.ansible/tmp/ansible-tmp-1771891235.282114-277-244501980400278/source",
                                    "unsafe_writes": false,
                                    "validate": null
                                }
                            },
                            "md5sum": "1a2a41ebe45b7659c62b285102100558",
                            "mode": "0644",
                            "owner": "root",
                            "size": 3489,
                            "src": "/var/lib/zuul/builds/7573dac089bc49b5b355588f880c7711/work/.ansible/tmp/ansible-tmp-1771891235.282114-277-244501980400278/source",
                            "state": "file",
                            "uid": 0
                        }
                    },
                    "role": {
                        "id": "0242ac1f-6606-c045-40cc-00000000001f",
                        "name": "validate-host",
                        "path": "/var/lib/zuul/builds/7573dac089bc49b5b355588f880c7711/trusted/project_1/github.com/osism/openinfra-zuul-jobs/roles/validate-host"
                    },
                    "task": {
                        "duration": {
                            "end": "2026-02-24T00:00:36.426018Z",
                            "start": "2026-02-24T00:00:35.250424Z"
                        },
                        "id": "0242ac1f-6606-c045-40cc-000000000028",
                        "name": "Write out all zuul information for each host"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "zuul_console",
                            "changed": false,
                            "invocation": {
                                "module_args": {
                                    "path": "/tmp/console-{log_uuid}.log",
                                    "port": 19885,
                                    "state": "present"
                                }
                            }
                        }
                    },
                    "role": {
                        "id": "0242ac1f-6606-c045-40cc-00000000002a",
                        "name": "prepare-workspace-log",
                        "path": "/var/lib/zuul/builds/7573dac089bc49b5b355588f880c7711/trusted/project_0/github.com/osism/zuul-config/roles/prepare-workspace-log"
                    },
                    "task": {
                        "duration": {
                            "end": "2026-02-24T00:00:36.785675Z",
                            "start": "2026-02-24T00:00:36.435979Z"
                        },
                        "id": "0242ac1f-6606-c045-40cc-00000000002c",
                        "name": "Start zuul_console daemon."
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "synchronize",
                            "changed": true,
                            "cmd": "/usr/bin/rsync --compress --delete-after --archive --no-owner --no-group --rsh='/usr/bin/ssh -S none -o Port=22 -o StrictHostKeyChecking=no -o UserKnownHostsFile=/dev/null' --out-format='<<CHANGED>>%i %n%L' /var/lib/zuul/builds/7573dac089bc49b5b355588f880c7711/work/src zuul-testbed02@81.163.193.140:.",
                            "invocation": {
                                "module_args": {
                                    "_local_rsync_password": null,
                                    "_local_rsync_path": "rsync",
                                    "_ssh_args": null,
                                    "_substitute_controller": false,
                                    "archive": true,
                                    "checksum": false,
                                    "compress": true,
                                    "copy_links": false,
                                    "delay_updates": false,
                                    "delete": true,
                                    "dest": "zuul-testbed02@81.163.193.140:.",
                                    "dest_port": 22,
                                    "dirs": false,
                                    "existing_only": false,
                                    "group": false,
                                    "link_dest": null,
                                    "links": null,
                                    "mode": "push",
                                    "owner": false,
                                    "partial": false,
                                    "perms": null,
                                    "private_key": null,
                                    "recursive": true,
                                    "rsync_opts": [],
                                    "rsync_path": null,
                                    "rsync_timeout": 0,
                                    "set_remote_user": true,
                                    "src": "/var/lib/zuul/builds/7573dac089bc49b5b355588f880c7711/work/src",
                                    "ssh_connection_multiplexing": false,
                                    "times": null,
                                    "use_ssh_args": false,
                                    "verify_host": false
                                }
                            },
                            "msg": ".d..t...... src/\n.d..t...... src/github.com/\n.d..t...... src/github.com/osism/\n.d..t...... src/github.com/osism/ansible-collection-commons/\n<f..t...... src/github.com/osism/ansible-collection-commons/.ansible-lint\n<f..t...... src/github.com/osism/ansible-collection-commons/.flake8\n<f..t...... src/github.com/osism/ansible-collection-commons/.gitignore\n<f..t...... src/github.com/osism/ansible-collection-commons/.yamllint.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/.zuul.yaml\n<f..t...... src/github.com/osism/ansible-collection-commons/LICENSE\n<f..t...... src/github.com/osism/ansible-collection-commons/README.md\n<f..t...... src/github.com/osism/ansible-collection-commons/galaxy.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/\n<f..t...... src/github.com/osism/ansible-collection-commons/.git/HEAD\n<f..t...... src/github.com/osism/ansible-collection-commons/.git/ORIG_HEAD\n<f..t...... src/github.com/osism/ansible-collection-commons/.git/config\n<f..t...... src/github.com/osism/ansible-collection-commons/.git/config.worktree\n<f..t...... src/github.com/osism/ansible-collection-commons/.git/description\n<f..t...... src/github.com/osism/ansible-collection-commons/.git/index\n<f..t...... src/github.com/osism/ansible-collection-commons/.git/packed-refs\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/branches/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/hooks/\n<f..t...... src/github.com/osism/ansible-collection-commons/.git/hooks/applypatch-msg.sample\n<f..t...... src/github.com/osism/ansible-collection-commons/.git/hooks/commit-msg.sample\n<f..t...... src/github.com/osism/ansible-collection-commons/.git/hooks/fsmonitor-watchman.sample\n<f..t...... src/github.com/osism/ansible-collection-commons/.git/hooks/post-update.sample\n<f..t...... src/github.com/osism/ansible-collection-commons/.git/hooks/pre-applypatch.sample\n<f..t...... src/github.com/osism/ansible-collection-commons/.git/hooks/pre-commit.sample\n<f..t...... src/github.com/osism/ansible-collection-commons/.git/hooks/pre-merge-commit.sample\n<f..t...... src/github.com/osism/ansible-collection-commons/.git/hooks/pre-push.sample\n<f..t...... src/github.com/osism/ansible-collection-commons/.git/hooks/pre-rebase.sample\n<f..t...... src/github.com/osism/ansible-collection-commons/.git/hooks/pre-receive.sample\n<f..t...... src/github.com/osism/ansible-collection-commons/.git/hooks/prepare-commit-msg.sample\n<f..t...... src/github.com/osism/ansible-collection-commons/.git/hooks/push-to-checkout.sample\n<f..t...... src/github.com/osism/ansible-collection-commons/.git/hooks/update.sample\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/info/\n<f..t...... src/github.com/osism/ansible-collection-commons/.git/info/exclude\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/logs/\n<f..t...... src/github.com/osism/ansible-collection-commons/.git/logs/HEAD\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/logs/refs/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/logs/refs/heads/\n<f..t...... src/github.com/osism/ansible-collection-commons/.git/logs/refs/heads/main\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/logs/refs/remotes/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/logs/refs/remotes/origin/\n<f..t...... src/github.com/osism/ansible-collection-commons/.git/logs/refs/remotes/origin/HEAD\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/01/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/02/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/04/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/05/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/06/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/07/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/08/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/09/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/0a/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/0b/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/0c/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/0d/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/0e/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/0f/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/10/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/11/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/12/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/13/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/14/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/15/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/16/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/17/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/18/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/19/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/1a/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/1b/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/1d/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/1f/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/20/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/21/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/22/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/23/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/24/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/25/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/26/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/27/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/28/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/29/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/2a/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/2b/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/2c/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/2e/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/2f/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/30/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/31/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/32/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/33/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/35/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/36/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/37/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/38/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/39/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/3a/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/3b/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/3c/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/3d/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/3e/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/40/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/41/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/42/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/43/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/44/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/45/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/46/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/48/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/49/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/4a/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/4b/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/4c/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/4d/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/4e/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/4f/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/50/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/52/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/53/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/54/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/55/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/56/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/57/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/58/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/59/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/5a/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/5b/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/5c/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/5d/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/5e/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/5f/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/60/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/61/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/62/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/63/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/64/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/66/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/67/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/68/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/69/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/6a/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/6b/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/6c/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/6d/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/6e/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/70/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/71/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/72/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/73/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/74/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/75/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/76/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/77/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/78/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/79/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/7a/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/7b/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/7c/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/7d/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/7f/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/80/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/81/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/82/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/83/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/84/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/85/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/86/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/87/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/88/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/89/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/8a/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/8b/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/8c/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/8d/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/8e/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/8f/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/90/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/91/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/93/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/94/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/95/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/96/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/97/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/98/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/9a/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/9b/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/9c/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/9f/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/a0/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/a1/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/a2/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/a3/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/a4/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/a5/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/a6/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/a7/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/a8/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/a9/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/aa/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/ab/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/ac/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/ad/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/ae/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/af/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/b0/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/b1/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/b2/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/b3/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/b4/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/b5/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/b6/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/b7/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/b8/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/b9/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/ba/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/bb/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/bc/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/bd/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/be/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/bf/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/c0/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/c3/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/c4/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/c5/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/c6/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/c7/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/c9/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/ca/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/cb/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/cc/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/ce/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/cf/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/d0/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/d1/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/d2/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/d3/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/d4/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/d5/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/d6/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/d8/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/d9/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/da/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/db/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/dc/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/de/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/df/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/e0/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/e2/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/e3/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/e4/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/e5/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/e7/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/e8/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/e9/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/ea/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/eb/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/ec/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/ed/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/ee/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/ef/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/f0/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/f2/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/f3/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/f4/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/f5/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/f6/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/f7/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/f8/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/f9/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/fa/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/fb/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/fc/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/fd/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/fe/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/ff/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/info/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/pack/\n.d..t...... src/github.com/osism/ansible-collection-commons/.git/refs/\n.d..t...... src/github.com/osism/ansible-collection-commons/.github/\n<f..t...... src/github.com/osism/ansible-collection-commons/.github/publish-collection.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/.github/renovate.json\n.d..t...... src/github.com/osism/ansible-collection-commons/.github/templates/\n<f..t...... src/github.com/osism/ansible-collection-commons/.github/templates/galaxy.yml.j2\n.d..t...... src/github.com/osism/ansible-collection-commons/.github/workflows/\n<f..t...... src/github.com/osism/ansible-collection-commons/.github/workflows/publish-collection.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/meta/\n<f..t...... src/github.com/osism/ansible-collection-commons/meta/runtime.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/molecule/\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/requirements.txt\n.d..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/collections.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/converge.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/molecule.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/certificates.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/cleanup.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/configfs.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/configuration.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/docker_compose.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/docker_login.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/facts.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/firewall.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/hostname.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/hosts.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/ipmitool.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/k9s.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/kernel_modules.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/known_hosts.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/limits.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/lynis.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/microcode.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/motd.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/network.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/operator.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/packages.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/podman.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/proxy.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/repository.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/resolvconf.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/runc.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/services.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/sosreport.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/sshconfig.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/state.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/sysctl.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/systohc.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/timezone.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/trivy.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/user.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/__init__.py\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/certificates.py\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/configfs.py\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/docker_compose.py\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/docker_login.py\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/facts.py\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/firewall.py\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/hostname.py\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/ipmitool.py\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/k9s.py\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/kernel_modules.py\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/known_hosts.py\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/limits.py\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/microcode.py\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/operator.py\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/resolvconf.py\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/runc.py\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/services.py\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/sosreport.py\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/sshconfig.py\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/state.py\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/sysctl.py\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/systohc.py\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/user.py\n.d..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/cleanup/\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/cleanup/__init__.py\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/cleanup/cloudinit.py\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/cleanup/packages.py\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/cleanup/packages_debian.py\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/cleanup/services.py\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/cleanup/timer_debian.py\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/cleanup/timer_redhat.py\n.d..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/configuration/\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/configuration/__init__.py\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/configuration/git.py\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/configuration/main.py\n.d..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/hosts/\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/hosts/__init__.py\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/hosts/type-block.py\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/hosts/type-template.py\n.d..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/lynis/\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/lynis/__init__.py\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/lynis/debian.py\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/lynis/main.py\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/lynis/redhat.py\n.d..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/motd/\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/motd/__init__.py\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/motd/debian.py\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/motd/main.py\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/motd/redhat.py\n.d..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/network/\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/network/__init__.py\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/network/interfaces.py\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/network/netplan.py\n.d..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/packages/\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/packages/__init__.py\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/packages/debian.py\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/packages/main.py\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/packages/redhat.py\n.d..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/podman/\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/podman/__init__.py\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/podman/debian.py\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/podman/main.py\n.d..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/proxy/\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/proxy/__init__.py\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/proxy/debian.py\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/proxy/main.py\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/proxy/redhat.py\n.d..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/repository/\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/repository/__init__.py\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/repository/centos.py\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/repository/debian.py\n.d..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/timezone/\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/timezone/__init__.py\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/timezone/debian.py\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/timezone/main.py\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/timezone/redhat.py\n.d..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/trivy/\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/trivy/__init__.py\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/trivy/debian.py\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/trivy/main.py\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/trivy/redhat.py\n.d..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/util/\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/util/__init__.py\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/util/util.py\n.d..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/certificates.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/cleanup.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/configfs.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/configuration.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/docker_compose.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/docker_login.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/facts.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/firewall.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/hostname.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/hosts.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/ipmitool.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/k9s.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/kernel_modules.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/known_hosts.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/limits.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/lynis.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/microcode.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/motd.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/network.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/operator.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/packages.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/podman.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/proxy.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/repository.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/resolvconf.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/runc.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/services.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/sosreport.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/sshconfig.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/state.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/sysctl.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/systohc.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/timezone.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/trivy.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/user.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/verify/\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/verify/certificates.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/verify/cleanup.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/verify/configfs.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/verify/configuration.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/verify/docker_compose.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/verify/facts.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/verify/firewall.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/verify/hostname.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/verify/hosts.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/verify/ipmitool.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/verify/k9s.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/verify/kernel_modules.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/verify/lynis.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/verify/microcode.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/verify/motd.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/verify/operator.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/verify/packages.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/verify/podman.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/verify/proxy.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/verify/repository.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/verify/resolvconf.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/verify/runc.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/verify/services.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/verify/sosreport.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/verify/sshconfig.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/verify/state.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/verify/sysctl.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/verify/systohc.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/verify/timezone.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/verify/trivy.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/plugins/\n.d..t...... src/github.com/osism/ansible-collection-commons/plugins/callback/\n<f..t...... src/github.com/osism/ansible-collection-commons/plugins/callback/still_alive.py\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/README.md\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/certificates/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/certificates/README.md\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/certificates/defaults/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/certificates/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/certificates/handlers/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/certificates/handlers/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/certificates/meta/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/certificates/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/certificates/tasks/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/certificates/tasks/install-Debian-family.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/certificates/tasks/install-RedHat-family.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/certificates/tasks/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/certificates/templates/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/certificates/templates/certificates.j2\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/certificates/vars/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/certificates/vars/Debian-family.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/certificates/vars/RedHat-family.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/cleanup/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/cleanup/README.md\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/cleanup/defaults/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/cleanup/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/cleanup/meta/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/cleanup/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/cleanup/tasks/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/cleanup/tasks/cloudinit.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/cleanup/tasks/main.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/cleanup/tasks/packages-Debian-family.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/cleanup/tasks/packages-RedHat-family.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/cleanup/tasks/services-Debian-family.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/cleanup/tasks/services-RedHat-family.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/cleanup/tasks/timers-Debian-family.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/cleanup/tasks/timers-RedHat-family.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/cleanup/vars/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/cleanup/vars/Debian-family.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/cleanup/vars/RedHat-family.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/configfs/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/configfs/README.md\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/configfs/defaults/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/configfs/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/configfs/meta/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/configfs/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/configfs/tasks/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/configfs/tasks/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/configuration/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/configuration/README.md\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/configuration/defaults/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/configuration/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/configuration/meta/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/configuration/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/configuration/tasks/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/configuration/tasks/git.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/configuration/tasks/main.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/configuration/tasks/netbox.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/configuration/tasks/update.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/configuration/templates/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/configuration/templates/config-git-netbox-private-key.j2\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/configuration/templates/config-git-private-key.j2\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/configuration/templates/config.j2\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/docker_compose/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/docker_compose/README.md\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/docker_compose/defaults/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/docker_compose/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/docker_compose/files/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/docker_compose/files/docker-compose\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/docker_compose/handlers/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/docker_compose/handlers/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/docker_compose/meta/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/docker_compose/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/docker_compose/tasks/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/docker_compose/tasks/install-Debian-family.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/docker_compose/tasks/install-RedHat-family.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/docker_compose/tasks/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/docker_compose/templates/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/docker_compose/templates/docker-compose.service.j2\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/docker_compose/templates/osism.target.j2\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/docker_login/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/docker_login/README.md\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/docker_login/defaults/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/docker_login/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/docker_login/meta/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/docker_login/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/docker_login/tasks/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/docker_login/tasks/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/facts/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/facts/README.md\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/facts/defaults/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/facts/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/facts/meta/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/facts/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/facts/tasks/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/facts/tasks/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/firewall/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/firewall/README.md\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/firewall/defaults/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/firewall/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/firewall/handlers/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/firewall/handlers/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/firewall/meta/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/firewall/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/firewall/tasks/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/firewall/tasks/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/firewall/templates/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/firewall/templates/ufw.conf.j2\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/hostname/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/hostname/README.md\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/hostname/defaults/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/hostname/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/hostname/handlers/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/hostname/handlers/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/hostname/meta/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/hostname/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/hostname/tasks/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/hostname/tasks/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/hostname/templates/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/hostname/templates/config.j2\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/hosts/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/hosts/README.md\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/hosts/defaults/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/hosts/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/hosts/meta/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/hosts/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/hosts/tasks/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/hosts/tasks/main.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/hosts/tasks/type-block.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/hosts/tasks/type-local.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/hosts/tasks/type-template.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/hosts/templates/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/hosts/templates/hosts-block.j2\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/hosts/templates/hosts-template.j2\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/ipmitool/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/ipmitool/README.md\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/ipmitool/defaults/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/ipmitool/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/ipmitool/meta/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/ipmitool/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/ipmitool/tasks/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/ipmitool/tasks/install-Debian-family.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/ipmitool/tasks/install-RedHat-family.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/ipmitool/tasks/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/ipmitool/templates/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/ipmitool/templates/module-load.conf.j2\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/kernel_modules/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/kernel_modules/README.md\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/kernel_modules/defaults/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/kernel_modules/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/kernel_modules/meta/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/kernel_modules/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/kernel_modules/tasks/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/kernel_modules/tasks/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/known_hosts/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/known_hosts/README.md\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/known_hosts/defaults/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/known_hosts/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/known_hosts/meta/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/known_hosts/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/known_hosts/tasks/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/known_hosts/tasks/main.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/known_hosts/tasks/write-scanned.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/known_hosts/tasks/write-static.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/limits/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/limits/README.md\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/limits/defaults/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/limits/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/limits/meta/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/limits/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/limits/tasks/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/limits/tasks/limits.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/limits/tasks/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/lynis/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/lynis/README.md\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/lynis/defaults/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/lynis/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/lynis/meta/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/lynis/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/lynis/tasks/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/lynis/tasks/install-Debian-family.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/lynis/tasks/install-RedHat-family.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/lynis/tasks/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/microcode/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/microcode/README.md\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/microcode/defaults/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/microcode/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/microcode/meta/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/microcode/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/microcode/tasks/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/microcode/tasks/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/motd/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/motd/README.md\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/motd/defaults/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/motd/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/motd/handlers/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/motd/handlers/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/motd/meta/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/motd/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/motd/tasks/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/motd/tasks/configure-Debian-family.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/motd/tasks/configure-RedHat-family.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/motd/tasks/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/motd/templates/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/motd/templates/motd.j2\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/network/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/network/README.md\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/network/defaults/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/network/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/network/handlers/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/network/handlers/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/network/meta/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/network/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/network/tasks/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/network/tasks/cleanup-interfaces.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/network/tasks/cleanup-netplan.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/network/tasks/cleanup-networkd.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/network/tasks/dummy-interfaces.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/network/tasks/interfaces-Debian-family.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/network/tasks/interfaces-RedHat-family.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/network/tasks/main.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/network/tasks/netplan-Debian-family.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/network/tasks/netplan-RedHat-family.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/network/tasks/network-extra-init.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/network/tasks/vxlan-interfaces.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/network/templates/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/network/templates/dummy.netdev.j2\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/network/templates/network-extra-init.service.j2\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/network/templates/network-extra-init.sh.j2\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/network/templates/vxlan.netdev.j2\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/network/templates/vxlan.network.j2\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/network/templates/interfaces/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/network/templates/interfaces/device.j2\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/network/templates/interfaces/interfaces.j2\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/network/templates/netplan/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/network/templates/netplan/01-osism.yaml.j2\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/network/templates/netplan/interfaces.j2\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/network/vars/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/network/vars/Debian-family.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/network/vars/RedHat-family.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/network/vars/interfaces.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/operator/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/operator/README.md\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/operator/defaults/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/operator/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/operator/meta/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/operator/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/operator/tasks/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/operator/tasks/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/operator/templates/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/operator/templates/operator-sudoers.j2\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/operator/vars/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/operator/vars/Debian-family.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/operator/vars/RedHat-family.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/packages/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/packages/README.md\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/packages/defaults/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/packages/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/packages/meta/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/packages/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/packages/tasks/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/packages/tasks/main.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/packages/tasks/package-Debian-family.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/packages/tasks/package-RedHat-family.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/packages/vars/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/packages/vars/Debian-family.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/packages/vars/RedHat-family.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/podman/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/podman/README.md\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/podman/defaults/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/podman/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/podman/meta/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/podman/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/podman/tasks/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/podman/tasks/config.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/podman/tasks/deploy.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/podman/tasks/install-Debian-family.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/podman/tasks/install-RedHat-family.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/podman/tasks/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/proxy/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/proxy/README.md\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/proxy/defaults/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/proxy/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/proxy/meta/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/proxy/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/proxy/tasks/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/proxy/tasks/Debian-family.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/proxy/tasks/RedHat-family.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/proxy/tasks/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/proxy/templates/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/proxy/templates/proxy-apt.j2\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/repository/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/repository/README.md\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/repository/defaults/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/repository/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/repository/handlers/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/repository/handlers/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/repository/meta/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/repository/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/repository/tasks/\n.L..t...... src/github.com/osism/ansible-collection-commons/roles/repository/tasks/AlmaLinux.yml -> RedHat.yml\n.L..t...... src/github.com/osism/ansible-collection-commons/roles/repository/tasks/CentOS.yml -> RedHat.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/repository/tasks/Debian.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/repository/tasks/RedHat.yml\n.L..t...... src/github.com/osism/ansible-collection-commons/roles/repository/tasks/Rocky.yml -> RedHat.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/repository/tasks/Ubuntu.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/repository/tasks/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/repository/templates/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/repository/templates/99osism.j2\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/repository/templates/source.list.j2\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/repository/templates/ubuntu.sources.j2\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/repository/vars/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/repository/vars/AlmaLinux-amd64.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/repository/vars/CentOS-x86_64.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/repository/vars/Debian-x86_64.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/repository/vars/Rocky-x86_64.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/repository/vars/Ubuntu-aarch64.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/repository/vars/Ubuntu-x86_64.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/resolvconf/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/resolvconf/README.md\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/resolvconf/defaults/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/resolvconf/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/resolvconf/handlers/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/resolvconf/handlers/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/resolvconf/meta/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/resolvconf/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/resolvconf/tasks/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/resolvconf/tasks/configure-Debian-family.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/resolvconf/tasks/configure-RedHat-family.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/resolvconf/tasks/configure-resolv.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/resolvconf/tasks/install-Debian-family.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/resolvconf/tasks/install-RedHat-family.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/resolvconf/tasks/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/resolvconf/templates/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/resolvconf/templates/resolved.conf.j2\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/runc/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/runc/README.md\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/runc/defaults/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/runc/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/runc/meta/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/runc/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/runc/tasks/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/runc/tasks/install-Debian-family.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/runc/tasks/install-RedHat-family.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/runc/tasks/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/services/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/services/README.md\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/services/defaults/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/services/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/services/meta/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/services/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/services/tasks/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/services/tasks/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/sosreport/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/sosreport/README.md\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/sosreport/defaults/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/sosreport/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/sosreport/meta/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/sosreport/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/sosreport/tasks/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/sosreport/tasks/install-Debian-family.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/sosreport/tasks/install-RedHat-family.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/sosreport/tasks/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/sosreport/vars/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/sosreport/vars/CentOS-dist.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/sosreport/vars/Debian-dist.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/sosreport/vars/Ubuntu-dist.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/sshconfig/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/sshconfig/README.md\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/sshconfig/defaults/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/sshconfig/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/sshconfig/meta/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/sshconfig/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/sshconfig/tasks/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/sshconfig/tasks/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/sshconfig/templates/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/sshconfig/templates/config.j2\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/state/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/state/README.md\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/state/defaults/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/state/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/state/meta/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/state/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/state/tasks/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/state/tasks/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/sysctl/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/sysctl/README.md\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/sysctl/defaults/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/sysctl/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/sysctl/meta/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/sysctl/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/sysctl/tasks/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/sysctl/tasks/main.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/sysctl/tasks/sysctl.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/systohc/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/systohc/README.md\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/systohc/defaults/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/systohc/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/systohc/meta/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/systohc/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/systohc/tasks/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/systohc/tasks/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/timezone/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/timezone/README.md\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/timezone/defaults/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/timezone/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/timezone/meta/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/timezone/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/timezone/tasks/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/timezone/tasks/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/trivy/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/trivy/README.md\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/trivy/defaults/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/trivy/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/trivy/meta/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/trivy/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/trivy/tasks/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/trivy/tasks/install-Debian-family.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/trivy/tasks/install-RedHat-family.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/trivy/tasks/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/user/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/user/README.md\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/user/defaults/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/user/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/user/meta/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/user/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/user/tasks/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/user/tasks/main.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/user/tasks/remote-key.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/user/tasks/type-default.yml\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/user/tasks/type-keyfile.yml\n.d..t...... src/github.com/osism/ansible-collection-commons/roles/user/templates/\n<f..t...... src/github.com/osism/ansible-collection-commons/roles/user/templates/config.j2\n.d..t...... src/github.com/osism/ansible-collection-services/\n<f..t...... src/github.com/osism/ansible-collection-services/.ansible-lint\n<f..t...... src/github.com/osism/ansible-collection-services/.flake8\n<f..t...... src/github.com/osism/ansible-collection-services/.gitignore\n<f..t...... src/github.com/osism/ansible-collection-services/.yamllint.yml\n<f..t...... src/github.com/osism/ansible-collection-services/.zuul.yaml\n<f..t...... src/github.com/osism/ansible-collection-services/LICENSE\n<f..t...... src/github.com/osism/ansible-collection-services/README.md\n<f..t...... src/github.com/osism/ansible-collection-services/galaxy.yml\n.d..t...... src/github.com/osism/ansible-collection-services/.git/\n<f..t...... src/github.com/osism/ansible-collection-services/.git/HEAD\n<f..t...... src/github.com/osism/ansible-collection-services/.git/ORIG_HEAD\n<f..t...... src/github.com/osism/ansible-collection-services/.git/config\n<f..t...... src/github.com/osism/ansible-collection-services/.git/config.worktree\n<f..t...... src/github.com/osism/ansible-collection-services/.git/description\n<f..t...... src/github.com/osism/ansible-collection-services/.git/index\n<f..t...... src/github.com/osism/ansible-collection-services/.git/packed-refs\n.d..t...... src/github.com/osism/ansible-collection-services/.git/branches/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/hooks/\n<f..t...... src/github.com/osism/ansible-collection-services/.git/hooks/applypatch-msg.sample\n<f..t...... src/github.com/osism/ansible-collection-services/.git/hooks/commit-msg.sample\n<f..t...... src/github.com/osism/ansible-collection-services/.git/hooks/fsmonitor-watchman.sample\n<f..t...... src/github.com/osism/ansible-collection-services/.git/hooks/post-update.sample\n<f..t...... src/github.com/osism/ansible-collection-services/.git/hooks/pre-applypatch.sample\n<f..t...... src/github.com/osism/ansible-collection-services/.git/hooks/pre-commit.sample\n<f..t...... src/github.com/osism/ansible-collection-services/.git/hooks/pre-merge-commit.sample\n<f..t...... src/github.com/osism/ansible-collection-services/.git/hooks/pre-push.sample\n<f..t...... src/github.com/osism/ansible-collection-services/.git/hooks/pre-rebase.sample\n<f..t...... src/github.com/osism/ansible-collection-services/.git/hooks/pre-receive.sample\n<f..t...... src/github.com/osism/ansible-collection-services/.git/hooks/prepare-commit-msg.sample\n<f..t...... src/github.com/osism/ansible-collection-services/.git/hooks/push-to-checkout.sample\n<f..t...... src/github.com/osism/ansible-collection-services/.git/hooks/update.sample\n.d..t...... src/github.com/osism/ansible-collection-services/.git/info/\n<f..t...... src/github.com/osism/ansible-collection-services/.git/info/exclude\n.d..t...... src/github.com/osism/ansible-collection-services/.git/logs/\n<f..t...... src/github.com/osism/ansible-collection-services/.git/logs/HEAD\n.d..t...... src/github.com/osism/ansible-collection-services/.git/logs/refs/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/logs/refs/heads/\n<f..t...... src/github.com/osism/ansible-collection-services/.git/logs/refs/heads/main\n.d..t...... src/github.com/osism/ansible-collection-services/.git/logs/refs/remotes/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/logs/refs/remotes/origin/\n<f..t...... src/github.com/osism/ansible-collection-services/.git/logs/refs/remotes/origin/HEAD\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/\ncd+++++++++ src/github.com/osism/ansible-collection-services/.git/objects/07/\n<f+++++++++ src/github.com/osism/ansible-collection-services/.git/objects/07/abeb3fe45e34bb6147f840abb924c456658985\ncd+++++++++ src/github.com/osism/ansible-collection-services/.git/objects/0a/\n<f+++++++++ src/github.com/osism/ansible-collection-services/.git/objects/0a/d6892495184ac5a0905065ee7b8f59b58e11ed\ncd+++++++++ src/github.com/osism/ansible-collection-services/.git/objects/0b/\n<f+++++++++ src/github.com/osism/ansible-collection-services/.git/objects/0b/cc82939de812d421ce7dbd024b2530ddaf7373\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/0c/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/10/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/1a/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/20/\ncd+++++++++ src/github.com/osism/ansible-collection-services/.git/objects/21/\n<f+++++++++ src/github.com/osism/ansible-collection-services/.git/objects/21/cf74b32ae0d8cff16a471330421627a77f059d\ncd+++++++++ src/github.com/osism/ansible-collection-services/.git/objects/23/\n<f+++++++++ src/github.com/osism/ansible-collection-services/.git/objects/23/f0e755006ad195b7a2720e2160a4596cace0a0\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/29/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/2b/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/2d/\ncd+++++++++ src/github.com/osism/ansible-collection-services/.git/objects/32/\n<f+++++++++ src/github.com/osism/ansible-collection-services/.git/objects/32/5879d45377e2eb83e8e826a04168d8a18585c7\n<f+++++++++ src/github.com/osism/ansible-collection-services/.git/objects/32/e3c0272e42e39c1b5073a3e99c0da8a2ddf038\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/33/\ncd+++++++++ src/github.com/osism/ansible-collection-services/.git/objects/4a/\n<f+++++++++ src/github.com/osism/ansible-collection-services/.git/objects/4a/0f26b3f2bb9570732719dca581d86c90dd7728\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/4f/\ncd+++++++++ src/github.com/osism/ansible-collection-services/.git/objects/50/\n<f+++++++++ src/github.com/osism/ansible-collection-services/.git/objects/50/4e662b4528f0145dd717c4f9c61aacdd466d2f\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/51/\ncd+++++++++ src/github.com/osism/ansible-collection-services/.git/objects/55/\n<f+++++++++ src/github.com/osism/ansible-collection-services/.git/objects/55/a2f430ce0c434d2c0c118608248b8e1814780a\ncd+++++++++ src/github.com/osism/ansible-collection-services/.git/objects/59/\n<f+++++++++ src/github.com/osism/ansible-collection-services/.git/objects/59/a248cb3e90c79960e8e660f5409daec7d03aca\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/66/\ncd+++++++++ src/github.com/osism/ansible-collection-services/.git/objects/6c/\n<f+++++++++ src/github.com/osism/ansible-collection-services/.git/objects/6c/81ed60282f5fbaefb6b19bed89b8ff9dde0479\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/72/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/7b/\ncd+++++++++ src/github.com/osism/ansible-collection-services/.git/objects/82/\n<f+++++++++ src/github.com/osism/ansible-collection-services/.git/objects/82/b8eb8b6a299a32fbe9d73f4997a28d7b6f9115\ncd+++++++++ src/github.com/osism/ansible-collection-services/.git/objects/85/\n<f+++++++++ src/github.com/osism/ansible-collection-services/.git/objects/85/b6ddd5e7addd50a6f504aa01af44f4956e23db\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/87/\ncd+++++++++ src/github.com/osism/ansible-collection-services/.git/objects/8b/\n<f+++++++++ src/github.com/osism/ansible-collection-services/.git/objects/8b/e9c887dd0903883b40bebfe6f91660cbd657a5\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/95/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/b2/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/d0/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/d7/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/dc/\ncd+++++++++ src/github.com/osism/ansible-collection-services/.git/objects/dd/\n<f+++++++++ src/github.com/osism/ansible-collection-services/.git/objects/dd/601554d21714fa353d94d36c26bc38ed71830b\ncd+++++++++ src/github.com/osism/ansible-collection-services/.git/objects/e0/\n<f+++++++++ src/github.com/osism/ansible-collection-services/.git/objects/e0/93ad7eaa58f0394d242749dbf97e648bd24afe\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/e5/\ncd+++++++++ src/github.com/osism/ansible-collection-services/.git/objects/e9/\n<f+++++++++ src/github.com/osism/ansible-collection-services/.git/objects/e9/159510cf7ca0393b8c2cd8ea649f1fb518fc26\ncd+++++++++ src/github.com/osism/ansible-collection-services/.git/objects/ea/\n<f+++++++++ src/github.com/osism/ansible-collection-services/.git/objects/ea/16ee541431840f00f27f68248afc3b2f69a869\ncd+++++++++ src/github.com/osism/ansible-collection-services/.git/objects/eb/\n<f+++++++++ src/github.com/osism/ansible-collection-services/.git/objects/eb/817d89ecab11ed060682c7912e3edf8bdb641e\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/info/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/objects/pack/\n.d..t...... src/github.com/osism/ansible-collection-services/.git/refs/\n.d..t...... src/github.com/osism/ansible-collection-services/.github/\n<f..t...... src/github.com/osism/ansible-collection-services/.github/publish-collection.yml\n<f..t...... src/github.com/osism/ansible-collection-services/.github/renovate.json\n.d..t...... src/github.com/osism/ansible-collection-services/.github/templates/\n<f..t...... src/github.com/osism/ansible-collection-services/.github/templates/galaxy.yml.j2\n.d..t...... src/github.com/osism/ansible-collection-services/.github/workflows/\n<f..t...... src/github.com/osism/ansible-collection-services/.github/workflows/publish-collection.yml\n.d..t...... src/github.com/osism/ansible-collection-services/meta/\n<f..t...... src/github.com/osism/ansible-collection-services/meta/runtime.yml\n.d..t...... src/github.com/osism/ansible-collection-services/molecule/\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/requirements.txt\n.d..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/collections.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/converge.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/molecule.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify.yml\n.d..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/adminer.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/auditd.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/cephclient.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/cgit.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/chrony.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/clamav.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/containerd.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/dnsdist.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/dnsmasq.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/docker.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/fail2ban.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/falco.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/frr.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/gnmic.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/hddtemp.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/homer.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/httpd.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/journald.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/lldpd.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/manager.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/netbird.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/netbox.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/netdata.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/nexus.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/openstackclient.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/opentelemetry_collector.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/osquery.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/phpmyadmin.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/rng.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/rsyslog.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/scaphandre.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/smartd.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/squid.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/sshd.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/stepca.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/substation.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/teleport.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/thanos_sidecar.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/traefik.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/tuned.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/wazuh_agent.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/wireguard.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/zabbix_agent.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/zuul.yml\n.d..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/__init__.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/adminer.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/auditd.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/cgit.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/dnsdist.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/dnsmasq.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/fail2ban.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/frr.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/gnmic.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/homer.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/httpd.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/journald.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/lldpd.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/manager.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/netbox.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/nexus.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/opentelemetry_collector.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/phpmyadmin.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/rng.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/rsyslog.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/scaphandre.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/smartd.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/squid.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/sshd.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/stepca.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/substation.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/thanos_sidecar.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/traefik.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/tuned.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/wireguard.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/zuul.py\n.d..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/cephclient/\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/cephclient/__init__.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/cephclient/container.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/cephclient/main.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/cephclient/package.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/cephclient/package_debian.py\n.d..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/chrony/\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/chrony/__init__.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/chrony/debian.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/chrony/main.py\n.d..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/clamav/\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/clamav/__init__.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/clamav/debian.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/clamav/main.py\n.d..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/containerd/\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/containerd/__init__.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/containerd/debian.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/containerd/main.py\n.d..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/docker/\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/docker/__init__.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/docker/debian.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/docker/main.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/docker/redhat.py\n.d..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/falco/\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/falco/__init__.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/falco/debian.py\n.d..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/hddtemp/\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/hddtemp/__init__.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/hddtemp/debian.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/hddtemp/redhat.py\n.d..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/netbird/\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/netbird/__init__.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/netbird/debian.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/netbird/redhat.py\n.d..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/netdata/\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/netdata/__init__.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/netdata/debian.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/netdata/main.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/netdata/redhat.py\n.d..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/openstackclient/\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/openstackclient/__init__.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/openstackclient/container.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/openstackclient/main.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/openstackclient/package-debian.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/openstackclient/package-redhat.py\n.d..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/osquery/\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/osquery/__init__.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/osquery/debian.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/osquery/main.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/osquery/redhat.py\n.d..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/teleport/\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/teleport/__init__.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/teleport/debian.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/teleport/main.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/teleport/redhat.py\n.d..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/util/\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/util/__init__.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/util/util.py\n.d..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/wazuh_agent/\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/wazuh_agent/__init__.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/wazuh_agent/debian.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/wazuh_agent/main.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/wazuh_agent/redhat.py\n.d..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/zabbix_agent/\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/zabbix_agent/__init__.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/zabbix_agent/debian.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/zabbix_agent/main.py\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/zabbix_agent/redhat.py\n.d..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/adminer.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/auditd.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/cephclient.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/cgit.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/chrony.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/clamav.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/containerd.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/dnsdist.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/dnsmasq.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/docker.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/fail2ban.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/falco.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/frr.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/gnmic.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/hddtemp.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/homer.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/httpd.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/journald.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/lldpd.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/manager.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/netbird.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/netbox.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/netdata.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/nexus.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/openstackclient.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/opentelemetry_collector.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/osquery.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/phpmyadmin.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/rng.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/rsyslog.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/scaphandre.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/smartd.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/squid.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/sshd.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/stepca.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/substation.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/teleport.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/thanos_sidecar.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/traefik.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/tuned.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/wazuh_agent.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/wireguard.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/zabbix_agent.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/zuul.yml\n.d..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/adminer.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/auditd.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/cephclient.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/cgit.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/chrony.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/clamav.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/containerd.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/dnsdist.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/dnsmasq.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/docker.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/fail2ban.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/falco.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/frr.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/gnmic.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/hddtemp.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/homer.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/httpd.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/journald.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/lldpd.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/netbird.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/netdata.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/osquery.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/rng.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/rsyslog.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/smartd.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/sshd.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/teleport.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/tuned.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/wazuh_agent.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/wireguard.yml\n<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/zabbix_agent.yml\n.d..t...... src/github.com/osism/ansible-collection-services/plugins/\n.d..t...... src/github.com/osism/ansible-collection-services/plugins/filter/\n<f..t...... src/github.com/osism/ansible-collection-services/plugins/filter/address.py\n.d..t...... src/github.com/osism/ansible-collection-services/plugins/modules/\n<f..t...... src/github.com/osism/ansible-collection-services/plugins/modules/kolla_container_facts.py\n<f..t...... src/github.com/osism/ansible-collection-services/plugins/modules/kolla_docker.py\n<f..t...... src/github.com/osism/ansible-collection-services/plugins/modules/kolla_toolbox.py\n.d..t...... src/github.com/osism/ansible-collection-services/roles/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/README.md\n.d..t...... src/github.com/osism/ansible-collection-services/roles/adminer/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/adminer/README.md\n.d..t...... src/github.com/osism/ansible-collection-services/roles/adminer/defaults/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/adminer/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/adminer/handlers/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/adminer/handlers/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/adminer/meta/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/adminer/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/adminer/tasks/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/adminer/tasks/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/adminer/templates/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/adminer/templates/docker-compose.yml.j2\n.d..t...... src/github.com/osism/ansible-collection-services/roles/auditd/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/auditd/README.md\n.d..t...... src/github.com/osism/ansible-collection-services/roles/auditd/defaults/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/auditd/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/auditd/handlers/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/auditd/handlers/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/auditd/meta/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/auditd/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/auditd/tasks/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/auditd/tasks/install-Debian-family.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/auditd/tasks/install-RedHat-family.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/auditd/tasks/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/auditd/templates/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/auditd/templates/auditd.conf.j2\n.d..t...... src/github.com/osism/ansible-collection-services/roles/auditd/templates/rules/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/auditd/templates/rules/20-neo23x0.rules.j2\n.d..t...... src/github.com/osism/ansible-collection-services/roles/cephclient/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/cephclient/README.md\n.d..t...... src/github.com/osism/ansible-collection-services/roles/cephclient/defaults/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/cephclient/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/cephclient/handlers/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/cephclient/handlers/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/cephclient/meta/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/cephclient/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/cephclient/tasks/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/cephclient/tasks/clean-container.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/cephclient/tasks/clean-package-Debian-family.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/cephclient/tasks/clean-package-RedHat-family.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/cephclient/tasks/container.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/cephclient/tasks/main.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/cephclient/tasks/package-Debian-family.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/cephclient/tasks/package-RedHat-family.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/cephclient/tasks/rook.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/cephclient/templates/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/cephclient/templates/ceph.conf.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/cephclient/templates/docker-compose.yml.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/cephclient/templates/keyring.j2\n.d..t...... src/github.com/osism/ansible-collection-services/roles/cephclient/templates/wrapper/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/cephclient/templates/wrapper/ceph-authtool.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/cephclient/templates/wrapper/ceph.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/cephclient/templates/wrapper/crushtool.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/cephclient/templates/wrapper/rados.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/cephclient/templates/wrapper/radosgw-admin.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/cephclient/templates/wrapper/rbd.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/cephclient/templates/wrapper/rook.j2\n.d..t...... src/github.com/osism/ansible-collection-services/roles/cgit/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/cgit/README.md\n.d..t...... src/github.com/osism/ansible-collection-services/roles/cgit/defaults/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/cgit/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/cgit/handlers/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/cgit/handlers/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/cgit/meta/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/cgit/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/cgit/tasks/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/cgit/tasks/config.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/cgit/tasks/main.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/cgit/tasks/service.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/cgit/templates/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/cgit/templates/cgitrc.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/cgit/templates/docker-compose.yml.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/cgit/templates/httpd-cgit.conf.j2\n.d..t...... src/github.com/osism/ansible-collection-services/roles/chrony/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/chrony/README.md\n.d..t...... src/github.com/osism/ansible-collection-services/roles/chrony/defaults/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/chrony/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/chrony/handlers/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/chrony/handlers/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/chrony/meta/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/chrony/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/chrony/tasks/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/chrony/tasks/install-Debian-family.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/chrony/tasks/install-RedHat-family.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/chrony/tasks/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/chrony/templates/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/chrony/templates/chrony.conf.j2\n.d..t...... src/github.com/osism/ansible-collection-services/roles/chrony/vars/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/chrony/vars/Debian-family.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/chrony/vars/RedHat-family.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/clamav/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/clamav/README.md\n.d..t...... src/github.com/osism/ansible-collection-services/roles/clamav/defaults/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/clamav/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/clamav/handlers/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/clamav/handlers/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/clamav/meta/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/clamav/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/clamav/tasks/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/clamav/tasks/initialize.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/clamav/tasks/install-Debian-family.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/clamav/tasks/install-RedHat-family.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/clamav/tasks/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/clamav/vars/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/clamav/vars/Debian-family.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/clamav/vars/RedHat-family.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/containerd/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/containerd/README.md\n.d..t...... src/github.com/osism/ansible-collection-services/roles/containerd/defaults/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/containerd/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/containerd/handlers/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/containerd/handlers/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/containerd/meta/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/containerd/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/containerd/tasks/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/containerd/tasks/config.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/containerd/tasks/install-Debian-family.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/containerd/tasks/install-RedHat-family.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/containerd/tasks/main.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/containerd/tasks/service.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/containerd/templates/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/containerd/templates/config.toml.j2\n.d..t...... src/github.com/osism/ansible-collection-services/roles/containerd/vars/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/containerd/vars/CentOS-dist.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/containerd/vars/Debian-dist.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/containerd/vars/Ubuntu-dist.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/dnsdist/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/dnsdist/README.md\n.d..t...... src/github.com/osism/ansible-collection-services/roles/dnsdist/defaults/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/dnsdist/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/dnsdist/handlers/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/dnsdist/handlers/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/dnsdist/meta/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/dnsdist/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/dnsdist/tasks/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/dnsdist/tasks/config.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/dnsdist/tasks/main.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/dnsdist/tasks/service.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/dnsdist/templates/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/dnsdist/templates/dnsdist.conf.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/dnsdist/templates/docker-compose.yml.j2\n.d..t...... src/github.com/osism/ansible-collection-services/roles/dnsmasq/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/dnsmasq/README.md\n.d..t...... src/github.com/osism/ansible-collection-services/roles/dnsmasq/defaults/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/dnsmasq/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/dnsmasq/handlers/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/dnsmasq/handlers/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/dnsmasq/meta/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/dnsmasq/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/dnsmasq/tasks/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/dnsmasq/tasks/config.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/dnsmasq/tasks/main.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/dnsmasq/tasks/service.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/dnsmasq/templates/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/dnsmasq/templates/dnsmasq.conf.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/dnsmasq/templates/docker-compose.yml.j2\n.d..t...... src/github.com/osism/ansible-collection-services/roles/docker/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/docker/README.md\n.d..t...... src/github.com/osism/ansible-collection-services/roles/docker/defaults/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/docker/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/docker/files/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/docker/files/.gitattributes\n<f..t...... src/github.com/osism/ansible-collection-services/roles/docker/files/limits.conf\n.d..t...... src/github.com/osism/ansible-collection-services/roles/docker/files/facts/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/docker/files/facts/docker_containers.fact\n<f..t...... src/github.com/osism/ansible-collection-services/roles/docker/files/facts/docker_images.fact\n.d..t...... src/github.com/osism/ansible-collection-services/roles/docker/handlers/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/docker/handlers/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/docker/meta/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/docker/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/docker/tasks/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/docker/tasks/bootstrap.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/docker/tasks/config.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/docker/tasks/facts.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/docker/tasks/install-docker-Debian-family.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/docker/tasks/install-docker-RedHat-family.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/docker/tasks/main.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/docker/tasks/service.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/docker/tasks/storage-block.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/docker/tasks/storage-zram.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/docker/templates/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/docker/templates/config.json.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/docker/templates/containerd_config.toml.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/docker/templates/daemon.json.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/docker/templates/overlay.conf.j2\n.d..t...... src/github.com/osism/ansible-collection-services/roles/docker/vars/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/docker/vars/Debian-family.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/docker/vars/RedHat-family.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/fail2ban/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/fail2ban/README.md\n.d..t...... src/github.com/osism/ansible-collection-services/roles/fail2ban/defaults/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/fail2ban/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/fail2ban/meta/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/fail2ban/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/fail2ban/tasks/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/fail2ban/tasks/install-Debian-family.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/fail2ban/tasks/install-RedHat-family.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/fail2ban/tasks/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/fail2ban/templates/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/fail2ban/templates/sshd.conf.j2\n.d..t...... src/github.com/osism/ansible-collection-services/roles/falco/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/falco/.gitignore\n<f..t...... src/github.com/osism/ansible-collection-services/roles/falco/README.md\n.d..t...... src/github.com/osism/ansible-collection-services/roles/falco/defaults/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/falco/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/falco/files/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/falco/files/falco.service\n.d..t...... src/github.com/osism/ansible-collection-services/roles/falco/handlers/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/falco/handlers/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/falco/meta/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/falco/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/falco/tasks/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/falco/tasks/install-Debian-family.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/falco/tasks/install-RedHat-family.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/falco/tasks/install-RedHat-family_broken.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/falco/tasks/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/frr/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/frr/README.md\n.d..t...... src/github.com/osism/ansible-collection-services/roles/frr/defaults/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/frr/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/frr/handlers/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/frr/handlers/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/frr/meta/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/frr/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/frr/tasks/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/frr/tasks/install-Debian-family.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/frr/tasks/install-RedHat-family.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/frr/tasks/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/frr/templates/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/frr/templates/daemons.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/frr/templates/frr_k3s_cilium.conf.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/frr/templates/frr_leaf.conf.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/frr/templates/frr_loadbalancer.conf.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/frr/templates/frr_loadbalancer_external_uplink.conf.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/frr/templates/frr_test.conf.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/frr/templates/vtysh.conf.j2\n.d..t...... src/github.com/osism/ansible-collection-services/roles/gnmic/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/gnmic/README.md\n.d..t...... src/github.com/osism/ansible-collection-services/roles/gnmic/defaults/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/gnmic/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/gnmic/handlers/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/gnmic/handlers/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/gnmic/meta/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/gnmic/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/gnmic/tasks/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/gnmic/tasks/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/gnmic/templates/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/gnmic/templates/docker-compose.yml.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/gnmic/templates/gnmic.yaml.j2\n.d..t...... src/github.com/osism/ansible-collection-services/roles/hddtemp/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/hddtemp/README.md\n.d..t...... src/github.com/osism/ansible-collection-services/roles/hddtemp/defaults/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/hddtemp/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/hddtemp/files/\n.d..t...... src/github.com/osism/ansible-collection-services/roles/hddtemp/files/RedHat-family/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/hddtemp/files/RedHat-family/hddtemp\n.d..t...... src/github.com/osism/ansible-collection-services/roles/hddtemp/handlers/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/hddtemp/handlers/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/hddtemp/meta/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/hddtemp/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/hddtemp/tasks/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/hddtemp/tasks/install-Debian-family.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/hddtemp/tasks/install-RedHat-family.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/hddtemp/tasks/main.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/hddtemp/tasks/service-Debian-family.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/hddtemp/tasks/service-RedHat-family.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/hddtemp/vars/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/hddtemp/vars/Debian-family.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/hddtemp/vars/RedHat-family.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/homer/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/homer/README.md\n.d..t...... src/github.com/osism/ansible-collection-services/roles/homer/defaults/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/homer/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/homer/handlers/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/homer/handlers/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/homer/meta/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/homer/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/homer/tasks/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/homer/tasks/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/homer/templates/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/homer/templates/config.yml.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/homer/templates/docker-compose.yml.j2\n.d..t...... src/github.com/osism/ansible-collection-services/roles/httpd/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/httpd/README.md\n.d..t...... src/github.com/osism/ansible-collection-services/roles/httpd/defaults/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/httpd/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/httpd/files/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/httpd/files/prepare-ironic-volume.sh\n.d..t...... src/github.com/osism/ansible-collection-services/roles/httpd/handlers/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/httpd/handlers/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/httpd/meta/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/httpd/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/httpd/tasks/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/httpd/tasks/ironic.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/httpd/tasks/main.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/httpd/tasks/sonic-ztp.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/httpd/templates/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/httpd/templates/docker-compose.yml.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/httpd/templates/htaccess.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/httpd/templates/httpd.conf.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/httpd/templates/post_install.sh.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/httpd/templates/ztp.json.j2\n.d..t...... src/github.com/osism/ansible-collection-services/roles/journald/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/journald/README.md\n.d..t...... src/github.com/osism/ansible-collection-services/roles/journald/defaults/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/journald/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/journald/handlers/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/journald/handlers/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/journald/meta/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/journald/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/journald/tasks/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/journald/tasks/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/journald/templates/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/journald/templates/journald.conf.j2\n.d..t...... src/github.com/osism/ansible-collection-services/roles/kepler/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/kepler/README.md\n.d..t...... src/github.com/osism/ansible-collection-services/roles/kepler/defaults/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/kepler/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/kepler/handlers/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/kepler/handlers/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/kepler/meta/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/kepler/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/kepler/tasks/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/kepler/tasks/config.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/kepler/tasks/main.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/kepler/tasks/service.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/kepler/templates/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/kepler/templates/config.yaml.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/kepler/templates/docker-compose.yml.j2\n.d..t...... src/github.com/osism/ansible-collection-services/roles/lldpd/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/lldpd/README.md\n.d..t...... src/github.com/osism/ansible-collection-services/roles/lldpd/defaults/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/lldpd/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/lldpd/handlers/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/lldpd/handlers/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/lldpd/meta/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/lldpd/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/lldpd/tasks/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/lldpd/tasks/install-Debian-family.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/lldpd/tasks/install-RedHat-family.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/lldpd/tasks/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/manager/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/README.md\n.d..t...... src/github.com/osism/ansible-collection-services/roles/manager/defaults/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/manager/handlers/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/handlers/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/manager/meta/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/manager/tasks/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/tasks/config-ansible.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/tasks/config-ara.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/tasks/config-celery.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/tasks/config-frontend.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/tasks/config-netbox.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/tasks/config-scripts.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/tasks/config-vault.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/tasks/config-wrapper.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/tasks/config.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/tasks/initialize-vault.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/tasks/initialize.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/tasks/install-Debian-family.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/tasks/install-RedHat-family.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/tasks/main.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/tasks/service.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/tasks/verify-versions.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/check-service-manager-versions.sh.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/config-netbox-secret.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/config-ssh-key.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/config-vault-key.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/config.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/docker-compose.yml.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/manager.service.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/vault.hcl.j2\n.d..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/env/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/env/all.env.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/env/ansible.env.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/env/ara-server.env.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/env/ara.env.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/env/client.env.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/env/conductor.env.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/env/frontend.env.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/env/inventory-reconciler.env.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/env/listener.env.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/env/mariadb.env.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/env/netbox.env.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/env/openstack.env.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/env/postgres.env.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/env/vault.env.j2\n.d..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/scripts/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/scripts/osism-include.j2\n.d..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/wrapper/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/wrapper/cilium.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/wrapper/flux.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/wrapper/hubble.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/wrapper/osism-ansible.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/wrapper/osism-ceph.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/wrapper/osism-custom.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/wrapper/osism-generic.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/wrapper/osism-infrastructure.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/wrapper/osism-kolla.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/wrapper/osism-manager.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/wrapper/osism-monitoring.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/wrapper/osism-openstack.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/wrapper/osism-run-without-secrets.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/wrapper/osism-run.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/wrapper/osism-state.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/wrapper/osism-update-docker.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/wrapper/osism-update-manager.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/wrapper/osism-validate.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/wrapper/osism.j2\n.d..t...... src/github.com/osism/ansible-collection-services/roles/manager/vars/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/vars/Debian-family.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/vars/RedHat-family.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/vars/scripts.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/vars/wrapper.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/netbird/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/netbird/README.md\n.d..t...... src/github.com/osism/ansible-collection-services/roles/netbird/defaults/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/netbird/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/netbird/handlers/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/netbird/handlers/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/netbird/meta/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/netbird/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/netbird/tasks/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/netbird/tasks/install-Debian-family.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/netbird/tasks/install-RedHat-family.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/netbird/tasks/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/netbird/vars/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/netbird/vars/Debian-family.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/netbird/vars/RedHat-family.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/netbox/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/README.md\n.d..t...... src/github.com/osism/ansible-collection-services/roles/netbox/defaults/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/netbox/files/\n.d..t...... src/github.com/osism/ansible-collection-services/roles/netbox/files/startup-scripts/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/files/startup-scripts/270_tags.py\n.d..t...... src/github.com/osism/ansible-collection-services/roles/netbox/handlers/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/handlers/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/netbox/meta/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/netbox/tasks/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/tasks/config-netbox.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/tasks/config-postgres.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/tasks/config.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/tasks/install-Debian-family.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/tasks/install-RedHat-family.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/tasks/main.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/tasks/rescue-service.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/tasks/restart-service.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/tasks/service.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/tasks/wait-for-healthy-service.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/netbox/templates/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/templates/config-netbox-database.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/templates/config-netbox-secret.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/templates/config-postgres-secret.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/templates/configuration.py.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/templates/docker-compose.yml.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/templates/netbox.service.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/templates/nginx-unit-1.26.json.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/templates/nginx-unit.json.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/templates/postgresql.conf.j2\n.d..t...... src/github.com/osism/ansible-collection-services/roles/netbox/templates/env/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/templates/env/netbox.env.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/templates/env/postgres.env.j2\n.d..t...... src/github.com/osism/ansible-collection-services/roles/netbox/templates/initializers/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/templates/initializers/custom_fields.yml.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/templates/initializers/device_roles.yml.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/templates/initializers/device_types.yml.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/templates/initializers/groups.yml.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/templates/initializers/manufacturers.yml.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/templates/initializers/object_permissions.yml.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/templates/initializers/prefix_vlan_roles.yml.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/templates/initializers/sites.yml.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/templates/initializers/tags.yml.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/templates/initializers/users.yml.j2\n.d..t...... src/github.com/osism/ansible-collection-services/roles/netbox/vars/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/vars/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/netdata/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/netdata/README.md\n.d..t...... src/github.com/osism/ansible-collection-services/roles/netdata/defaults/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/netdata/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/netdata/handlers/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/netdata/handlers/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/netdata/meta/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/netdata/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/netdata/tasks/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/netdata/tasks/client.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/netdata/tasks/config.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/netdata/tasks/install-Debian-family.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/netdata/tasks/install-RedHat-family.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/netdata/tasks/main.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/netdata/tasks/server.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/netdata/templates/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/netdata/templates/netdata.conf.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/netdata/templates/stream.conf.j2\n.d..t...... src/github.com/osism/ansible-collection-services/roles/nexus/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/nexus/README.md\n.d..t...... src/github.com/osism/ansible-collection-services/roles/nexus/defaults/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/nexus/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/nexus/files/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/nexus/files/create_repos_from_list.groovy\n<f..t...... src/github.com/osism/ansible-collection-services/roles/nexus/files/setup_http_proxy.groovy\n<f..t...... src/github.com/osism/ansible-collection-services/roles/nexus/files/setup_realms.groovy\n<f..t...... src/github.com/osism/ansible-collection-services/roles/nexus/files/update_admin_password.groovy\n.d..t...... src/github.com/osism/ansible-collection-services/roles/nexus/handlers/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/nexus/handlers/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/nexus/meta/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/nexus/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/nexus/tasks/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/nexus/tasks/call-script.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/nexus/tasks/config.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/nexus/tasks/declare-script.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/nexus/tasks/initialize.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/nexus/tasks/main.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/nexus/tasks/service.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/nexus/templates/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/nexus/templates/docker-compose.yml.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/nexus/templates/nexus.env.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/nexus/templates/nexus.properties.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/nexus/templates/nexus.service.j2\n.d..t...... src/github.com/osism/ansible-collection-services/roles/openstackclient/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/openstackclient/README.md\n.d..t...... src/github.com/osism/ansible-collection-services/roles/openstackclient/defaults/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/openstackclient/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/openstackclient/handlers/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/openstackclient/handlers/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/openstackclient/meta/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/openstackclient/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/openstackclient/tasks/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/openstackclient/tasks/container-Debian-family.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/openstackclient/tasks/container-RedHat-family.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/openstackclient/tasks/main.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/openstackclient/tasks/package-Debian-family.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/openstackclient/tasks/package-RedHat-family.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/openstackclient/templates/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/openstackclient/templates/docker-compose.yml.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/openstackclient/templates/openstack.j2\n.d..t...... src/github.com/osism/ansible-collection-services/roles/opentelemetry_collector/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/opentelemetry_collector/README.md\n.d..t...... src/github.com/osism/ansible-collection-services/roles/opentelemetry_collector/defaults/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/opentelemetry_collector/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/opentelemetry_collector/handlers/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/opentelemetry_collector/handlers/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/opentelemetry_collector/meta/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/opentelemetry_collector/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/opentelemetry_collector/tasks/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/opentelemetry_collector/tasks/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/opentelemetry_collector/templates/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/opentelemetry_collector/templates/docker-compose.yml.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/opentelemetry_collector/templates/otel-collector-config.yml.j2\n.d..t...... src/github.com/osism/ansible-collection-services/roles/osquery/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/osquery/README.md\n.d..t...... src/github.com/osism/ansible-collection-services/roles/osquery/defaults/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/osquery/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/osquery/handlers/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/osquery/handlers/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/osquery/meta/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/osquery/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/osquery/tasks/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/osquery/tasks/install-Debian-family.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/osquery/tasks/install-RedHat-family.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/osquery/tasks/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/phpmyadmin/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/phpmyadmin/README.md\n.d..t...... src/github.com/osism/ansible-collection-services/roles/phpmyadmin/defaults/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/phpmyadmin/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/phpmyadmin/handlers/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/phpmyadmin/handlers/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/phpmyadmin/meta/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/phpmyadmin/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/phpmyadmin/tasks/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/phpmyadmin/tasks/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/phpmyadmin/templates/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/phpmyadmin/templates/docker-compose.yml.j2\n.d..t...... src/github.com/osism/ansible-collection-services/roles/rng/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/rng/README.md\n.d..t...... src/github.com/osism/ansible-collection-services/roles/rng/defaults/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/rng/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/rng/meta/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/rng/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/rng/tasks/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/rng/tasks/install-Debian-family.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/rng/tasks/install-RedHat-family.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/rng/tasks/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/rsyslog/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/rsyslog/README.md\n.d..t...... src/github.com/osism/ansible-collection-services/roles/rsyslog/defaults/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/rsyslog/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/rsyslog/handlers/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/rsyslog/handlers/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/rsyslog/meta/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/rsyslog/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/rsyslog/tasks/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/rsyslog/tasks/additional-log-server.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/rsyslog/tasks/fluentd.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/rsyslog/tasks/install-Debian-family.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/rsyslog/tasks/install-RedHat-family.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/rsyslog/tasks/logrotate.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/rsyslog/tasks/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/rsyslog/templates/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/rsyslog/templates/70-fluentd.conf.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/rsyslog/templates/71-additional-log-server.conf.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/rsyslog/templates/rsyslog-logrotate.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/rsyslog/templates/rsyslog.conf.j2\n.d..t...... src/github.com/osism/ansible-collection-services/roles/rsyslog/vars/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/rsyslog/vars/CentOS-dist.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/rsyslog/vars/Debian-dist.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/rsyslog/vars/Ubuntu-dist.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/scaphandre/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/scaphandre/README.md\n.d..t...... src/github.com/osism/ansible-collection-services/roles/scaphandre/defaults/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/scaphandre/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/scaphandre/handlers/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/scaphandre/handlers/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/scaphandre/meta/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/scaphandre/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/scaphandre/tasks/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/scaphandre/tasks/config.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/scaphandre/tasks/main.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/scaphandre/tasks/service.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/scaphandre/templates/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/scaphandre/templates/docker-compose.yml.j2\n.d..t...... src/github.com/osism/ansible-collection-services/roles/smartd/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/smartd/README.md\n.d..t...... src/github.com/osism/ansible-collection-services/roles/smartd/defaults/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/smartd/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/smartd/files/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/smartd/files/smartmontools\n.d..t...... src/github.com/osism/ansible-collection-services/roles/smartd/handlers/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/smartd/handlers/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/smartd/meta/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/smartd/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/smartd/tasks/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/smartd/tasks/install-Debian-family.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/smartd/tasks/install-RedHat-family.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/smartd/tasks/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/squid/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/squid/README.md\n.d..t...... src/github.com/osism/ansible-collection-services/roles/squid/defaults/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/squid/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/squid/handlers/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/squid/handlers/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/squid/meta/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/squid/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/squid/tasks/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/squid/tasks/install-Debian-family.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/squid/tasks/install-RedHat-family.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/squid/tasks/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/squid/templates/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/squid/templates/docker-compose.yml.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/squid/templates/osism.conf.j2\n.d..t...... src/github.com/osism/ansible-collection-services/roles/sshd/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/sshd/README.md\n.d..t...... src/github.com/osism/ansible-collection-services/roles/sshd/defaults/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/sshd/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/sshd/handlers/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/sshd/handlers/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/sshd/meta/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/sshd/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/sshd/tasks/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/sshd/tasks/install-Debian-family.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/sshd/tasks/install-RedHat-family.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/sshd/tasks/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/stepca/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/stepca/README.md\n.d..t...... src/github.com/osism/ansible-collection-services/roles/stepca/defaults/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/stepca/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/stepca/handlers/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/stepca/handlers/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/stepca/tasks/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/stepca/tasks/config-wrapper.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/stepca/tasks/config.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/stepca/tasks/main.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/stepca/tasks/provisioner-acme.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/stepca/tasks/service.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/stepca/templates/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/stepca/templates/docker-compose.yml.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/stepca/templates/stepca.env.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/stepca/templates/stepca.service.j2\n.d..t...... src/github.com/osism/ansible-collection-services/roles/stepca/templates/wrapper/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/stepca/templates/wrapper/step.j2\n.d..t...... src/github.com/osism/ansible-collection-services/roles/substation/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/substation/README.md\n.d..t...... src/github.com/osism/ansible-collection-services/roles/substation/defaults/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/substation/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/substation/files/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/substation/files/entrypoint.sh\n.d..t...... src/github.com/osism/ansible-collection-services/roles/substation/handlers/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/substation/handlers/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/substation/meta/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/substation/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/substation/tasks/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/substation/tasks/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/substation/templates/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/substation/templates/docker-compose.yml.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/substation/templates/substation.j2\n.d..t...... src/github.com/osism/ansible-collection-services/roles/teleport/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/teleport/README.md\n.d..t...... src/github.com/osism/ansible-collection-services/roles/teleport/defaults/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/teleport/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/teleport/handlers/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/teleport/handlers/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/teleport/meta/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/teleport/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/teleport/tasks/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/teleport/tasks/install-Debian-family.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/teleport/tasks/install-RedHat-family.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/teleport/tasks/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/teleport/templates/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/teleport/templates/teleport.yaml.j2\n.d..t...... src/github.com/osism/ansible-collection-services/roles/thanos_sidecar/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/thanos_sidecar/README.md\n.d..t...... src/github.com/osism/ansible-collection-services/roles/thanos_sidecar/defaults/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/thanos_sidecar/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/thanos_sidecar/handlers/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/thanos_sidecar/handlers/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/thanos_sidecar/meta/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/thanos_sidecar/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/thanos_sidecar/tasks/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/thanos_sidecar/tasks/config.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/thanos_sidecar/tasks/main.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/thanos_sidecar/tasks/service.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/thanos_sidecar/templates/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/thanos_sidecar/templates/docker-compose.yml.j2\n.d..t...... src/github.com/osism/ansible-collection-services/roles/traefik/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/traefik/README.md\n.d..t...... src/github.com/osism/ansible-collection-services/roles/traefik/defaults/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/traefik/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/traefik/handlers/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/traefik/handlers/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/traefik/meta/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/traefik/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/traefik/tasks/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/traefik/tasks/config.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/traefik/tasks/main.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/traefik/tasks/service.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/traefik/templates/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/traefik/templates/certificates.yml.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/traefik/templates/docker-compose.yml.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/traefik/templates/dynamic.yml.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/traefik/templates/traefik-cert.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/traefik/templates/traefik-key.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/traefik/templates/traefik.env.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/traefik/templates/traefik.yml.j2\n.d..t...... src/github.com/osism/ansible-collection-services/roles/tuned/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/tuned/README.md\n.d..t...... src/github.com/osism/ansible-collection-services/roles/tuned/defaults/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/tuned/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/tuned/meta/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/tuned/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/tuned/tasks/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/tuned/tasks/install-Debian-family.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/tuned/tasks/install-RedHat-family.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/tuned/tasks/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/wazuh_agent/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/wazuh_agent/README.md\n.d..t...... src/github.com/osism/ansible-collection-services/roles/wazuh_agent/defaults/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/wazuh_agent/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/wazuh_agent/handlers/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/wazuh_agent/handlers/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/wazuh_agent/meta/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/wazuh_agent/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/wazuh_agent/tasks/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/wazuh_agent/tasks/install-Debian-family.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/wazuh_agent/tasks/install-RedHat-family.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/wazuh_agent/tasks/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/wazuh_agent/templates/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/wazuh_agent/templates/client.keys.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/wazuh_agent/templates/ossec.conf.j2\n.d..t...... src/github.com/osism/ansible-collection-services/roles/wireguard/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/wireguard/README.md\n.d..t...... src/github.com/osism/ansible-collection-services/roles/wireguard/defaults/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/wireguard/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/wireguard/handlers/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/wireguard/handlers/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/wireguard/meta/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/wireguard/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/wireguard/tasks/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/wireguard/tasks/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/wireguard/templates/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/wireguard/templates/client.conf.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/wireguard/templates/server.conf.j2\n.d..t...... src/github.com/osism/ansible-collection-services/roles/zabbix_agent/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/zabbix_agent/README.md\n.d..t...... src/github.com/osism/ansible-collection-services/roles/zabbix_agent/defaults/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/zabbix_agent/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/zabbix_agent/handlers/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/zabbix_agent/handlers/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/zabbix_agent/meta/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/zabbix_agent/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/zabbix_agent/tasks/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/zabbix_agent/tasks/install-Debian-family.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/zabbix_agent/tasks/install-RedHat-family.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/zabbix_agent/tasks/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/zabbix_agent/templates/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/zabbix_agent/templates/zabbix_agent2.conf.j2\n.d..t...... src/github.com/osism/ansible-collection-services/roles/zuul/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/zuul/README.md\n.d..t...... src/github.com/osism/ansible-collection-services/roles/zuul/defaults/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/zuul/defaults/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/zuul/files/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/zuul/files/builder-logging.conf\n<f..t...... src/github.com/osism/ansible-collection-services/roles/zuul/files/executor-logging.conf\n<f..t...... src/github.com/osism/ansible-collection-services/roles/zuul/files/launcher-logging.conf\n<f..t...... src/github.com/osism/ansible-collection-services/roles/zuul/files/logging.conf\n<f..t...... src/github.com/osism/ansible-collection-services/roles/zuul/files/mime.types\n<f..t...... src/github.com/osism/ansible-collection-services/roles/zuul/files/openssl.cnf\n<f..t...... src/github.com/osism/ansible-collection-services/roles/zuul/files/web-logging.conf\n<f..t...... src/github.com/osism/ansible-collection-services/roles/zuul/files/zk-ca.sh\n.d..t...... src/github.com/osism/ansible-collection-services/roles/zuul/handlers/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/zuul/handlers/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/zuul/meta/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/zuul/meta/main.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/zuul/tasks/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/zuul/tasks/copy_tasks.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/zuul/tasks/generate_certs_and_keys.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/zuul/tasks/main.yml\n<f..t...... src/github.com/osism/ansible-collection-services/roles/zuul/tasks/template_tasks.yml\n.d..t...... src/github.com/osism/ansible-collection-services/roles/zuul/templates/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/zuul/templates/docker-compose.yaml.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/zuul/templates/httpd.conf.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/zuul/templates/main.yaml.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/zuul/templates/nodepool.yaml.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/zuul/templates/wait-to-start-gearman.sh.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/zuul/templates/wait-to-start.sh.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/zuul/templates/zoo.cfg.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/zuul/templates/zuul.conf.j2\n.d..t...... src/github.com/osism/ansible-collection-services/roles/zuul/templates/env/\n<f..t...... src/github.com/osism/ansible-collection-services/roles/zuul/templates/env/executor.env.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/zuul/templates/env/mariadb.env.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/zuul/templates/env/nodepool_builder.env.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/zuul/templates/env/nodepool_launcher.env.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/zuul/templates/env/scheduler.env.j2\n<f..t...... src/github.com/osism/ansible-collection-services/roles/zuul/templates/env/web.env.j2\n.d..t...... src/github.com/osism/terraform-base/\n<f..t...... src/github.com/osism/terraform-base/.yamllint.yml\n<f..t...... src/github.com/osism/terraform-base/.zuul.yaml\n<f..t...... src/github.com/osism/terraform-base/LICENSE\n<f..t...... src/github.com/osism/terraform-base/README.md\n.d..t...... src/github.com/osism/terraform-base/.git/\n<f..t...... src/github.com/osism/terraform-base/.git/HEAD\n<f..t...... src/github.com/osism/terraform-base/.git/ORIG_HEAD\n<f..t...... src/github.com/osism/terraform-base/.git/config\n<f..t...... src/github.com/osism/terraform-base/.git/config.worktree\n<f..t...... src/github.com/osism/terraform-base/.git/description\n<f..t...... src/github.com/osism/terraform-base/.git/index\n<f..t...... src/github.com/osism/terraform-base/.git/packed-refs\n.d..t...... src/github.com/osism/terraform-base/.git/branches/\n.d..t...... src/github.com/osism/terraform-base/.git/hooks/\n<f..t...... src/github.com/osism/terraform-base/.git/hooks/applypatch-msg.sample\n<f..t...... src/github.com/osism/terraform-base/.git/hooks/commit-msg.sample\n<f..t...... src/github.com/osism/terraform-base/.git/hooks/fsmonitor-watchman.sample\n<f..t...... src/github.com/osism/terraform-base/.git/hooks/post-update.sample\n<f..t...... src/github.com/osism/terraform-base/.git/hooks/pre-applypatch.sample\n<f..t...... src/github.com/osism/terraform-base/.git/hooks/pre-commit.sample\n<f..t...... src/github.com/osism/terraform-base/.git/hooks/pre-merge-commit.sample\n<f..t...... src/github.com/osism/terraform-base/.git/hooks/pre-push.sample\n<f..t...... src/github.com/osism/terraform-base/.git/hooks/pre-rebase.sample\n<f..t...... src/github.com/osism/terraform-base/.git/hooks/pre-receive.sample\n<f..t...... src/github.com/osism/terraform-base/.git/hooks/prepare-commit-msg.sample\n<f..t...... src/github.com/osism/terraform-base/.git/hooks/push-to-checkout.sample\n<f..t...... src/github.com/osism/terraform-base/.git/hooks/update.sample\n.d..t...... src/github.com/osism/terraform-base/.git/info/\n<f..t...... src/github.com/osism/terraform-base/.git/info/exclude\n.d..t...... src/github.com/osism/terraform-base/.git/logs/\n<f..t...... src/github.com/osism/terraform-base/.git/logs/HEAD\n.d..t...... src/github.com/osism/terraform-base/.git/logs/refs/\n.d..t...... src/github.com/osism/terraform-base/.git/logs/refs/heads/\n<f..t...... src/github.com/osism/terraform-base/.git/logs/refs/heads/main\n.d..t...... src/github.com/osism/terraform-base/.git/logs/refs/remotes/\n.d..t...... src/github.com/osism/terraform-base/.git/logs/refs/remotes/origin/\n<f..t...... src/github.com/osism/terraform-base/.git/logs/refs/remotes/origin/HEAD\n.d..t...... src/github.com/osism/terraform-base/.git/objects/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/00/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/01/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/02/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/05/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/06/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/07/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/08/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/0a/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/0b/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/0d/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/0f/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/10/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/12/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/15/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/16/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/17/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/1b/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/1c/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/1f/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/20/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/21/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/24/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/26/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/29/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/2c/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/36/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/37/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/3a/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/3d/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/41/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/43/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/46/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/47/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/48/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/49/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/4b/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/4d/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/52/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/55/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/56/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/58/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/5d/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/61/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/63/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/66/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/67/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/68/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/6d/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/70/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/72/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/76/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/7c/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/7f/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/80/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/81/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/82/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/83/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/84/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/85/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/86/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/88/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/8c/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/8d/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/90/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/91/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/92/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/95/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/96/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/98/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/9b/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/9e/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/a2/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/a3/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/a6/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/a7/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/a9/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/aa/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/af/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/b0/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/b3/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/b4/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/b8/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/b9/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/ba/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/bc/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/bf/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/c0/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/c1/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/c4/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/c7/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/c8/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/cb/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/ce/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/cf/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/d2/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/d3/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/d6/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/d7/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/d9/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/da/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/db/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/dc/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/dd/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/de/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/df/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/e1/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/e2/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/e3/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/e4/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/e8/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/ea/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/ee/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/ef/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/f1/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/f2/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/f4/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/f9/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/fb/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/ff/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/info/\n.d..t...... src/github.com/osism/terraform-base/.git/objects/pack/\n.d..t...... src/github.com/osism/terraform-base/.git/refs/\n.d..t...... src/github.com/osism/terraform-base/.github/\n<f..t...... src/github.com/osism/terraform-base/.github/renovate.json\n.d..t...... src/github.com/osism/terraform-base/playbooks/\n<f..t...... src/github.com/osism/terraform-base/playbooks/validate.yml\n.d..t...... src/github.com/osism/terraform-base/testbed-default/\n<f..t...... src/github.com/osism/terraform-base/testbed-default/data.tf\n<f..t...... src/github.com/osism/terraform-base/testbed-default/main.tf\n<f..t...... src/github.com/osism/terraform-base/testbed-default/manager.tf\n<f..t...... src/github.com/osism/terraform-base/testbed-default/neutron.tf\n<f..t...... src/github.com/osism/terraform-base/testbed-default/nodes.tf\n<f..t...... src/github.com/osism/terraform-base/testbed-default/nova.tf\n<f..t...... src/github.com/osism/terraform-base/testbed-default/outputs.tf\n<f..t...... src/github.com/osism/terraform-base/testbed-default/provider.tf\n<f..t...... src/github.com/osism/terraform-base/testbed-default/variables.tf\n.d..t...... src/github.com/osism/terraform-base/testbed-default/customisations/\n<f..t...... src/github.com/osism/terraform-base/testbed-default/customisations/access_floatingip_custom.tf\n<f..t...... src/github.com/osism/terraform-base/testbed-default/customisations/access_ipv4_custom.tf\n<f..t...... src/github.com/osism/terraform-base/testbed-default/customisations/access_ipv6_custom.tf\n<f..t...... src/github.com/osism/terraform-base/testbed-default/customisations/default_custom.tf\n<f..t...... src/github.com/osism/terraform-base/testbed-default/customisations/external_api_custom.tf\n<f..t...... src/github.com/osism/terraform-base/testbed-default/customisations/neutron_floatingip_custom.tf\n.d..t...... src/github.com/osism/terraform-base/testbed-default/environments/\n<f..t...... src/github.com/osism/terraform-base/testbed-default/environments/default.tfvars\n.d..t...... src/github.com/osism/terraform-base/testbed-default/overrides/\n<f..t...... src/github.com/osism/terraform-base/testbed-default/overrides/manager_boot_from_image_override.tf\n<f..t...... src/github.com/osism/terraform-base/testbed-default/overrides/manager_boot_from_volume_override.tf\n<f..t...... src/github.com/osism/terraform-base/testbed-default/overrides/neutron_availability_zone_hints_network_override.tf\n<f..t...... src/github.com/osism/terraform-base/testbed-default/overrides/neutron_availability_zone_hints_router_override.tf\n<f..t...... src/github.com/osism/terraform-base/testbed-default/overrides/neutron_router_enable_snat_override.tf\n<f..t...... src/github.com/osism/terraform-base/testbed-default/overrides/nodes_boot_from_image_override.tf\n<f..t...... src/github.com/osism/terraform-base/testbed-default/overrides/nodes_boot_from_volume_override.tf\n<f..t...... src/github.com/osism/terraform-base/testbed-default/overrides/nodes_use_ephemeral_storage_override.tf\n.d..t...... src/github.com/osism/testbed/\n<f..t...... src/github.com/osism/testbed/.ansible-lint\n<f..t...... src/github.com/osism/testbed/.flake8\n<f..t...... src/github.com/osism/testbed/.gitignore\n<f..t...... src/github.com/osism/testbed/.yamllint.yml\n<f..t...... src/github.com/osism/testbed/.zuul.yaml\n<f..t...... src/github.com/osism/testbed/LICENSE\n<f..t...... src/github.com/osism/testbed/Makefile\n<f..t...... src/github.com/osism/testbed/README.md\n<f..t...... src/github.com/osism/testbed/ansible.cfg\n<f..t...... src/github.com/osism/testbed/gilt.yml\n<f..t...... src/github.com/osism/testbed/requirements.txt\n<f..t...... src/github.com/osism/testbed/requirements.yml\n<f..t...... src/github.com/osism/testbed/tox.ini\n.d..t...... src/github.com/osism/testbed/.git/\n<f..t...... src/github.com/osism/testbed/.git/HEAD\n<f..t...... src/github.com/osism/testbed/.git/ORIG_HEAD\n<f.st...... src/github.com/osism/testbed/.git/config\n<f..t...... src/github.com/osism/testbed/.git/config.worktree\n<f..t...... src/github.com/osism/testbed/.git/description\n<f..t...... src/github.com/osism/testbed/.git/index\n<f..t...... src/github.com/osism/testbed/.git/packed-refs\n.d..t...... src/github.com/osism/testbed/.git/branches/\n.d..t...... src/github.com/osism/testbed/.git/hooks/\n<f..t...... src/github.com/osism/testbed/.git/hooks/applypatch-msg.sample\n<f..t...... src/github.com/osism/testbed/.git/hooks/commit-msg.sample\n<f..t...... src/github.com/osism/testbed/.git/hooks/fsmonitor-watchman.sample\n<f..t...... src/github.com/osism/testbed/.git/hooks/post-update.sample\n<f..t...... src/github.com/osism/testbed/.git/hooks/pre-applypatch.sample\n<f..t...... src/github.com/osism/testbed/.git/hooks/pre-commit.sample\n<f..t...... src/github.com/osism/testbed/.git/hooks/pre-merge-commit.sample\n<f..t...... src/github.com/osism/testbed/.git/hooks/pre-push.sample\n<f..t...... src/github.com/osism/testbed/.git/hooks/pre-rebase.sample\n<f..t...... src/github.com/osism/testbed/.git/hooks/pre-receive.sample\n<f..t...... src/github.com/osism/testbed/.git/hooks/prepare-commit-msg.sample\n<f..t...... src/github.com/osism/testbed/.git/hooks/push-to-checkout.sample\n<f..t...... src/github.com/osism/testbed/.git/hooks/update.sample\n.d..t...... src/github.com/osism/testbed/.git/info/\n<f..t...... src/github.com/osism/testbed/.git/info/exclude\n.d..t...... src/github.com/osism/testbed/.git/logs/\n<f.st...... src/github.com/osism/testbed/.git/logs/HEAD\ncd+++++++++ src/github.com/osism/testbed/.git/logs/refs/\ncd+++++++++ src/github.com/osism/testbed/.git/logs/refs/heads/\n<f+++++++++ src/github.com/osism/testbed/.git/logs/refs/heads/main\ncd+++++++++ src/github.com/osism/testbed/.git/logs/refs/remotes/\ncd+++++++++ src/github.com/osism/testbed/.git/logs/refs/remotes/origin/\n<f+++++++++ src/github.com/osism/testbed/.git/logs/refs/remotes/origin/HEAD\n.d..t...... src/github.com/osism/testbed/.git/objects/\n.d..t...... src/github.com/osism/testbed/.git/objects/00/\n.d..t...... src/github.com/osism/testbed/.git/objects/01/\n.d..t...... src/github.com/osism/testbed/.git/objects/02/\n.d..t...... src/github.com/osism/testbed/.git/objects/03/\n.d..t...... src/github.com/osism/testbed/.git/objects/04/\n.d..t...... src/github.com/osism/testbed/.git/objects/05/\n.d..t...... src/github.com/osism/testbed/.git/objects/06/\n.d..t...... src/github.com/osism/testbed/.git/objects/07/\n<f..t...... src/github.com/osism/testbed/.git/objects/07/1d016fb2341fa1b57d73e6c901348ad5a5cc85\n.d..t...... src/github.com/osism/testbed/.git/objects/08/\n.d..t...... src/github.com/osism/testbed/.git/objects/09/\n.d..t...... src/github.com/osism/testbed/.git/objects/0a/\n.d..t...... src/github.com/osism/testbed/.git/objects/0b/\n.d..t...... src/github.com/osism/testbed/.git/objects/0c/\n.d..t...... src/github.com/osism/testbed/.git/objects/0d/\n.d..t...... src/github.com/osism/testbed/.git/objects/0e/\n.d..t...... src/github.com/osism/testbed/.git/objects/0f/\n.d..t...... src/github.com/osism/testbed/.git/objects/10/\n.d..t...... src/github.com/osism/testbed/.git/objects/11/\n.d..t...... src/github.com/osism/testbed/.git/objects/12/\n.d..t...... src/github.com/osism/testbed/.git/objects/13/\n.d..t...... src/github.com/osism/testbed/.git/objects/14/\n.d..t...... src/github.com/osism/testbed/.git/objects/15/\n.d..t...... src/github.com/osism/testbed/.git/objects/16/\n.d..t...... src/github.com/osism/testbed/.git/objects/17/\n.d..t...... src/github.com/osism/testbed/.git/objects/18/\n.d..t...... src/github.com/osism/testbed/.git/objects/19/\n.d..t...... src/github.com/osism/testbed/.git/objects/1a/\n.d..t...... src/github.com/osism/testbed/.git/objects/1b/\n.d..t...... src/github.com/osism/testbed/.git/objects/1c/\n.d..t...... src/github.com/osism/testbed/.git/objects/1d/\n.d..t...... src/github.com/osism/testbed/.git/objects/1e/\n.d..t...... src/github.com/osism/testbed/.git/objects/1f/\n.d..t...... src/github.com/osism/testbed/.git/objects/20/\n.d..t...... src/github.com/osism/testbed/.git/objects/21/\n.d..t...... src/github.com/osism/testbed/.git/objects/22/\n.d..t...... src/github.com/osism/testbed/.git/objects/23/\n.d..t...... src/github.com/osism/testbed/.git/objects/24/\n.d..t...... src/github.com/osism/testbed/.git/objects/25/\n.d..t...... src/github.com/osism/testbed/.git/objects/26/\n.d..t...... src/github.com/osism/testbed/.git/objects/27/\n.d..t...... src/github.com/osism/testbed/.git/objects/28/\n.d..t...... src/github.com/osism/testbed/.git/objects/29/\n.d..t...... src/github.com/osism/testbed/.git/objects/2a/\n.d..t...... src/github.com/osism/testbed/.git/objects/2b/\n.d..t...... src/github.com/osism/testbed/.git/objects/2c/\n.d..t...... src/github.com/osism/testbed/.git/objects/2d/\n.d..t...... src/github.com/osism/testbed/.git/objects/2e/\n.d..t...... src/github.com/osism/testbed/.git/objects/2f/\n.d..t...... src/github.com/osism/testbed/.git/objects/30/\n.d..t...... src/github.com/osism/testbed/.git/objects/31/\n.d..t...... src/github.com/osism/testbed/.git/objects/32/\n.d..t...... src/github.com/osism/testbed/.git/objects/33/\n.d..t...... src/github.com/osism/testbed/.git/objects/34/\n.d..t...... src/github.com/osism/testbed/.git/objects/35/\n.d..t...... src/github.com/osism/testbed/.git/objects/36/\n.d..t...... src/github.com/osism/testbed/.git/objects/37/\n.d..t...... src/github.com/osism/testbed/.git/objects/38/\n.d..t...... src/github.com/osism/testbed/.git/objects/39/\n.d..t...... src/github.com/osism/testbed/.git/objects/3a/\n.d..t...... src/github.com/osism/testbed/.git/objects/3b/\n.d..t...... src/github.com/osism/testbed/.git/objects/3c/\n.d..t...... src/github.com/osism/testbed/.git/objects/3d/\n.d..t...... src/github.com/osism/testbed/.git/objects/3e/\n.d..t...... src/github.com/osism/testbed/.git/objects/3f/\n.d..t...... src/github.com/osism/testbed/.git/objects/40/\n.d..t...... src/github.com/osism/testbed/.git/objects/41/\n.d..t...... src/github.com/osism/testbed/.git/objects/42/\n.d..t...... src/github.com/osism/testbed/.git/objects/43/\n.d..t...... src/github.com/osism/testbed/.git/objects/44/\n.d..t...... src/github.com/osism/testbed/.git/objects/45/\n.d..t...... src/github.com/osism/testbed/.git/objects/46/\n.d..t...... src/github.com/osism/testbed/.git/objects/47/\n.d..t...... src/github.com/osism/testbed/.git/objects/48/\n.d..t...... src/github.com/osism/testbed/.git/objects/49/\n.d..t...... src/github.com/osism/testbed/.git/objects/4a/\n.d..t...... src/github.com/osism/testbed/.git/objects/4c/\n.d..t...... src/github.com/osism/testbed/.git/objects/4d/\n.d..t...... src/github.com/osism/testbed/.git/objects/4e/\n.d..t...... src/github.com/osism/testbed/.git/objects/4f/\n.d..t...... src/github.com/osism/testbed/.git/objects/50/\n.d..t...... src/github.com/osism/testbed/.git/objects/51/\n.d..t...... src/github.com/osism/testbed/.git/objects/52/\n.d..t...... src/github.com/osism/testbed/.git/objects/53/\n.d..t...... src/github.com/osism/testbed/.git/objects/54/\n.d..t...... src/github.com/osism/testbed/.git/objects/55/\n.d..t...... src/github.com/osism/testbed/.git/objects/56/\n.d..t...... src/github.com/osism/testbed/.git/objects/57/\n.d..t...... src/github.com/osism/testbed/.git/objects/58/\n.d..t...... src/github.com/osism/testbed/.git/objects/59/\n.d..t...... src/github.com/osism/testbed/.git/objects/5a/\n.d..t...... src/github.com/osism/testbed/.git/objects/5b/\n.d..t...... src/github.com/osism/testbed/.git/objects/5c/\n.d..t...... src/github.com/osism/testbed/.git/objects/5d/\n.d..t...... src/github.com/osism/testbed/.git/objects/5e/\n.d..t...... src/github.com/osism/testbed/.git/objects/5f/\n.d..t...... src/github.com/osism/testbed/.git/objects/60/\n.d..t...... src/github.com/osism/testbed/.git/objects/61/\n.d..t...... src/github.com/osism/testbed/.git/objects/64/\n.d..t...... src/github.com/osism/testbed/.git/objects/65/\n.d..t...... src/github.com/osism/testbed/.git/objects/66/\n.d..t...... src/github.com/osism/testbed/.git/objects/67/\n.d..t...... src/github.com/osism/testbed/.git/objects/68/\n.d..t...... src/github.com/osism/testbed/.git/objects/69/\n.d..t...... src/github.com/osism/testbed/.git/objects/6a/\n.d..t...... src/github.com/osism/testbed/.git/objects/6b/\n.d..t...... src/github.com/osism/testbed/.git/objects/6c/\n.d..t...... src/github.com/osism/testbed/.git/objects/6d/\n.d..t...... src/github.com/osism/testbed/.git/objects/6e/\n.d..t...... src/github.com/osism/testbed/.git/objects/6f/\n.d..t...... src/github.com/osism/testbed/.git/objects/70/\n.d..t...... src/github.com/osism/testbed/.git/objects/71/\n.d..t...... src/github.com/osism/testbed/.git/objects/72/\n.d..t...... src/github.com/osism/testbed/.git/objects/73/\n.d..t...... src/github.com/osism/testbed/.git/objects/74/\n.d..t...... src/github.com/osism/testbed/.git/objects/75/\n.d..t...... src/github.com/osism/testbed/.git/objects/76/\n.d..t...... src/github.com/osism/testbed/.git/objects/77/\n.d..t...... src/github.com/osism/testbed/.git/objects/78/\n.d..t...... src/github.com/osism/testbed/.git/objects/79/\n.d..t...... src/github.com/osism/testbed/.git/objects/7a/\n.d..t...... src/github.com/osism/testbed/.git/objects/7b/\n.d..t...... src/github.com/osism/testbed/.git/objects/7c/\n.d..t...... src/github.com/osism/testbed/.git/objects/7d/\n.d..t...... src/github.com/osism/testbed/.git/objects/7e/\n.d..t...... src/github.com/osism/testbed/.git/objects/7f/\n.d..t...... src/github.com/osism/testbed/.git/objects/80/\n.d..t...... src/github.com/osism/testbed/.git/objects/81/\n.d..t...... src/github.com/osism/testbed/.git/objects/82/\n.d..t...... src/github.com/osism/testbed/.git/objects/83/\n.d..t...... src/github.com/osism/testbed/.git/objects/84/\n.d..t...... src/github.com/osism/testbed/.git/objects/85/\n.d..t...... src/github.com/osism/testbed/.git/objects/86/\n.d..t...... src/github.com/osism/testbed/.git/objects/87/\n.d..t...... src/github.com/osism/testbed/.git/objects/8a/\n.d..t...... src/github.com/osism/testbed/.git/objects/8b/\n.d..t...... src/github.com/osism/testbed/.git/objects/8c/\n.d..t...... src/github.com/osism/testbed/.git/objects/8e/\n.d..t...... src/github.com/osism/testbed/.git/objects/8f/\n.d..t...... src/github.com/osism/testbed/.git/objects/90/\n.d..t...... src/github.com/osism/testbed/.git/objects/91/\n.d..t...... src/github.com/osism/testbed/.git/objects/92/\n.d..t...... src/github.com/osism/testbed/.git/objects/93/\n.d..t...... src/github.com/osism/testbed/.git/objects/94/\n.d..t...... src/github.com/osism/testbed/.git/objects/95/\n.d..t...... src/github.com/osism/testbed/.git/objects/96/\n.d..t...... src/github.com/osism/testbed/.git/objects/97/\n.d..t...... src/github.com/osism/testbed/.git/objects/98/\n.d..t...... src/github.com/osism/testbed/.git/objects/9a/\n.d..t...... src/github.com/osism/testbed/.git/objects/9c/\n.d..t...... src/github.com/osism/testbed/.git/objects/9e/\n.d..t...... src/github.com/osism/testbed/.git/objects/9f/\n.d..t...... src/github.com/osism/testbed/.git/objects/a0/\n.d..t...... src/github.com/osism/testbed/.git/objects/a1/\n.d..t...... src/github.com/osism/testbed/.git/objects/a2/\n.d..t...... src/github.com/osism/testbed/.git/objects/a3/\n.d..t...... src/github.com/osism/testbed/.git/objects/a4/\n.d..t...... src/github.com/osism/testbed/.git/objects/a5/\n.d..t...... src/github.com/osism/testbed/.git/objects/a6/\n.d..t...... src/github.com/osism/testbed/.git/objects/a7/\n.d..t...... src/github.com/osism/testbed/.git/objects/a8/\n.d..t...... src/github.com/osism/testbed/.git/objects/a9/\n.d..t...... src/github.com/osism/testbed/.git/objects/aa/\n.d..t...... src/github.com/osism/testbed/.git/objects/ab/\n.d..t...... src/github.com/osism/testbed/.git/objects/ac/\n.d..t...... src/github.com/osism/testbed/.git/objects/ad/\n.d..t...... src/github.com/osism/testbed/.git/objects/ae/\n.d..t...... src/github.com/osism/testbed/.git/objects/af/\n.d..t...... src/github.com/osism/testbed/.git/objects/b0/\n.d..t...... src/github.com/osism/testbed/.git/objects/b1/\n.d..t...... src/github.com/osism/testbed/.git/objects/b2/\n.d..t...... src/github.com/osism/testbed/.git/objects/b3/\n.d..t...... src/github.com/osism/testbed/.git/objects/b4/\n.d..t...... src/github.com/osism/testbed/.git/objects/b5/\n.d..t...... src/github.com/osism/testbed/.git/objects/b6/\n.d..t...... src/github.com/osism/testbed/.git/objects/b7/\n.d..t...... src/github.com/osism/testbed/.git/objects/b8/\n.d..t...... src/github.com/osism/testbed/.git/objects/b9/\n.d..t...... src/github.com/osism/testbed/.git/objects/ba/\n.d..t...... src/github.com/osism/testbed/.git/objects/bb/\n.d..t...... src/github.com/osism/testbed/.git/objects/bc/\n.d..t...... src/github.com/osism/testbed/.git/objects/bd/\n.d..t...... src/github.com/osism/testbed/.git/objects/be/\n.d..t...... src/github.com/osism/testbed/.git/objects/bf/\n.d..t...... src/github.com/osism/testbed/.git/objects/c0/\n.d..t...... src/github.com/osism/testbed/.git/objects/c1/\n.d..t...... src/github.com/osism/testbed/.git/objects/c2/\n.d..t...... src/github.com/osism/testbed/.git/objects/c3/\n.d..t...... src/github.com/osism/testbed/.git/objects/c4/\n.d..t...... src/github.com/osism/testbed/.git/objects/c5/\n.d..t...... src/github.com/osism/testbed/.git/objects/c6/\n.d..t...... src/github.com/osism/testbed/.git/objects/c7/\n.d..t...... src/github.com/osism/testbed/.git/objects/c8/\n.d..t...... src/github.com/osism/testbed/.git/objects/c9/\n.d..t...... src/github.com/osism/testbed/.git/objects/ca/\n.d..t...... src/github.com/osism/testbed/.git/objects/cb/\n.d..t...... src/github.com/osism/testbed/.git/objects/cc/\n.d..t...... src/github.com/osism/testbed/.git/objects/cd/\n.d..t...... src/github.com/osism/testbed/.git/objects/ce/\n.d..t...... src/github.com/osism/testbed/.git/objects/cf/\n.d..t...... src/github.com/osism/testbed/.git/objects/d0/\n.d..t...... src/github.com/osism/testbed/.git/objects/d2/\n.d..t...... src/github.com/osism/testbed/.git/objects/d3/\n.d..t...... src/github.com/osism/testbed/.git/objects/d5/\n.d..t...... src/github.com/osism/testbed/.git/objects/d6/\n.d..t...... src/github.com/osism/testbed/.git/objects/d7/\n.d..t...... src/github.com/osism/testbed/.git/objects/d8/\n.d..t...... src/github.com/osism/testbed/.git/objects/d9/\n.d..t...... src/github.com/osism/testbed/.git/objects/da/\n.d..t...... src/github.com/osism/testbed/.git/objects/db/\n.d..t...... src/github.com/osism/testbed/.git/objects/dc/\n.d..t...... src/github.com/osism/testbed/.git/objects/dd/\n.d..t...... src/github.com/osism/testbed/.git/objects/de/\n.d..t...... src/github.com/osism/testbed/.git/objects/df/\n.d..t...... src/github.com/osism/testbed/.git/objects/e0/\n.d..t...... src/github.com/osism/testbed/.git/objects/e2/\n.d..t...... src/github.com/osism/testbed/.git/objects/e3/\n.d..t...... src/github.com/osism/testbed/.git/objects/e4/\n.d..t...... src/github.com/osism/testbed/.git/objects/e5/\n.d..t...... src/github.com/osism/testbed/.git/objects/e6/\n.d..t...... src/github.com/osism/testbed/.git/objects/e7/\n.d..t...... src/github.com/osism/testbed/.git/objects/e8/\n.d..t...... src/github.com/osism/testbed/.git/objects/e9/\n.d..t...... src/github.com/osism/testbed/.git/objects/ea/\n.d..t...... src/github.com/osism/testbed/.git/objects/eb/\n.d..t...... src/github.com/osism/testbed/.git/objects/ec/\n.d..t...... src/github.com/osism/testbed/.git/objects/ed/\n.d..t...... src/github.com/osism/testbed/.git/objects/ee/\n.d..t...... src/github.com/osism/testbed/.git/objects/ef/\n.d..t...... src/github.com/osism/testbed/.git/objects/f0/\n.d..t...... src/github.com/osism/testbed/.git/objects/f1/\n.d..t...... src/github.com/osism/testbed/.git/objects/f2/\n.d..t...... src/github.com/osism/testbed/.git/objects/f3/\n.d..t...... src/github.com/osism/testbed/.git/objects/f4/\n.d..t...... src/github.com/osism/testbed/.git/objects/f5/\n.d..t...... src/github.com/osism/testbed/.git/objects/f6/\n.d..t...... src/github.com/osism/testbed/.git/objects/f8/\n.d..t...... src/github.com/osism/testbed/.git/objects/f9/\n.d..t...... src/github.com/osism/testbed/.git/objects/fa/\n.d..t...... src/github.com/osism/testbed/.git/objects/fb/\n.d..t...... src/github.com/osism/testbed/.git/objects/fc/\n.d..t...... src/github.com/osism/testbed/.git/objects/fd/\n.d..t...... src/github.com/osism/testbed/.git/objects/fe/\n.d..t...... src/github.com/osism/testbed/.git/objects/ff/\n.d..t...... src/github.com/osism/testbed/.git/objects/info/\n.d..t...... src/github.com/osism/testbed/.git/objects/pack/\n.d..t...... src/github.com/osism/testbed/.git/refs/\n.d..t...... src/github.com/osism/testbed/.github/\n<f..t...... src/github.com/osism/testbed/.github/renovate.json\n.d..t...... src/github.com/osism/testbed/.github/workflows/\n<f..t...... src/github.com/osism/testbed/.github/workflows/update-manager-images.yml\n.d..t...... src/github.com/osism/testbed/ansible/\n<f..t...... src/github.com/osism/testbed/ansible/check-local-versions.yml\n<f..t...... src/github.com/osism/testbed/ansible/localhost_inventory.yaml\n<f..t...... src/github.com/osism/testbed/ansible/manager-part-0.yml\n<f..t...... src/github.com/osism/testbed/ansible/manager-part-1.yml\n<f..t...... src/github.com/osism/testbed/ansible/manager-part-2.yml\n<f..t...... src/github.com/osism/testbed/ansible/manager-part-3.yml\n.d..t...... src/github.com/osism/testbed/containerlab/\n<f..t...... src/github.com/osism/testbed/containerlab/README.md\n<f..t...... src/github.com/osism/testbed/containerlab/osism.clab.yml\n.d..t...... src/github.com/osism/testbed/containerlab/files/\n<f..t...... src/github.com/osism/testbed/containerlab/files/daemons\n<f..t...... src/github.com/osism/testbed/containerlab/files/frr01.conf\n<f..t...... src/github.com/osism/testbed/containerlab/files/frr02.conf\n<f..t...... src/github.com/osism/testbed/containerlab/files/vtysh.conf\n.d..t...... src/github.com/osism/testbed/contrib/\n<f..t...... src/github.com/osism/testbed/contrib/analyze-tempest-logs.sh\n<f..t...... src/github.com/osism/testbed/contrib/analyze-zuul-logs.py\n<f..t...... src/github.com/osism/testbed/contrib/elasticsearch_random_data_load.sh\n<f..t...... src/github.com/osism/testbed/contrib/mysql_random_data_load.sh\n<f..t...... src/github.com/osism/testbed/contrib/mysql_random_data_load.sql\n<f..t...... src/github.com/osism/testbed/contrib/semver2.sh\n<f..t...... src/github.com/osism/testbed/contrib/setup-testbed.py\n.d..t...... src/github.com/osism/testbed/contrib/ansible/\n<f..t...... src/github.com/osism/testbed/contrib/ansible/clouds.yaml.sample\n<f..t...... src/github.com/osism/testbed/contrib/ansible/dns.yml\n<f..t...... src/github.com/osism/testbed/contrib/ansible/secure.yaml.sample\n.d..t...... src/github.com/osism/testbed/contrib/assets/\n<f..t...... src/github.com/osism/testbed/contrib/assets/operations-dashboard.png\n.d..t...... src/github.com/osism/testbed/contrib/integration-tests/\n<f..t...... src/github.com/osism/testbed/contrib/integration-tests/docker-compose.yaml\n<f..t...... src/github.com/osism/testbed/contrib/integration-tests/run_test.sh\n.d..t...... src/github.com/osism/testbed/contrib/integration-tests/tests/\n<f..t...... src/github.com/osism/testbed/contrib/integration-tests/tests/new_user.robot\n.d..t...... src/github.com/osism/testbed/contrib/nagios-plugins/\n<f..t...... src/github.com/osism/testbed/contrib/nagios-plugins/check_elasticsearch\n<f..t...... src/github.com/osism/testbed/contrib/nagios-plugins/check_galera_cluster\n<f..t...... src/github.com/osism/testbed/contrib/nagios-plugins/check_rabbitmq_cluster\n.d..t...... src/github.com/osism/testbed/contrib/openstack-cli/\n<f..t...... src/github.com/osism/testbed/contrib/openstack-cli/README.md\n<f..t...... src/github.com/osism/testbed/contrib/openstack-cli/clouds-public.yaml\n<f..t...... src/github.com/osism/testbed/contrib/openstack-cli/clouds.yaml.example\n<f..t...... src/github.com/osism/testbed/contrib/openstack-cli/secure.yaml.example\n.d..t...... src/github.com/osism/testbed/contrib/ownca/\n<f..t...... src/github.com/osism/testbed/contrib/ownca/README.md\n<f..t...... src/github.com/osism/testbed/contrib/ownca/create_ca.yml\n<f..t...... src/github.com/osism/testbed/contrib/ownca/create_manager.yml\n<f..t...... src/github.com/osism/testbed/contrib/ownca/create_wildcard.yml\n.d..t...... src/github.com/osism/testbed/environments/\n<f..t...... src/github.com/osism/testbed/environments/.vault_pass\n<f..t...... src/github.com/osism/testbed/environments/ansible.cfg\n<f..t...... src/github.com/osism/testbed/environments/configuration.yml\n<f..t...... src/github.com/osism/testbed/environments/images.yml\n<f..t...... src/github.com/osism/testbed/environments/secrets.yml\n.d..t...... src/github.com/osism/testbed/environments/ceph.test/\n<f..t...... src/github.com/osism/testbed/environments/ceph.test/configuration.yml\n<f..t...... src/github.com/osism/testbed/environments/ceph.test/images.yml\n<f..t...... src/github.com/osism/testbed/environments/ceph.test/secrets.yml\n.d..t...... src/github.com/osism/testbed/environments/ceph/\n<f..t...... src/github.com/osism/testbed/environments/ceph/configuration.yml\n<f..t...... src/github.com/osism/testbed/environments/ceph/images.yml\n<f..t...... src/github.com/osism/testbed/environments/ceph/secrets.yml\n.d..t...... src/github.com/osism/testbed/environments/custom/\n<f..t...... src/github.com/osism/testbed/environments/custom/configuration.yml\n<f..t...... src/github.com/osism/testbed/environments/custom/images.yml\n<f..t...... src/github.com/osism/testbed/environments/custom/playbook-baremetal-bootstrap.yml\n<f..t...... src/github.com/osism/testbed/environments/custom/playbook-baremetal-netbox.yml\n<f..t...... src/github.com/osism/testbed/environments/custom/playbook-baremetal-pull-images.yml\n<f..t...... src/github.com/osism/testbed/environments/custom/playbook-facts.yml\n<f..t...... src/github.com/osism/testbed/environments/custom/playbook-migrate-proxysql.yml\n<f..t...... src/github.com/osism/testbed/environments/custom/playbook-pull-images.yml\n<f..t...... src/github.com/osism/testbed/environments/custom/playbook-wipe-partitions.yml\n<f..t...... src/github.com/osism/testbed/environments/custom/playbook-workarounds-ceph.yml\n<f..t...... src/github.com/osism/testbed/environments/custom/playbook-workarounds.yml\n<f..t...... src/github.com/osism/testbed/environments/custom/secrets.yml\n.d..t...... src/github.com/osism/testbed/environments/custom/files/\n<f..t...... src/github.com/osism/testbed/environments/custom/files/baremetal-tenks-override.yml\n<f..t...... src/github.com/osism/testbed/environments/custom/files/testbed_ceph_devices.fact\n<f..t...... src/github.com/osism/testbed/environments/custom/files/testbed_ceph_devices_all.fact\n<f..t...... src/github.com/osism/testbed/environments/custom/files/testbed_ceph_osd_devices.fact\n<f..t...... src/github.com/osism/testbed/environments/custom/files/testbed_ceph_osd_devices_all.fact\n<f..t...... src/github.com/osism/testbed/environments/custom/files/testbed_network_devices.fact\n.d..t...... src/github.com/osism/testbed/environments/custom/templates/\n<f..t...... src/github.com/osism/testbed/environments/custom/templates/baremetal-netbox-device.yml.j2\n<f..t...... src/github.com/osism/testbed/environments/custom/templates/ironic-interface.service.j2\n<f..t...... src/github.com/osism/testbed/environments/custom/templates/workarounds.service.j2\n<f..t...... src/github.com/osism/testbed/environments/custom/templates/workarounds.sh.j2\n.d..t...... src/github.com/osism/testbed/environments/generic/\n<f..t...... src/github.com/osism/testbed/environments/generic/configuration.yml\n<f..t...... src/github.com/osism/testbed/environments/generic/images.yml\n<f..t...... src/github.com/osism/testbed/environments/generic/secrets.yml\n.d..t...... src/github.com/osism/testbed/environments/infrastructure/\n<f..t...... src/github.com/osism/testbed/environments/infrastructure/configuration.yml\n<f..t...... src/github.com/osism/testbed/environments/infrastructure/images.yml\n<f..t...... src/github.com/osism/testbed/environments/infrastructure/secrets.yml\n.d..t...... src/github.com/osism/testbed/environments/infrastructure/files/\n.d..t...... src/github.com/osism/testbed/environments/infrastructure/files/ceph/\n<f..t...... src/github.com/osism/testbed/environments/infrastructure/files/ceph/.gitkeep\n.d..t...... src/github.com/osism/testbed/environments/infrastructure/files/keycloak/\n<f..t...... src/github.com/osism/testbed/environments/infrastructure/files/keycloak/cert.crt\n<f..t...... src/github.com/osism/testbed/environments/infrastructure/files/keycloak/private_key.pem\n.d..t...... src/github.com/osism/testbed/environments/kolla/\n<f.st...... src/github.com/osism/testbed/environments/kolla/configuration.yml\n<f..t...... src/github.com/osism/testbed/environments/kolla/images.yml\n<f..t...... src/github.com/osism/testbed/environments/kolla/secrets.yml\n.d..t...... src/github.com/osism/testbed/environments/kolla/certificates/\n<f..t...... src/github.com/osism/testbed/environments/kolla/certificates/haproxy-internal.pem\n<f..t...... src/github.com/osism/testbed/environments/kolla/certificates/haproxy.pem\n.d..t...... src/github.com/osism/testbed/environments/kolla/certificates/ca/\n<f..t...... src/github.com/osism/testbed/environments/kolla/certificates/ca/testbed.crt\n.d..t...... src/github.com/osism/testbed/environments/kolla/files/\n.d..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/\n<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/cinder.conf\n<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/galera.cnf\n<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/global.conf\n<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/grafana.ini\n<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/keystone.conf\n<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/magnum.conf\n<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/neutron.conf\n.d..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/ceilometer/\n<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/ceilometer/event_pipeline.yaml\n<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/ceilometer/pipeline.yaml\n.d..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/cinder/\n<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/cinder/ceph.conf\n<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/cinder/cinder-scheduler.conf\n<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/cinder/cinder-volume.conf\n.d..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/cinder/cinder-backup/\n<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/cinder/cinder-backup/.gitkeep\n.d..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/cinder/cinder-volume/\n<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/cinder/cinder-volume/.gitkeep\n.d..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/glance/\n<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/glance/ceph.conf\n<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/glance/glance-api.conf\n.d..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/gnocchi/\n<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/gnocchi/ceph.conf\n.d..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/grafana/\n<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/grafana/provisioning.yaml\n.d..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/grafana/dashboards/\n<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/grafana/dashboards/.gitkeep\n.d..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/haproxy/\n.d..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/haproxy/services.d/\n<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/haproxy/services.d/haproxy.cfg\n.d..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/ironic/\n<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/ironic/ironic-agent.initramfs\n<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/ironic/ironic-agent.kernel\n<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/ironic/ironic-conductor.conf\n<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/ironic/ironic-dnsmasq.conf\n.d..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/magnum/\n<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/magnum/.gitkeep\n.d..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/manila/\n<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/manila/ceph.conf\n.d..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/neutron/\n<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/neutron/ml2_conf.ini\n.d..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/nova/\n<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/nova/ceph.conf\n<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/nova/nova-compute.conf\n.d..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/octavia/\n<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/octavia/client.cert-and-key.pem\n<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/octavia/client_ca.cert.pem\n<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/octavia/server_ca.cert.pem\n<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/octavia/server_ca.key.pem\n.d..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/prometheus/\n<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/prometheus/prometheus-alertmanager.yml\n.d..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/prometheus/prometheus.yml.d/\n<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/prometheus/prometheus.yml.d/50-ceph.yml\n.d..t...... src/github.com/osism/testbed/environments/kubernetes/\n<f..t...... src/github.com/osism/testbed/environments/kubernetes/configuration.yml\n<f..t...... src/github.com/osism/testbed/environments/kubernetes/images.yml\n<f..t...... src/github.com/osism/testbed/environments/kubernetes/playbook-nginx.yml\n<f..t...... src/github.com/osism/testbed/environments/kubernetes/secrets.yml\n.d..t...... src/github.com/osism/testbed/environments/kubernetes/files/\n<f..t...... src/github.com/osism/testbed/environments/kubernetes/files/nginx.yml\n.d..t...... src/github.com/osism/testbed/environments/manager/\n<f..t...... src/github.com/osism/testbed/environments/manager/configuration.yml\n<f..t...... src/github.com/osism/testbed/environments/manager/hosts\n<f..t...... src/github.com/osism/testbed/environments/manager/images.yml\n<f..t...... src/github.com/osism/testbed/environments/manager/secrets.yml\n.d..t...... src/github.com/osism/testbed/environments/manager/files/\n<f..t...... src/github.com/osism/testbed/environments/manager/files/conductor.yml\n.d..t...... src/github.com/osism/testbed/environments/monitoring/\n<f..t...... src/github.com/osism/testbed/environments/monitoring/configuration.yml\n<f..t...... src/github.com/osism/testbed/environments/monitoring/images.yml\n<f..t...... src/github.com/osism/testbed/environments/monitoring/secrets.yml\n.d..t...... src/github.com/osism/testbed/environments/openstack/\n<f..t...... src/github.com/osism/testbed/environments/openstack/clouds.yml\n<f..t...... src/github.com/osism/testbed/environments/openstack/configuration.yml\n<f..t...... src/github.com/osism/testbed/environments/openstack/id_rsa.test\n<f..t...... src/github.com/osism/testbed/environments/openstack/id_rsa.test.pub\n<f..t...... src/github.com/osism/testbed/environments/openstack/images.yml\n<f..t...... src/github.com/osism/testbed/environments/openstack/openrc\n<f..t...... src/github.com/osism/testbed/environments/openstack/playbook-bootstrap-basic.yml\n<f..t...... src/github.com/osism/testbed/environments/openstack/playbook-bootstrap-ceph-rgw.yml\n<f..t...... src/github.com/osism/testbed/environments/openstack/playbook-bootstrap-flavors.yml\n<f..t...... src/github.com/osism/testbed/environments/openstack/playbook-bootstrap-images.yml\n<f..t...... src/github.com/osism/testbed/environments/openstack/playbook-test-identity.yml\n<f..t...... src/github.com/osism/testbed/environments/openstack/playbook-test.yml\n<f..t...... src/github.com/osism/testbed/environments/openstack/secrets.yml\n<f..t...... src/github.com/osism/testbed/environments/openstack/secure.yml\n<f..t...... src/github.com/osism/testbed/environments/openstack/testbed.pem\n.d..t...... src/github.com/osism/testbed/environments/openstack/files/\n.d..t...... src/github.com/osism/testbed/environments/openstack/files/tempest/\n<f..t...... src/github.com/osism/testbed/environments/openstack/files/tempest/exclude.lst\n<f..t...... src/github.com/osism/testbed/environments/openstack/files/tempest/include-scs-compatible.lst\n.d..t...... src/github.com/osism/testbed/inventory/\n<f..t...... src/github.com/osism/testbed/inventory/10-custom\n<f..t...... src/github.com/osism/testbed/inventory/20-roles\n<f..t...... src/github.com/osism/testbed/inventory/99-overwrite\n.d..t...... src/github.com/osism/testbed/inventory/group_vars/\n<f..t...... src/github.com/osism/testbed/inventory/group_vars/testbed-external-nodes.yml\n<f..t...... src/github.com/osism/testbed/inventory/group_vars/testbed-managers.yml\n<f..t...... src/github.com/osism/testbed/inventory/group_vars/testbed-nodes.yml\n.d..t...... src/github.com/osism/testbed/inventory/group_vars/all/\n<f..t...... src/github.com/osism/testbed/inventory/group_vars/all/ansible.yml\n<f..t...... src/github.com/osism/testbed/inventory/group_vars/all/kolla.yml\n<f..t...... src/github.com/osism/testbed/inventory/group_vars/all/registries.yml\n<f..t...... src/github.com/osism/testbed/inventory/group_vars/all/vxlan.yml\n.d..t...... src/github.com/osism/testbed/inventory/host_vars/\n.d..t...... src/github.com/osism/testbed/inventory/host_vars/testbed-manager/\n<f..t...... src/github.com/osism/testbed/inventory/host_vars/testbed-manager/vars.yml\n.d..t...... src/github.com/osism/testbed/inventory/host_vars/testbed-node-0/\n<f..t...... src/github.com/osism/testbed/inventory/host_vars/testbed-node-0/vars.yml\n.d..t...... src/github.com/osism/testbed/inventory/host_vars/testbed-node-1/\n<f..t...... src/github.com/osism/testbed/inventory/host_vars/testbed-node-1/vars.yml\n.d..t...... src/github.com/osism/testbed/inventory/host_vars/testbed-node-2/\n<f..t...... src/github.com/osism/testbed/inventory/host_vars/testbed-node-2/vars.yml\n.d..t...... src/github.com/osism/testbed/inventory/host_vars/testbed-node-3/\n<f..t...... src/github.com/osism/testbed/inventory/host_vars/testbed-node-3/vars.yml\n.d..t...... src/github.com/osism/testbed/inventory/host_vars/testbed-node-4/\n<f..t...... src/github.com/osism/testbed/inventory/host_vars/testbed-node-4/vars.yml\n.d..t...... src/github.com/osism/testbed/inventory/host_vars/testbed-node-5/\n<f..t...... src/github.com/osism/testbed/inventory/host_vars/testbed-node-5/vars.yml\n.d..t...... src/github.com/osism/testbed/inventory/host_vars/testbed-node-6/\n<f..t...... src/github.com/osism/testbed/inventory/host_vars/testbed-node-6/vars.yml\n.d..t...... src/github.com/osism/testbed/inventory/host_vars/testbed-node-7/\n<f..t...... src/github.com/osism/testbed/inventory/host_vars/testbed-node-7/vars.yml\n.d..t...... src/github.com/osism/testbed/inventory/host_vars/testbed-node-8/\n<f..t...... src/github.com/osism/testbed/inventory/host_vars/testbed-node-8/vars.yml\n.d..t...... src/github.com/osism/testbed/inventory/host_vars/testbed-node-9/\n<f..t...... src/github.com/osism/testbed/inventory/host_vars/testbed-node-9/vars.yml\n.d..t...... src/github.com/osism/testbed/netbox/\n.d..t...... src/github.com/osism/testbed/netbox/devicetypes/\n.d..t...... src/github.com/osism/testbed/netbox/devicetypes/Edgecore/\n<f..t...... src/github.com/osism/testbed/netbox/devicetypes/Edgecore/5835-54X-O-AC-F.yaml\n<f..t...... src/github.com/osism/testbed/netbox/devicetypes/Edgecore/7726-32X-O-AC-F.yaml\n.d..t...... src/github.com/osism/testbed/netbox/devicetypes/Other/\n<f..t...... src/github.com/osism/testbed/netbox/devicetypes/Other/baremetal-device.yml\n<f..t...... src/github.com/osism/testbed/netbox/devicetypes/Other/baremetal-housing.yml\n<f..t...... src/github.com/osism/testbed/netbox/devicetypes/Other/manager.yml\n<f..t...... src/github.com/osism/testbed/netbox/devicetypes/Other/node.yml\n.d..t...... src/github.com/osism/testbed/netbox/moduletypes/\n<f..t...... src/github.com/osism/testbed/netbox/moduletypes/.gitkeep\n.d..t...... src/github.com/osism/testbed/netbox/resources/\n<f..t...... src/github.com/osism/testbed/netbox/resources/100-initialise.yml\n<f..t...... src/github.com/osism/testbed/netbox/resources/200-rack-1000.yml\n<f..t...... src/github.com/osism/testbed/netbox/resources/300-testbed-manager.yml\n<f..t...... src/github.com/osism/testbed/netbox/resources/300-testbed-node-0.yml\n<f..t...... src/github.com/osism/testbed/netbox/resources/300-testbed-node-1.yml\n<f..t...... src/github.com/osism/testbed/netbox/resources/300-testbed-node-2.yml\n<f..t...... src/github.com/osism/testbed/netbox/resources/300-testbed-node-3.yml\n<f..t...... src/github.com/osism/testbed/netbox/resources/300-testbed-node-4.yml\n<f..t...... src/github.com/osism/testbed/netbox/resources/300-testbed-node-5.yml\n<f..t...... src/github.com/osism/testbed/netbox/resources/300-testbed-node-6.yml\n<f..t...... src/github.com/osism/testbed/netbox/resources/300-testbed-node-7.yml\n<f..t...... src/github.com/osism/testbed/netbox/resources/300-testbed-node-8.yml\n<f..t...... src/github.com/osism/testbed/netbox/resources/300-testbed-node-9.yml\n<f..t...... src/github.com/osism/testbed/netbox/resources/300-testbed-switch-0.yml\n<f..t...... src/github.com/osism/testbed/netbox/resources/300-testbed-switch-1.yml\n<f..t...... src/github.com/osism/testbed/netbox/resources/300-testbed-switch-2.yml\n<f..t...... src/github.com/osism/testbed/netbox/resources/300-testbed-switch-3.yml\n<f..t...... src/github.com/osism/testbed/netbox/resources/300-testbed-switch-oob.yml\n.d..t...... src/github.com/osism/testbed/network/\n<f..t...... src/github.com/osism/testbed/network/iptables.sh\n<f..t...... src/github.com/osism/testbed/network/vxlan.sh\n.d..t...... src/github.com/osism/testbed/playbooks/\n<f..t...... src/github.com/osism/testbed/playbooks/cleanup.yml\n<f..t...... src/github.com/osism/testbed/playbooks/deploy.yml\n<f..t...... src/github.com/osism/testbed/playbooks/post.yml\n<f..t...... src/github.com/osism/testbed/playbooks/pre.yml\n<f..t...... src/github.com/osism/testbed/playbooks/update-stable.yml\n<f..t...... src/github.com/osism/testbed/playbooks/upgrade-stable.yml\n<f..t...... src/github.com/osism/testbed/playbooks/upgrade.yml\n.d..t...... src/github.com/osism/testbed/playbooks/templates/\n<f..t...... src/github.com/osism/testbed/playbooks/templates/local.env.j2\n.d..t...... src/github.com/osism/testbed/playbooks/vars/\n<f..t...... src/github.com/osism/testbed/playbooks/vars/mappings.yml\n<f..t...... src/github.com/osism/testbed/playbooks/vars/repositories.yml\n.d..t...... src/github.com/osism/testbed/scripts/\n<f..t...... src/github.com/osism/testbed/scripts/bootstrap-services.sh\n<f..t...... src/github.com/osism/testbed/scripts/bootstrap.sh\n<f..t...... src/github.com/osism/testbed/scripts/check-ceph.sh\n<f..t...... src/github.com/osism/testbed/scripts/check-services.sh\n<f..t...... src/github.com/osism/testbed/scripts/check.sh\n<f..t...... src/github.com/osism/testbed/scripts/deploy-baremetal.sh\n<f..t...... src/github.com/osism/testbed/scripts/deploy-ceph.sh\n<f..t...... src/github.com/osism/testbed/scripts/deploy-in-a-nutshell.sh\n<f..t...... src/github.com/osism/testbed/scripts/deploy-manager.sh\n<f..t...... src/github.com/osism/testbed/scripts/deploy-services.sh\n<f..t...... src/github.com/osism/testbed/scripts/deploy.sh\n<f..t...... src/github.com/osism/testbed/scripts/disable-ara.sh\n<f..t...... src/github.com/osism/testbed/scripts/disable-local-registry.sh\n<f..t...... src/github.com/osism/testbed/scripts/enable-ara.sh\n<f..t...... src/github.com/osism/testbed/scripts/enable-resource-nodes.sh\n<f..t...... src/github.com/osism/testbed/scripts/include.sh\n<f..t...... src/github.com/osism/testbed/scripts/manager-version.sh\n<f..t...... src/github.com/osism/testbed/scripts/prepare-ceph-configuration.sh\n<f..t...... src/github.com/osism/testbed/scripts/prepare-wireguard-configuration.sh\n<f..t...... src/github.com/osism/testbed/scripts/pull-images.sh\n<f..t...... src/github.com/osism/testbed/scripts/set-ceph-version.sh\n<f..t...... src/github.com/osism/testbed/scripts/set-docker-registry.sh\n<f..t...... src/github.com/osism/testbed/scripts/set-kolla-namespace.sh\n<f..t...... src/github.com/osism/testbed/scripts/set-manager-version.sh\n<f..t...... src/github.com/osism/testbed/scripts/set-openstack-version.sh\n<f..t...... src/github.com/osism/testbed/scripts/sync-configuration-repository.sh\n<f..t...... src/github.com/osism/testbed/scripts/upgrade-ceph.sh\n<f..t...... src/github.com/osism/testbed/scripts/upgrade-manager.sh\n<f..t...... src/github.com/osism/testbed/scripts/upgrade-services.sh\n<f..t...... src/github.com/osism/testbed/scripts/upgrade.sh\n<f..t...... src/github.com/osism/testbed/scripts/wrapper-gilt.sh\n.d..t...... src/github.com/osism/testbed/scripts/bootstrap/\n<f..t...... src/github.com/osism/testbed/scripts/bootstrap/000-netbox.sh\n<f..t...... src/github.com/osism/testbed/scripts/bootstrap/300-openstack.sh\n<f..t...... src/github.com/osism/testbed/scripts/bootstrap/301-openstack-octavia-amhpora-image.sh\n<f..t...... src/github.com/osism/testbed/scripts/bootstrap/302-openstack-k8s-clusterapi-images.sh\n.d..t...... src/github.com/osism/testbed/scripts/check/\n<f..t...... src/github.com/osism/testbed/scripts/check/100-ceph-with-ansible.sh\n<f..t...... src/github.com/osism/testbed/scripts/check/100-ceph-with-rook.sh\n<f..t...... src/github.com/osism/testbed/scripts/check/200-infrastructure.sh\n<f..t...... src/github.com/osism/testbed/scripts/check/300-openstack.sh\n<f..t...... src/github.com/osism/testbed/scripts/check/302-openstack-scs-compatible.sh\n<f..t...... src/github.com/osism/testbed/scripts/check/302-openstack-with-tempest.sh\n<f..t...... src/github.com/osism/testbed/scripts/check/303-prometheus-alert-status.sh\n<f..t...... src/github.com/osism/testbed/scripts/check/500-kubernetes.sh\n.d..t...... src/github.com/osism/testbed/scripts/customisations/\n<f..t...... src/github.com/osism/testbed/scripts/customisations/external-api.sh\n.d..t...... src/github.com/osism/testbed/scripts/deploy/\n<f..t...... src/github.com/osism/testbed/scripts/deploy/000-manager.sh\n<f..t...... src/github.com/osism/testbed/scripts/deploy/001-helpers.sh\n<f..t...... src/github.com/osism/testbed/scripts/deploy/100-ceph-with-ansible.sh\n<f..t...... src/github.com/osism/testbed/scripts/deploy/100-ceph-with-rook.sh\n<f..t...... src/github.com/osism/testbed/scripts/deploy/200-infrastructure.sh\n<f..t...... src/github.com/osism/testbed/scripts/deploy/300-openstack.sh\n<f..t...... src/github.com/osism/testbed/scripts/deploy/310-openstack-extended.sh\n<f..t...... src/github.com/osism/testbed/scripts/deploy/320-openstack-minimal.sh\n<f..t...... src/github.com/osism/testbed/scripts/deploy/400-monitoring.sh\n<f..t...... src/github.com/osism/testbed/scripts/deploy/500-kubernetes.sh\n<f..t...... src/github.com/osism/testbed/scripts/deploy/510-clusterapi.sh\n.d..t...... src/github.com/osism/testbed/scripts/upgrade/\n<f..t...... src/github.com/osism/testbed/scripts/upgrade/100-ceph-with-ansible.sh\n<f..t...... src/github.com/osism/testbed/scripts/upgrade/100-ceph-with-rook.sh\n<f..t...... src/github.com/osism/testbed/scripts/upgrade/200-infrastructure.sh\n<f..t...... src/github.com/osism/testbed/scripts/upgrade/300-openstack.sh\n<f..t...... src/github.com/osism/testbed/scripts/upgrade/310-openstack-extended.sh\n<f..t...... src/github.com/osism/testbed/scripts/upgrade/320-openstack-minimal.sh\n<f..t...... src/github.com/osism/testbed/scripts/upgrade/400-monitoring.sh\n<f..t...... src/github.com/osism/testbed/scripts/upgrade/500-kubernetes.sh\n<f..t...... src/github.com/osism/testbed/scripts/upgrade/510-clusterapi.sh\n.d..t...... src/github.com/osism/testbed/terraform/\n<f..t...... src/github.com/osism/testbed/terraform/Makefile\n<f..t...... src/github.com/osism/testbed/terraform/README.md\n<f..t...... src/github.com/osism/testbed/terraform/backend.env.sample\n<f..t...... src/github.com/osism/testbed/terraform/backend.tf.sample\n<f..t...... src/github.com/osism/testbed/terraform/clouds-public.yaml\n<f..t...... src/github.com/osism/testbed/terraform/clouds.yaml.sample\n<f..t...... src/github.com/osism/testbed/terraform/local.env.sample\n<f..t...... src/github.com/osism/testbed/terraform/secure.yaml.sample\n<f..t...... src/github.com/osism/testbed/terraform/terraformrc.sample\n.d..t...... src/github.com/osism/testbed/terraform/environments/\n<f..t...... src/github.com/osism/testbed/terraform/environments/artcodix.tfvars\n<f..t...... src/github.com/osism/testbed/terraform/environments/cah-f1a.tfvars\n<f..t...... src/github.com/osism/testbed/terraform/environments/ci-centos-stream-9.tfvars\n<f..t...... src/github.com/osism/testbed/terraform/environments/ci-debian-12.tfvars\n<f..t...... src/github.com/osism/testbed/terraform/environments/ci-ubuntu-24.04.tfvars\n<f..t...... src/github.com/osism/testbed/terraform/environments/ci.tfvars\n<f..t...... src/github.com/osism/testbed/terraform/environments/cleura.tfvars\n<f..t...... src/github.com/osism/testbed/terraform/environments/cloudpod.tfvars\n<f..t...... src/github.com/osism/testbed/terraform/environments/default.tfvars\n<f..t...... src/github.com/osism/testbed/terraform/environments/exigo.tfvars\n<f..t...... src/github.com/osism/testbed/terraform/environments/gx-scs.tfvars\n<f..t...... src/github.com/osism/testbed/terraform/environments/huaweicloud.tfvars\n<f..t...... src/github.com/osism/testbed/terraform/environments/otc.tfvars\n<f..t...... src/github.com/osism/testbed/terraform/environments/ovh.tfvars\n<f..t...... src/github.com/osism/testbed/terraform/environments/pluscloudopen.tfvars\n<f..t...... src/github.com/osism/testbed/terraform/environments/regiocloud.tfvars\n<f..t...... src/github.com/osism/testbed/terraform/environments/wavestack.tfvars\n.d..t...... src/github.com/osism/testbed/terraform/scripts/\n<f..t...... src/github.com/osism/testbed/terraform/scripts/cleanup.py\n*deleting   src/github.com/osism/testbed/.git/FETCH_HEAD\n*deleting   src/github.com/osism/testbed/.git/COMMIT_EDITMSG\n*deleting   src/github.com/osism/testbed/.git/refs/zuul/fetch\n*deleting   src/github.com/osism/testbed/.git/refs/zuul/b28b7af69320201d1cf206ebf28373980add1451\n*deleting   src/github.com/osism/testbed/.git/refs/zuul/\n*deleting   src/github.com/osism/testbed/.git/refs/remotes/origin/main\n*deleting   src/github.com/osism/testbed/.git/refs/remotes/origin/\n*deleting   src/github.com/osism/testbed/.git/refs/remotes/\n*deleting   src/github.com/osism/testbed/.git/refs/heads/main\n*deleting   src/github.com/osism/testbed/.git/refs/heads/\n*deleting   src/github.com/osism/testbed/ansible/ansible-manager-part-1.log\n*deleting   src/github.com/osism/testbed/ansible/ansible-manager-part-0.log\n*deleting   src/github.com/osism/testbed/ansible/.id_rsa.ci.pub\n*deleting   src/github.com/osism/testbed/ansible/.id_rsa.ci\n*deleting   src/github.com/osism/testbed/terraform/terraform.tfstate.d/ci/terraform.tfstate\n*deleting   src/github.com/osism/testbed/terraform/terraform.tfstate.d/ci/\n*deleting   src/github.com/osism/testbed/terraform/terraform.tfstate.d/\n*deleting   src/github.com/osism/testbed/terraform/overrides/nodes_use_ephemeral_storage_override.tf\n*deleting   src/github.com/osism/testbed/terraform/overrides/nodes_boot_from_volume_override.tf\n*deleting   src/github.com/osism/testbed/terraform/overrides/nodes_boot_from_image_override.tf\n*deleting   src/github.com/osism/testbed/terraform/overrides/neutron_router_enable_snat_override.tf\n*deleting   src/github.com/osism/testbed/terraform/overrides/neutron_availability_zone_hints_router_override.tf\n*deleting   src/github.com/osism/testbed/terraform/overrides/neutron_availability_zone_hints_network_override.tf\n*deleting   src/github.com/osism/testbed/terraform/overrides/manager_boot_from_volume_override.tf\n*deleting   src/github.com/osism/testbed/terraform/overrides/manager_boot_from_image_override.tf\n*deleting   src/github.com/osism/testbed/terraform/overrides/\n*deleting   src/github.com/osism/testbed/terraform/customisations/neutron_floatingip_custom.tf\n*deleting   src/github.com/osism/testbed/terraform/customisations/external_api_custom.tf\n*deleting   src/github.com/osism/testbed/terraform/customisations/default_custom.tf\n*deleting   src/github.com/osism/testbed/terraform/customisations/access_ipv6_custom.tf\n*deleting   src/github.com/osism/testbed/terraform/customisations/access_ipv4_custom.tf\n*deleting   src/github.com/osism/testbed/terraform/customisations/access_floatingip_custom.tf\n*deleting   src/github.com/osism/testbed/terraform/customisations/\n*deleting   src/github.com/osism/testbed/terraform/.terraform/providers/registry.opentofu.org/terraform-provider-openstack/openstack/3.4.0/linux_amd64/terraform-provider-openstack_v3.4.0\n*deleting   src/github.com/osism/testbed/terraform/.terraform/providers/registry.opentofu.org/terraform-provider-openstack/openstack/3.4.0/linux_amd64/README.md\n*deleting   src/github.com/osism/testbed/terraform/.terraform/providers/registry.opentofu.org/terraform-provider-openstack/openstack/3.4.0/linux_amd64/LICENSE\n*deleting   src/github.com/osism/testbed/terraform/.terraform/providers/registry.opentofu.org/terraform-provider-openstack/openstack/3.4.0/linux_amd64/CHANGELOG.md\n*deleting   src/github.com/osism/testbed/terraform/.terraform/providers/registry.opentofu.org/terraform-provider-openstack/openstack/3.4.0/linux_amd64/\n*deleting   src/github.com/osism/testbed/terraform/.terraform/providers/registry.opentofu.org/terraform-provider-openstack/openstack/3.4.0/linux_amd64.lock\n*deleting   src/github.com/osism/testbed/terraform/.terraform/providers/registry.opentofu.org/terraform-provider-openstack/openstack/3.4.0/\n*deleting   src/github.com/osism/testbed/terraform/.terraform/providers/registry.opentofu.org/terraform-provider-openstack/openstack/\n*deleting   src/github.com/osism/testbed/terraform/.terraform/providers/registry.opentofu.org/terraform-provider-openstack/\n*deleting   src/github.com/osism/testbed/terraform/.terraform/providers/registry.opentofu.org/hashicorp/null/3.2.4/linux_amd64/terraform-provider-null\n*deleting   src/github.com/osism/testbed/terraform/.terraform/providers/registry.opentofu.org/hashicorp/null/3.2.4/linux_amd64/README.md\n*deleting   src/github.com/osism/testbed/terraform/.terraform/providers/registry.opentofu.org/hashicorp/null/3.2.4/linux_amd64/LICENSE\n*deleting   src/github.com/osism/testbed/terraform/.terraform/providers/registry.opentofu.org/hashicorp/null/3.2.4/linux_amd64/CHANGELOG.md\n*deleting   src/github.com/osism/testbed/terraform/.terraform/providers/registry.opentofu.org/hashicorp/null/3.2.4/linux_amd64/\n*deleting   src/github.com/osism/testbed/terraform/.terraform/providers/registry.opentofu.org/hashicorp/null/3.2.4/linux_amd64.lock\n*deleting   src/github.com/osism/testbed/terraform/.terraform/providers/registry.opentofu.org/hashicorp/null/3.2.4/\n*deleting   src/github.com/osism/testbed/terraform/.terraform/providers/registry.opentofu.org/hashicorp/null/\n*deleting   src/github.com/osism/testbed/terraform/.terraform/providers/registry.opentofu.org/hashicorp/local/2.7.0/linux_amd64/terraform-provider-local\n*deleting   src/github.com/osism/testbed/terraform/.terraform/providers/registry.opentofu.org/hashicorp/local/2.7.0/linux_amd64/README.md\n*deleting   src/github.com/osism/testbed/terraform/.terraform/providers/registry.opentofu.org/hashicorp/local/2.7.0/linux_amd64/LICENSE\n*deleting   src/github.com/osism/testbed/terraform/.terraform/providers/registry.opentofu.org/hashicorp/local/2.7.0/linux_amd64/CHANGELOG.md\n*deleting   src/github.com/osism/testbed/terraform/.terraform/providers/registry.opentofu.org/hashicorp/local/2.7.0/linux_amd64/\n*deleting   src/github.com/osism/testbed/terraform/.terraform/providers/registry.opentofu.org/hashicorp/local/2.7.0/linux_amd64.lock\n*deleting   src/github.com/osism/testbed/terraform/.terraform/providers/registry.opentofu.org/hashicorp/local/2.7.0/\n*deleting   src/github.com/osism/testbed/terraform/.terraform/providers/registry.opentofu.org/hashicorp/local/\n*deleting   src/github.com/osism/testbed/terraform/.terraform/providers/registry.opentofu.org/hashicorp/\n*deleting   src/github.com/osism/testbed/terraform/.terraform/providers/registry.opentofu.org/\n*deleting   src/github.com/osism/testbed/terraform/.terraform/providers/\n*deleting   src/github.com/osism/testbed/terraform/.terraform/environment\n*deleting   src/github.com/osism/testbed/terraform/.terraform/\n*deleting   src/github.com/osism/testbed/terraform/variables.tf\n*deleting   src/github.com/osism/testbed/terraform/provider.tf\n*deleting   src/github.com/osism/testbed/terraform/outputs.tf\n*deleting   src/github.com/osism/testbed/terraform/nova.tf\n*deleting   src/github.com/osism/testbed/terraform/nodes_boot_from_volume_override.tf\n*deleting   src/github.com/osism/testbed/terraform/nodes.tf\n*deleting   src/github.com/osism/testbed/terraform/neutron_floatingip_custom.tf\n*deleting   src/github.com/osism/testbed/terraform/neutron_availability_zone_hints_router_override.tf\n*deleting   src/github.com/osism/testbed/terraform/neutron_availability_zone_hints_network_override.tf\n*deleting   src/github.com/osism/testbed/terraform/neutron.tf\n*deleting   src/github.com/osism/testbed/terraform/manager_boot_from_volume_override.tf\n*deleting   src/github.com/osism/testbed/terraform/manager.tf\n*deleting   src/github.com/osism/testbed/terraform/main.tf\n*deleting   src/github.com/osism/testbed/terraform/local.env\n*deleting   src/github.com/osism/testbed/terraform/inventory.ci\n*deleting   src/github.com/osism/testbed/terraform/default_custom.tf\n*deleting   src/github.com/osism/testbed/terraform/data.tf\n*deleting   src/github.com/osism/testbed/terraform/ci.auto.tfvars\n*deleting   src/github.com/osism/testbed/terraform/access_floatingip_custom.tf\n*deleting   src/github.com/osism/testbed/terraform/.terraform.lock.hcl\n*deleting   src/github.com/osism/testbed/terraform/.id_rsa.ci.pub\n*deleting   src/github.com/osism/testbed/terraform/.id_rsa.ci\n*deleting   src/github.com/osism/testbed/terraform/.deploy.ci\n*deleting   src/github.com/osism/testbed/terraform/.MANAGER_ADDRESS.ci\n",
                            "rc": 0,
                            "stdout_lines": [
                                ".d..t...... src/",
                                ".d..t...... src/github.com/",
                                ".d..t...... src/github.com/osism/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/.ansible-lint",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/.flake8",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/.gitignore",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/.yamllint.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/.zuul.yaml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/LICENSE",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/README.md",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/galaxy.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/.git/HEAD",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/.git/ORIG_HEAD",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/.git/config",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/.git/config.worktree",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/.git/description",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/.git/index",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/.git/packed-refs",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/branches/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/hooks/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/.git/hooks/applypatch-msg.sample",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/.git/hooks/commit-msg.sample",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/.git/hooks/fsmonitor-watchman.sample",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/.git/hooks/post-update.sample",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/.git/hooks/pre-applypatch.sample",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/.git/hooks/pre-commit.sample",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/.git/hooks/pre-merge-commit.sample",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/.git/hooks/pre-push.sample",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/.git/hooks/pre-rebase.sample",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/.git/hooks/pre-receive.sample",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/.git/hooks/prepare-commit-msg.sample",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/.git/hooks/push-to-checkout.sample",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/.git/hooks/update.sample",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/info/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/.git/info/exclude",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/logs/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/.git/logs/HEAD",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/logs/refs/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/logs/refs/heads/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/.git/logs/refs/heads/main",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/logs/refs/remotes/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/logs/refs/remotes/origin/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/.git/logs/refs/remotes/origin/HEAD",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/01/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/02/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/04/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/05/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/06/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/07/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/08/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/09/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/0a/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/0b/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/0c/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/0d/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/0e/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/0f/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/10/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/11/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/12/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/13/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/14/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/15/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/16/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/17/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/18/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/19/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/1a/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/1b/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/1d/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/1f/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/20/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/21/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/22/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/23/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/24/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/25/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/26/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/27/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/28/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/29/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/2a/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/2b/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/2c/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/2e/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/2f/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/30/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/31/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/32/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/33/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/35/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/36/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/37/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/38/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/39/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/3a/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/3b/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/3c/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/3d/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/3e/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/40/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/41/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/42/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/43/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/44/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/45/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/46/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/48/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/49/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/4a/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/4b/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/4c/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/4d/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/4e/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/4f/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/50/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/52/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/53/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/54/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/55/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/56/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/57/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/58/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/59/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/5a/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/5b/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/5c/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/5d/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/5e/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/5f/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/60/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/61/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/62/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/63/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/64/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/66/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/67/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/68/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/69/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/6a/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/6b/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/6c/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/6d/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/6e/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/70/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/71/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/72/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/73/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/74/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/75/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/76/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/77/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/78/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/79/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/7a/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/7b/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/7c/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/7d/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/7f/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/80/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/81/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/82/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/83/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/84/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/85/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/86/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/87/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/88/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/89/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/8a/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/8b/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/8c/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/8d/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/8e/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/8f/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/90/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/91/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/93/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/94/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/95/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/96/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/97/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/98/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/9a/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/9b/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/9c/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/9f/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/a0/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/a1/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/a2/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/a3/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/a4/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/a5/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/a6/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/a7/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/a8/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/a9/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/aa/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/ab/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/ac/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/ad/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/ae/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/af/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/b0/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/b1/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/b2/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/b3/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/b4/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/b5/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/b6/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/b7/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/b8/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/b9/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/ba/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/bb/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/bc/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/bd/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/be/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/bf/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/c0/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/c3/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/c4/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/c5/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/c6/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/c7/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/c9/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/ca/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/cb/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/cc/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/ce/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/cf/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/d0/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/d1/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/d2/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/d3/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/d4/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/d5/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/d6/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/d8/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/d9/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/da/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/db/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/dc/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/de/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/df/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/e0/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/e2/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/e3/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/e4/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/e5/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/e7/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/e8/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/e9/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/ea/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/eb/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/ec/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/ed/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/ee/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/ef/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/f0/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/f2/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/f3/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/f4/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/f5/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/f6/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/f7/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/f8/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/f9/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/fa/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/fb/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/fc/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/fd/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/fe/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/ff/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/info/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/objects/pack/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.git/refs/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.github/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/.github/publish-collection.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/.github/renovate.json",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.github/templates/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/.github/templates/galaxy.yml.j2",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/.github/workflows/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/.github/workflows/publish-collection.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/meta/runtime.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/molecule/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/requirements.txt",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/collections.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/converge.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/molecule.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/certificates.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/cleanup.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/configfs.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/configuration.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/docker_compose.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/docker_login.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/facts.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/firewall.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/hostname.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/hosts.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/ipmitool.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/k9s.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/kernel_modules.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/known_hosts.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/limits.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/lynis.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/microcode.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/motd.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/network.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/operator.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/packages.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/podman.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/proxy.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/repository.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/resolvconf.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/runc.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/services.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/sosreport.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/sshconfig.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/state.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/sysctl.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/systohc.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/timezone.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/trivy.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/prepare/user.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/__init__.py",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/certificates.py",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/configfs.py",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/docker_compose.py",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/docker_login.py",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/facts.py",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/firewall.py",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/hostname.py",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/ipmitool.py",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/k9s.py",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/kernel_modules.py",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/known_hosts.py",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/limits.py",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/microcode.py",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/operator.py",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/resolvconf.py",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/runc.py",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/services.py",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/sosreport.py",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/sshconfig.py",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/state.py",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/sysctl.py",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/systohc.py",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/user.py",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/cleanup/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/cleanup/__init__.py",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/cleanup/cloudinit.py",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/cleanup/packages.py",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/cleanup/packages_debian.py",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/cleanup/services.py",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/cleanup/timer_debian.py",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/cleanup/timer_redhat.py",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/configuration/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/configuration/__init__.py",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/configuration/git.py",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/configuration/main.py",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/hosts/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/hosts/__init__.py",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/hosts/type-block.py",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/hosts/type-template.py",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/lynis/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/lynis/__init__.py",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/lynis/debian.py",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/lynis/main.py",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/lynis/redhat.py",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/motd/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/motd/__init__.py",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/motd/debian.py",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/motd/main.py",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/motd/redhat.py",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/network/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/network/__init__.py",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/network/interfaces.py",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/network/netplan.py",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/packages/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/packages/__init__.py",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/packages/debian.py",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/packages/main.py",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/packages/redhat.py",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/podman/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/podman/__init__.py",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/podman/debian.py",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/podman/main.py",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/proxy/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/proxy/__init__.py",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/proxy/debian.py",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/proxy/main.py",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/proxy/redhat.py",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/repository/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/repository/__init__.py",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/repository/centos.py",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/repository/debian.py",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/timezone/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/timezone/__init__.py",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/timezone/debian.py",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/timezone/main.py",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/timezone/redhat.py",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/trivy/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/trivy/__init__.py",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/trivy/debian.py",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/trivy/main.py",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/trivy/redhat.py",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/util/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/util/__init__.py",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/tests/util/util.py",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/certificates.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/cleanup.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/configfs.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/configuration.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/docker_compose.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/docker_login.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/facts.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/firewall.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/hostname.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/hosts.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/ipmitool.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/k9s.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/kernel_modules.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/known_hosts.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/limits.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/lynis.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/microcode.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/motd.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/network.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/operator.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/packages.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/podman.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/proxy.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/repository.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/resolvconf.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/runc.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/services.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/sosreport.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/sshconfig.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/state.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/sysctl.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/systohc.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/timezone.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/trivy.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/vars/user.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/verify/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/verify/certificates.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/verify/cleanup.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/verify/configfs.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/verify/configuration.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/verify/docker_compose.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/verify/facts.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/verify/firewall.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/verify/hostname.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/verify/hosts.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/verify/ipmitool.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/verify/k9s.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/verify/kernel_modules.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/verify/lynis.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/verify/microcode.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/verify/motd.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/verify/operator.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/verify/packages.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/verify/podman.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/verify/proxy.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/verify/repository.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/verify/resolvconf.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/verify/runc.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/verify/services.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/verify/sosreport.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/verify/sshconfig.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/verify/state.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/verify/sysctl.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/verify/systohc.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/verify/timezone.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/molecule/delegated/verify/trivy.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/plugins/",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/plugins/callback/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/plugins/callback/still_alive.py",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/certificates/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/certificates/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/certificates/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/certificates/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/certificates/handlers/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/certificates/handlers/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/certificates/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/certificates/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/certificates/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/certificates/tasks/install-Debian-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/certificates/tasks/install-RedHat-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/certificates/tasks/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/certificates/templates/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/certificates/templates/certificates.j2",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/certificates/vars/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/certificates/vars/Debian-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/certificates/vars/RedHat-family.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/cleanup/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/cleanup/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/cleanup/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/cleanup/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/cleanup/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/cleanup/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/cleanup/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/cleanup/tasks/cloudinit.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/cleanup/tasks/main.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/cleanup/tasks/packages-Debian-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/cleanup/tasks/packages-RedHat-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/cleanup/tasks/services-Debian-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/cleanup/tasks/services-RedHat-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/cleanup/tasks/timers-Debian-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/cleanup/tasks/timers-RedHat-family.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/cleanup/vars/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/cleanup/vars/Debian-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/cleanup/vars/RedHat-family.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/configfs/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/configfs/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/configfs/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/configfs/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/configfs/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/configfs/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/configfs/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/configfs/tasks/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/configuration/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/configuration/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/configuration/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/configuration/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/configuration/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/configuration/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/configuration/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/configuration/tasks/git.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/configuration/tasks/main.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/configuration/tasks/netbox.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/configuration/tasks/update.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/configuration/templates/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/configuration/templates/config-git-netbox-private-key.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/configuration/templates/config-git-private-key.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/configuration/templates/config.j2",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/docker_compose/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/docker_compose/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/docker_compose/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/docker_compose/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/docker_compose/files/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/docker_compose/files/docker-compose",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/docker_compose/handlers/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/docker_compose/handlers/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/docker_compose/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/docker_compose/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/docker_compose/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/docker_compose/tasks/install-Debian-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/docker_compose/tasks/install-RedHat-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/docker_compose/tasks/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/docker_compose/templates/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/docker_compose/templates/docker-compose.service.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/docker_compose/templates/osism.target.j2",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/docker_login/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/docker_login/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/docker_login/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/docker_login/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/docker_login/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/docker_login/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/docker_login/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/docker_login/tasks/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/facts/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/facts/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/facts/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/facts/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/facts/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/facts/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/facts/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/facts/tasks/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/firewall/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/firewall/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/firewall/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/firewall/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/firewall/handlers/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/firewall/handlers/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/firewall/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/firewall/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/firewall/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/firewall/tasks/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/firewall/templates/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/firewall/templates/ufw.conf.j2",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/hostname/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/hostname/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/hostname/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/hostname/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/hostname/handlers/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/hostname/handlers/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/hostname/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/hostname/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/hostname/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/hostname/tasks/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/hostname/templates/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/hostname/templates/config.j2",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/hosts/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/hosts/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/hosts/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/hosts/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/hosts/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/hosts/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/hosts/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/hosts/tasks/main.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/hosts/tasks/type-block.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/hosts/tasks/type-local.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/hosts/tasks/type-template.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/hosts/templates/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/hosts/templates/hosts-block.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/hosts/templates/hosts-template.j2",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/ipmitool/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/ipmitool/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/ipmitool/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/ipmitool/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/ipmitool/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/ipmitool/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/ipmitool/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/ipmitool/tasks/install-Debian-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/ipmitool/tasks/install-RedHat-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/ipmitool/tasks/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/ipmitool/templates/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/ipmitool/templates/module-load.conf.j2",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/kernel_modules/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/kernel_modules/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/kernel_modules/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/kernel_modules/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/kernel_modules/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/kernel_modules/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/kernel_modules/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/kernel_modules/tasks/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/known_hosts/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/known_hosts/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/known_hosts/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/known_hosts/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/known_hosts/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/known_hosts/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/known_hosts/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/known_hosts/tasks/main.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/known_hosts/tasks/write-scanned.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/known_hosts/tasks/write-static.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/limits/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/limits/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/limits/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/limits/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/limits/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/limits/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/limits/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/limits/tasks/limits.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/limits/tasks/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/lynis/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/lynis/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/lynis/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/lynis/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/lynis/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/lynis/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/lynis/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/lynis/tasks/install-Debian-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/lynis/tasks/install-RedHat-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/lynis/tasks/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/microcode/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/microcode/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/microcode/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/microcode/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/microcode/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/microcode/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/microcode/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/microcode/tasks/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/motd/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/motd/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/motd/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/motd/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/motd/handlers/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/motd/handlers/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/motd/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/motd/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/motd/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/motd/tasks/configure-Debian-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/motd/tasks/configure-RedHat-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/motd/tasks/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/motd/templates/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/motd/templates/motd.j2",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/network/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/network/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/network/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/network/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/network/handlers/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/network/handlers/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/network/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/network/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/network/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/network/tasks/cleanup-interfaces.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/network/tasks/cleanup-netplan.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/network/tasks/cleanup-networkd.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/network/tasks/dummy-interfaces.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/network/tasks/interfaces-Debian-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/network/tasks/interfaces-RedHat-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/network/tasks/main.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/network/tasks/netplan-Debian-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/network/tasks/netplan-RedHat-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/network/tasks/network-extra-init.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/network/tasks/vxlan-interfaces.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/network/templates/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/network/templates/dummy.netdev.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/network/templates/network-extra-init.service.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/network/templates/network-extra-init.sh.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/network/templates/vxlan.netdev.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/network/templates/vxlan.network.j2",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/network/templates/interfaces/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/network/templates/interfaces/device.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/network/templates/interfaces/interfaces.j2",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/network/templates/netplan/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/network/templates/netplan/01-osism.yaml.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/network/templates/netplan/interfaces.j2",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/network/vars/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/network/vars/Debian-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/network/vars/RedHat-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/network/vars/interfaces.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/operator/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/operator/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/operator/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/operator/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/operator/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/operator/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/operator/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/operator/tasks/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/operator/templates/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/operator/templates/operator-sudoers.j2",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/operator/vars/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/operator/vars/Debian-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/operator/vars/RedHat-family.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/packages/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/packages/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/packages/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/packages/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/packages/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/packages/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/packages/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/packages/tasks/main.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/packages/tasks/package-Debian-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/packages/tasks/package-RedHat-family.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/packages/vars/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/packages/vars/Debian-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/packages/vars/RedHat-family.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/podman/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/podman/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/podman/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/podman/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/podman/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/podman/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/podman/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/podman/tasks/config.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/podman/tasks/deploy.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/podman/tasks/install-Debian-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/podman/tasks/install-RedHat-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/podman/tasks/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/proxy/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/proxy/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/proxy/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/proxy/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/proxy/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/proxy/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/proxy/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/proxy/tasks/Debian-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/proxy/tasks/RedHat-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/proxy/tasks/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/proxy/templates/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/proxy/templates/proxy-apt.j2",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/repository/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/repository/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/repository/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/repository/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/repository/handlers/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/repository/handlers/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/repository/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/repository/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/repository/tasks/",
                                ".L..t...... src/github.com/osism/ansible-collection-commons/roles/repository/tasks/AlmaLinux.yml -> RedHat.yml",
                                ".L..t...... src/github.com/osism/ansible-collection-commons/roles/repository/tasks/CentOS.yml -> RedHat.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/repository/tasks/Debian.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/repository/tasks/RedHat.yml",
                                ".L..t...... src/github.com/osism/ansible-collection-commons/roles/repository/tasks/Rocky.yml -> RedHat.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/repository/tasks/Ubuntu.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/repository/tasks/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/repository/templates/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/repository/templates/99osism.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/repository/templates/source.list.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/repository/templates/ubuntu.sources.j2",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/repository/vars/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/repository/vars/AlmaLinux-amd64.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/repository/vars/CentOS-x86_64.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/repository/vars/Debian-x86_64.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/repository/vars/Rocky-x86_64.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/repository/vars/Ubuntu-aarch64.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/repository/vars/Ubuntu-x86_64.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/resolvconf/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/resolvconf/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/resolvconf/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/resolvconf/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/resolvconf/handlers/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/resolvconf/handlers/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/resolvconf/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/resolvconf/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/resolvconf/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/resolvconf/tasks/configure-Debian-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/resolvconf/tasks/configure-RedHat-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/resolvconf/tasks/configure-resolv.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/resolvconf/tasks/install-Debian-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/resolvconf/tasks/install-RedHat-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/resolvconf/tasks/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/resolvconf/templates/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/resolvconf/templates/resolved.conf.j2",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/runc/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/runc/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/runc/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/runc/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/runc/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/runc/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/runc/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/runc/tasks/install-Debian-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/runc/tasks/install-RedHat-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/runc/tasks/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/services/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/services/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/services/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/services/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/services/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/services/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/services/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/services/tasks/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/sosreport/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/sosreport/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/sosreport/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/sosreport/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/sosreport/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/sosreport/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/sosreport/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/sosreport/tasks/install-Debian-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/sosreport/tasks/install-RedHat-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/sosreport/tasks/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/sosreport/vars/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/sosreport/vars/CentOS-dist.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/sosreport/vars/Debian-dist.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/sosreport/vars/Ubuntu-dist.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/sshconfig/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/sshconfig/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/sshconfig/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/sshconfig/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/sshconfig/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/sshconfig/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/sshconfig/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/sshconfig/tasks/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/sshconfig/templates/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/sshconfig/templates/config.j2",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/state/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/state/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/state/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/state/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/state/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/state/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/state/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/state/tasks/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/sysctl/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/sysctl/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/sysctl/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/sysctl/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/sysctl/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/sysctl/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/sysctl/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/sysctl/tasks/main.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/sysctl/tasks/sysctl.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/systohc/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/systohc/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/systohc/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/systohc/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/systohc/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/systohc/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/systohc/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/systohc/tasks/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/timezone/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/timezone/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/timezone/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/timezone/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/timezone/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/timezone/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/timezone/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/timezone/tasks/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/trivy/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/trivy/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/trivy/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/trivy/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/trivy/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/trivy/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/trivy/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/trivy/tasks/install-Debian-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/trivy/tasks/install-RedHat-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/trivy/tasks/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/user/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/user/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/user/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/user/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/user/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/user/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/user/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/user/tasks/main.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/user/tasks/remote-key.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/user/tasks/type-default.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/user/tasks/type-keyfile.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-commons/roles/user/templates/",
                                "<f..t...... src/github.com/osism/ansible-collection-commons/roles/user/templates/config.j2",
                                ".d..t...... src/github.com/osism/ansible-collection-services/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/.ansible-lint",
                                "<f..t...... src/github.com/osism/ansible-collection-services/.flake8",
                                "<f..t...... src/github.com/osism/ansible-collection-services/.gitignore",
                                "<f..t...... src/github.com/osism/ansible-collection-services/.yamllint.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/.zuul.yaml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/LICENSE",
                                "<f..t...... src/github.com/osism/ansible-collection-services/README.md",
                                "<f..t...... src/github.com/osism/ansible-collection-services/galaxy.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/.git/HEAD",
                                "<f..t...... src/github.com/osism/ansible-collection-services/.git/ORIG_HEAD",
                                "<f..t...... src/github.com/osism/ansible-collection-services/.git/config",
                                "<f..t...... src/github.com/osism/ansible-collection-services/.git/config.worktree",
                                "<f..t...... src/github.com/osism/ansible-collection-services/.git/description",
                                "<f..t...... src/github.com/osism/ansible-collection-services/.git/index",
                                "<f..t...... src/github.com/osism/ansible-collection-services/.git/packed-refs",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/branches/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/hooks/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/.git/hooks/applypatch-msg.sample",
                                "<f..t...... src/github.com/osism/ansible-collection-services/.git/hooks/commit-msg.sample",
                                "<f..t...... src/github.com/osism/ansible-collection-services/.git/hooks/fsmonitor-watchman.sample",
                                "<f..t...... src/github.com/osism/ansible-collection-services/.git/hooks/post-update.sample",
                                "<f..t...... src/github.com/osism/ansible-collection-services/.git/hooks/pre-applypatch.sample",
                                "<f..t...... src/github.com/osism/ansible-collection-services/.git/hooks/pre-commit.sample",
                                "<f..t...... src/github.com/osism/ansible-collection-services/.git/hooks/pre-merge-commit.sample",
                                "<f..t...... src/github.com/osism/ansible-collection-services/.git/hooks/pre-push.sample",
                                "<f..t...... src/github.com/osism/ansible-collection-services/.git/hooks/pre-rebase.sample",
                                "<f..t...... src/github.com/osism/ansible-collection-services/.git/hooks/pre-receive.sample",
                                "<f..t...... src/github.com/osism/ansible-collection-services/.git/hooks/prepare-commit-msg.sample",
                                "<f..t...... src/github.com/osism/ansible-collection-services/.git/hooks/push-to-checkout.sample",
                                "<f..t...... src/github.com/osism/ansible-collection-services/.git/hooks/update.sample",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/info/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/.git/info/exclude",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/logs/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/.git/logs/HEAD",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/logs/refs/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/logs/refs/heads/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/.git/logs/refs/heads/main",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/logs/refs/remotes/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/logs/refs/remotes/origin/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/.git/logs/refs/remotes/origin/HEAD",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/",
                                "cd+++++++++ src/github.com/osism/ansible-collection-services/.git/objects/07/",
                                "<f+++++++++ src/github.com/osism/ansible-collection-services/.git/objects/07/abeb3fe45e34bb6147f840abb924c456658985",
                                "cd+++++++++ src/github.com/osism/ansible-collection-services/.git/objects/0a/",
                                "<f+++++++++ src/github.com/osism/ansible-collection-services/.git/objects/0a/d6892495184ac5a0905065ee7b8f59b58e11ed",
                                "cd+++++++++ src/github.com/osism/ansible-collection-services/.git/objects/0b/",
                                "<f+++++++++ src/github.com/osism/ansible-collection-services/.git/objects/0b/cc82939de812d421ce7dbd024b2530ddaf7373",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/0c/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/10/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/1a/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/20/",
                                "cd+++++++++ src/github.com/osism/ansible-collection-services/.git/objects/21/",
                                "<f+++++++++ src/github.com/osism/ansible-collection-services/.git/objects/21/cf74b32ae0d8cff16a471330421627a77f059d",
                                "cd+++++++++ src/github.com/osism/ansible-collection-services/.git/objects/23/",
                                "<f+++++++++ src/github.com/osism/ansible-collection-services/.git/objects/23/f0e755006ad195b7a2720e2160a4596cace0a0",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/29/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/2b/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/2d/",
                                "cd+++++++++ src/github.com/osism/ansible-collection-services/.git/objects/32/",
                                "<f+++++++++ src/github.com/osism/ansible-collection-services/.git/objects/32/5879d45377e2eb83e8e826a04168d8a18585c7",
                                "<f+++++++++ src/github.com/osism/ansible-collection-services/.git/objects/32/e3c0272e42e39c1b5073a3e99c0da8a2ddf038",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/33/",
                                "cd+++++++++ src/github.com/osism/ansible-collection-services/.git/objects/4a/",
                                "<f+++++++++ src/github.com/osism/ansible-collection-services/.git/objects/4a/0f26b3f2bb9570732719dca581d86c90dd7728",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/4f/",
                                "cd+++++++++ src/github.com/osism/ansible-collection-services/.git/objects/50/",
                                "<f+++++++++ src/github.com/osism/ansible-collection-services/.git/objects/50/4e662b4528f0145dd717c4f9c61aacdd466d2f",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/51/",
                                "cd+++++++++ src/github.com/osism/ansible-collection-services/.git/objects/55/",
                                "<f+++++++++ src/github.com/osism/ansible-collection-services/.git/objects/55/a2f430ce0c434d2c0c118608248b8e1814780a",
                                "cd+++++++++ src/github.com/osism/ansible-collection-services/.git/objects/59/",
                                "<f+++++++++ src/github.com/osism/ansible-collection-services/.git/objects/59/a248cb3e90c79960e8e660f5409daec7d03aca",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/66/",
                                "cd+++++++++ src/github.com/osism/ansible-collection-services/.git/objects/6c/",
                                "<f+++++++++ src/github.com/osism/ansible-collection-services/.git/objects/6c/81ed60282f5fbaefb6b19bed89b8ff9dde0479",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/72/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/7b/",
                                "cd+++++++++ src/github.com/osism/ansible-collection-services/.git/objects/82/",
                                "<f+++++++++ src/github.com/osism/ansible-collection-services/.git/objects/82/b8eb8b6a299a32fbe9d73f4997a28d7b6f9115",
                                "cd+++++++++ src/github.com/osism/ansible-collection-services/.git/objects/85/",
                                "<f+++++++++ src/github.com/osism/ansible-collection-services/.git/objects/85/b6ddd5e7addd50a6f504aa01af44f4956e23db",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/87/",
                                "cd+++++++++ src/github.com/osism/ansible-collection-services/.git/objects/8b/",
                                "<f+++++++++ src/github.com/osism/ansible-collection-services/.git/objects/8b/e9c887dd0903883b40bebfe6f91660cbd657a5",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/95/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/b2/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/d0/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/d7/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/dc/",
                                "cd+++++++++ src/github.com/osism/ansible-collection-services/.git/objects/dd/",
                                "<f+++++++++ src/github.com/osism/ansible-collection-services/.git/objects/dd/601554d21714fa353d94d36c26bc38ed71830b",
                                "cd+++++++++ src/github.com/osism/ansible-collection-services/.git/objects/e0/",
                                "<f+++++++++ src/github.com/osism/ansible-collection-services/.git/objects/e0/93ad7eaa58f0394d242749dbf97e648bd24afe",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/e5/",
                                "cd+++++++++ src/github.com/osism/ansible-collection-services/.git/objects/e9/",
                                "<f+++++++++ src/github.com/osism/ansible-collection-services/.git/objects/e9/159510cf7ca0393b8c2cd8ea649f1fb518fc26",
                                "cd+++++++++ src/github.com/osism/ansible-collection-services/.git/objects/ea/",
                                "<f+++++++++ src/github.com/osism/ansible-collection-services/.git/objects/ea/16ee541431840f00f27f68248afc3b2f69a869",
                                "cd+++++++++ src/github.com/osism/ansible-collection-services/.git/objects/eb/",
                                "<f+++++++++ src/github.com/osism/ansible-collection-services/.git/objects/eb/817d89ecab11ed060682c7912e3edf8bdb641e",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/info/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/objects/pack/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.git/refs/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.github/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/.github/publish-collection.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/.github/renovate.json",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.github/templates/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/.github/templates/galaxy.yml.j2",
                                ".d..t...... src/github.com/osism/ansible-collection-services/.github/workflows/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/.github/workflows/publish-collection.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/meta/runtime.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/molecule/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/requirements.txt",
                                ".d..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/collections.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/converge.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/molecule.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/adminer.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/auditd.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/cephclient.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/cgit.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/chrony.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/clamav.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/containerd.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/dnsdist.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/dnsmasq.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/docker.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/fail2ban.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/falco.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/frr.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/gnmic.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/hddtemp.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/homer.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/httpd.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/journald.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/lldpd.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/manager.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/netbird.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/netbox.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/netdata.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/nexus.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/openstackclient.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/opentelemetry_collector.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/osquery.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/phpmyadmin.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/rng.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/rsyslog.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/scaphandre.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/smartd.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/squid.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/sshd.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/stepca.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/substation.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/teleport.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/thanos_sidecar.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/traefik.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/tuned.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/wazuh_agent.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/wireguard.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/zabbix_agent.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/prepare/zuul.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/__init__.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/adminer.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/auditd.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/cgit.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/dnsdist.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/dnsmasq.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/fail2ban.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/frr.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/gnmic.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/homer.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/httpd.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/journald.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/lldpd.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/manager.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/netbox.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/nexus.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/opentelemetry_collector.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/phpmyadmin.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/rng.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/rsyslog.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/scaphandre.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/smartd.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/squid.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/sshd.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/stepca.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/substation.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/thanos_sidecar.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/traefik.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/tuned.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/wireguard.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/zuul.py",
                                ".d..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/cephclient/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/cephclient/__init__.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/cephclient/container.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/cephclient/main.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/cephclient/package.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/cephclient/package_debian.py",
                                ".d..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/chrony/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/chrony/__init__.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/chrony/debian.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/chrony/main.py",
                                ".d..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/clamav/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/clamav/__init__.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/clamav/debian.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/clamav/main.py",
                                ".d..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/containerd/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/containerd/__init__.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/containerd/debian.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/containerd/main.py",
                                ".d..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/docker/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/docker/__init__.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/docker/debian.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/docker/main.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/docker/redhat.py",
                                ".d..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/falco/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/falco/__init__.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/falco/debian.py",
                                ".d..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/hddtemp/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/hddtemp/__init__.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/hddtemp/debian.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/hddtemp/redhat.py",
                                ".d..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/netbird/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/netbird/__init__.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/netbird/debian.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/netbird/redhat.py",
                                ".d..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/netdata/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/netdata/__init__.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/netdata/debian.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/netdata/main.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/netdata/redhat.py",
                                ".d..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/openstackclient/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/openstackclient/__init__.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/openstackclient/container.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/openstackclient/main.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/openstackclient/package-debian.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/openstackclient/package-redhat.py",
                                ".d..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/osquery/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/osquery/__init__.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/osquery/debian.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/osquery/main.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/osquery/redhat.py",
                                ".d..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/teleport/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/teleport/__init__.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/teleport/debian.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/teleport/main.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/teleport/redhat.py",
                                ".d..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/util/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/util/__init__.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/util/util.py",
                                ".d..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/wazuh_agent/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/wazuh_agent/__init__.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/wazuh_agent/debian.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/wazuh_agent/main.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/wazuh_agent/redhat.py",
                                ".d..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/zabbix_agent/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/zabbix_agent/__init__.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/zabbix_agent/debian.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/zabbix_agent/main.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/tests/zabbix_agent/redhat.py",
                                ".d..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/adminer.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/auditd.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/cephclient.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/cgit.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/chrony.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/clamav.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/containerd.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/dnsdist.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/dnsmasq.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/docker.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/fail2ban.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/falco.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/frr.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/gnmic.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/hddtemp.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/homer.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/httpd.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/journald.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/lldpd.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/manager.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/netbird.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/netbox.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/netdata.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/nexus.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/openstackclient.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/opentelemetry_collector.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/osquery.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/phpmyadmin.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/rng.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/rsyslog.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/scaphandre.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/smartd.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/squid.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/sshd.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/stepca.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/substation.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/teleport.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/thanos_sidecar.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/traefik.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/tuned.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/wazuh_agent.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/wireguard.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/zabbix_agent.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/vars/zuul.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/adminer.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/auditd.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/cephclient.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/cgit.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/chrony.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/clamav.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/containerd.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/dnsdist.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/dnsmasq.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/docker.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/fail2ban.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/falco.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/frr.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/gnmic.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/hddtemp.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/homer.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/httpd.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/journald.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/lldpd.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/netbird.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/netdata.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/osquery.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/rng.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/rsyslog.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/smartd.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/sshd.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/teleport.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/tuned.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/wazuh_agent.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/wireguard.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/molecule/delegated/verify/zabbix_agent.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/plugins/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/plugins/filter/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/plugins/filter/address.py",
                                ".d..t...... src/github.com/osism/ansible-collection-services/plugins/modules/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/plugins/modules/kolla_container_facts.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/plugins/modules/kolla_docker.py",
                                "<f..t...... src/github.com/osism/ansible-collection-services/plugins/modules/kolla_toolbox.py",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/adminer/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/adminer/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/adminer/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/adminer/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/adminer/handlers/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/adminer/handlers/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/adminer/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/adminer/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/adminer/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/adminer/tasks/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/adminer/templates/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/adminer/templates/docker-compose.yml.j2",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/auditd/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/auditd/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/auditd/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/auditd/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/auditd/handlers/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/auditd/handlers/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/auditd/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/auditd/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/auditd/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/auditd/tasks/install-Debian-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/auditd/tasks/install-RedHat-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/auditd/tasks/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/auditd/templates/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/auditd/templates/auditd.conf.j2",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/auditd/templates/rules/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/auditd/templates/rules/20-neo23x0.rules.j2",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/cephclient/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/cephclient/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/cephclient/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/cephclient/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/cephclient/handlers/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/cephclient/handlers/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/cephclient/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/cephclient/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/cephclient/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/cephclient/tasks/clean-container.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/cephclient/tasks/clean-package-Debian-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/cephclient/tasks/clean-package-RedHat-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/cephclient/tasks/container.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/cephclient/tasks/main.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/cephclient/tasks/package-Debian-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/cephclient/tasks/package-RedHat-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/cephclient/tasks/rook.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/cephclient/templates/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/cephclient/templates/ceph.conf.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/cephclient/templates/docker-compose.yml.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/cephclient/templates/keyring.j2",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/cephclient/templates/wrapper/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/cephclient/templates/wrapper/ceph-authtool.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/cephclient/templates/wrapper/ceph.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/cephclient/templates/wrapper/crushtool.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/cephclient/templates/wrapper/rados.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/cephclient/templates/wrapper/radosgw-admin.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/cephclient/templates/wrapper/rbd.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/cephclient/templates/wrapper/rook.j2",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/cgit/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/cgit/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/cgit/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/cgit/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/cgit/handlers/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/cgit/handlers/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/cgit/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/cgit/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/cgit/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/cgit/tasks/config.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/cgit/tasks/main.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/cgit/tasks/service.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/cgit/templates/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/cgit/templates/cgitrc.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/cgit/templates/docker-compose.yml.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/cgit/templates/httpd-cgit.conf.j2",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/chrony/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/chrony/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/chrony/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/chrony/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/chrony/handlers/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/chrony/handlers/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/chrony/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/chrony/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/chrony/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/chrony/tasks/install-Debian-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/chrony/tasks/install-RedHat-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/chrony/tasks/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/chrony/templates/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/chrony/templates/chrony.conf.j2",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/chrony/vars/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/chrony/vars/Debian-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/chrony/vars/RedHat-family.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/clamav/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/clamav/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/clamav/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/clamav/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/clamav/handlers/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/clamav/handlers/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/clamav/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/clamav/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/clamav/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/clamav/tasks/initialize.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/clamav/tasks/install-Debian-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/clamav/tasks/install-RedHat-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/clamav/tasks/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/clamav/vars/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/clamav/vars/Debian-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/clamav/vars/RedHat-family.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/containerd/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/containerd/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/containerd/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/containerd/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/containerd/handlers/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/containerd/handlers/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/containerd/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/containerd/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/containerd/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/containerd/tasks/config.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/containerd/tasks/install-Debian-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/containerd/tasks/install-RedHat-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/containerd/tasks/main.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/containerd/tasks/service.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/containerd/templates/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/containerd/templates/config.toml.j2",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/containerd/vars/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/containerd/vars/CentOS-dist.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/containerd/vars/Debian-dist.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/containerd/vars/Ubuntu-dist.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/dnsdist/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/dnsdist/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/dnsdist/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/dnsdist/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/dnsdist/handlers/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/dnsdist/handlers/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/dnsdist/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/dnsdist/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/dnsdist/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/dnsdist/tasks/config.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/dnsdist/tasks/main.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/dnsdist/tasks/service.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/dnsdist/templates/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/dnsdist/templates/dnsdist.conf.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/dnsdist/templates/docker-compose.yml.j2",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/dnsmasq/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/dnsmasq/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/dnsmasq/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/dnsmasq/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/dnsmasq/handlers/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/dnsmasq/handlers/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/dnsmasq/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/dnsmasq/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/dnsmasq/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/dnsmasq/tasks/config.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/dnsmasq/tasks/main.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/dnsmasq/tasks/service.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/dnsmasq/templates/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/dnsmasq/templates/dnsmasq.conf.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/dnsmasq/templates/docker-compose.yml.j2",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/docker/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/docker/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/docker/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/docker/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/docker/files/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/docker/files/.gitattributes",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/docker/files/limits.conf",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/docker/files/facts/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/docker/files/facts/docker_containers.fact",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/docker/files/facts/docker_images.fact",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/docker/handlers/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/docker/handlers/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/docker/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/docker/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/docker/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/docker/tasks/bootstrap.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/docker/tasks/config.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/docker/tasks/facts.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/docker/tasks/install-docker-Debian-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/docker/tasks/install-docker-RedHat-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/docker/tasks/main.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/docker/tasks/service.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/docker/tasks/storage-block.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/docker/tasks/storage-zram.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/docker/templates/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/docker/templates/config.json.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/docker/templates/containerd_config.toml.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/docker/templates/daemon.json.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/docker/templates/overlay.conf.j2",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/docker/vars/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/docker/vars/Debian-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/docker/vars/RedHat-family.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/fail2ban/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/fail2ban/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/fail2ban/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/fail2ban/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/fail2ban/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/fail2ban/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/fail2ban/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/fail2ban/tasks/install-Debian-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/fail2ban/tasks/install-RedHat-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/fail2ban/tasks/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/fail2ban/templates/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/fail2ban/templates/sshd.conf.j2",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/falco/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/falco/.gitignore",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/falco/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/falco/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/falco/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/falco/files/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/falco/files/falco.service",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/falco/handlers/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/falco/handlers/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/falco/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/falco/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/falco/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/falco/tasks/install-Debian-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/falco/tasks/install-RedHat-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/falco/tasks/install-RedHat-family_broken.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/falco/tasks/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/frr/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/frr/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/frr/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/frr/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/frr/handlers/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/frr/handlers/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/frr/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/frr/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/frr/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/frr/tasks/install-Debian-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/frr/tasks/install-RedHat-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/frr/tasks/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/frr/templates/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/frr/templates/daemons.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/frr/templates/frr_k3s_cilium.conf.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/frr/templates/frr_leaf.conf.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/frr/templates/frr_loadbalancer.conf.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/frr/templates/frr_loadbalancer_external_uplink.conf.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/frr/templates/frr_test.conf.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/frr/templates/vtysh.conf.j2",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/gnmic/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/gnmic/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/gnmic/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/gnmic/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/gnmic/handlers/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/gnmic/handlers/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/gnmic/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/gnmic/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/gnmic/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/gnmic/tasks/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/gnmic/templates/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/gnmic/templates/docker-compose.yml.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/gnmic/templates/gnmic.yaml.j2",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/hddtemp/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/hddtemp/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/hddtemp/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/hddtemp/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/hddtemp/files/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/hddtemp/files/RedHat-family/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/hddtemp/files/RedHat-family/hddtemp",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/hddtemp/handlers/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/hddtemp/handlers/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/hddtemp/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/hddtemp/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/hddtemp/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/hddtemp/tasks/install-Debian-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/hddtemp/tasks/install-RedHat-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/hddtemp/tasks/main.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/hddtemp/tasks/service-Debian-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/hddtemp/tasks/service-RedHat-family.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/hddtemp/vars/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/hddtemp/vars/Debian-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/hddtemp/vars/RedHat-family.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/homer/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/homer/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/homer/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/homer/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/homer/handlers/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/homer/handlers/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/homer/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/homer/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/homer/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/homer/tasks/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/homer/templates/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/homer/templates/config.yml.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/homer/templates/docker-compose.yml.j2",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/httpd/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/httpd/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/httpd/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/httpd/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/httpd/files/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/httpd/files/prepare-ironic-volume.sh",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/httpd/handlers/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/httpd/handlers/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/httpd/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/httpd/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/httpd/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/httpd/tasks/ironic.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/httpd/tasks/main.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/httpd/tasks/sonic-ztp.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/httpd/templates/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/httpd/templates/docker-compose.yml.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/httpd/templates/htaccess.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/httpd/templates/httpd.conf.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/httpd/templates/post_install.sh.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/httpd/templates/ztp.json.j2",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/journald/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/journald/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/journald/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/journald/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/journald/handlers/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/journald/handlers/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/journald/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/journald/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/journald/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/journald/tasks/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/journald/templates/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/journald/templates/journald.conf.j2",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/kepler/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/kepler/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/kepler/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/kepler/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/kepler/handlers/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/kepler/handlers/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/kepler/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/kepler/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/kepler/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/kepler/tasks/config.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/kepler/tasks/main.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/kepler/tasks/service.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/kepler/templates/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/kepler/templates/config.yaml.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/kepler/templates/docker-compose.yml.j2",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/lldpd/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/lldpd/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/lldpd/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/lldpd/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/lldpd/handlers/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/lldpd/handlers/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/lldpd/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/lldpd/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/lldpd/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/lldpd/tasks/install-Debian-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/lldpd/tasks/install-RedHat-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/lldpd/tasks/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/manager/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/manager/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/manager/handlers/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/handlers/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/manager/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/manager/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/tasks/config-ansible.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/tasks/config-ara.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/tasks/config-celery.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/tasks/config-frontend.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/tasks/config-netbox.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/tasks/config-scripts.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/tasks/config-vault.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/tasks/config-wrapper.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/tasks/config.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/tasks/initialize-vault.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/tasks/initialize.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/tasks/install-Debian-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/tasks/install-RedHat-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/tasks/main.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/tasks/service.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/tasks/verify-versions.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/check-service-manager-versions.sh.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/config-netbox-secret.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/config-ssh-key.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/config-vault-key.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/config.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/docker-compose.yml.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/manager.service.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/vault.hcl.j2",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/env/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/env/all.env.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/env/ansible.env.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/env/ara-server.env.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/env/ara.env.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/env/client.env.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/env/conductor.env.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/env/frontend.env.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/env/inventory-reconciler.env.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/env/listener.env.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/env/mariadb.env.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/env/netbox.env.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/env/openstack.env.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/env/postgres.env.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/env/vault.env.j2",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/scripts/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/scripts/osism-include.j2",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/wrapper/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/wrapper/cilium.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/wrapper/flux.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/wrapper/hubble.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/wrapper/osism-ansible.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/wrapper/osism-ceph.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/wrapper/osism-custom.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/wrapper/osism-generic.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/wrapper/osism-infrastructure.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/wrapper/osism-kolla.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/wrapper/osism-manager.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/wrapper/osism-monitoring.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/wrapper/osism-openstack.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/wrapper/osism-run-without-secrets.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/wrapper/osism-run.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/wrapper/osism-state.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/wrapper/osism-update-docker.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/wrapper/osism-update-manager.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/wrapper/osism-validate.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/templates/wrapper/osism.j2",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/manager/vars/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/vars/Debian-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/vars/RedHat-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/vars/scripts.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/manager/vars/wrapper.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/netbird/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/netbird/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/netbird/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/netbird/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/netbird/handlers/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/netbird/handlers/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/netbird/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/netbird/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/netbird/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/netbird/tasks/install-Debian-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/netbird/tasks/install-RedHat-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/netbird/tasks/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/netbird/vars/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/netbird/vars/Debian-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/netbird/vars/RedHat-family.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/netbox/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/netbox/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/netbox/files/",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/netbox/files/startup-scripts/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/files/startup-scripts/270_tags.py",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/netbox/handlers/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/handlers/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/netbox/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/netbox/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/tasks/config-netbox.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/tasks/config-postgres.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/tasks/config.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/tasks/install-Debian-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/tasks/install-RedHat-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/tasks/main.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/tasks/rescue-service.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/tasks/restart-service.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/tasks/service.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/tasks/wait-for-healthy-service.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/netbox/templates/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/templates/config-netbox-database.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/templates/config-netbox-secret.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/templates/config-postgres-secret.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/templates/configuration.py.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/templates/docker-compose.yml.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/templates/netbox.service.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/templates/nginx-unit-1.26.json.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/templates/nginx-unit.json.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/templates/postgresql.conf.j2",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/netbox/templates/env/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/templates/env/netbox.env.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/templates/env/postgres.env.j2",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/netbox/templates/initializers/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/templates/initializers/custom_fields.yml.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/templates/initializers/device_roles.yml.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/templates/initializers/device_types.yml.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/templates/initializers/groups.yml.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/templates/initializers/manufacturers.yml.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/templates/initializers/object_permissions.yml.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/templates/initializers/prefix_vlan_roles.yml.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/templates/initializers/sites.yml.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/templates/initializers/tags.yml.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/templates/initializers/users.yml.j2",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/netbox/vars/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/netbox/vars/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/netdata/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/netdata/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/netdata/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/netdata/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/netdata/handlers/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/netdata/handlers/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/netdata/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/netdata/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/netdata/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/netdata/tasks/client.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/netdata/tasks/config.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/netdata/tasks/install-Debian-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/netdata/tasks/install-RedHat-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/netdata/tasks/main.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/netdata/tasks/server.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/netdata/templates/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/netdata/templates/netdata.conf.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/netdata/templates/stream.conf.j2",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/nexus/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/nexus/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/nexus/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/nexus/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/nexus/files/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/nexus/files/create_repos_from_list.groovy",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/nexus/files/setup_http_proxy.groovy",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/nexus/files/setup_realms.groovy",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/nexus/files/update_admin_password.groovy",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/nexus/handlers/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/nexus/handlers/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/nexus/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/nexus/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/nexus/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/nexus/tasks/call-script.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/nexus/tasks/config.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/nexus/tasks/declare-script.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/nexus/tasks/initialize.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/nexus/tasks/main.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/nexus/tasks/service.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/nexus/templates/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/nexus/templates/docker-compose.yml.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/nexus/templates/nexus.env.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/nexus/templates/nexus.properties.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/nexus/templates/nexus.service.j2",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/openstackclient/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/openstackclient/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/openstackclient/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/openstackclient/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/openstackclient/handlers/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/openstackclient/handlers/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/openstackclient/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/openstackclient/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/openstackclient/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/openstackclient/tasks/container-Debian-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/openstackclient/tasks/container-RedHat-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/openstackclient/tasks/main.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/openstackclient/tasks/package-Debian-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/openstackclient/tasks/package-RedHat-family.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/openstackclient/templates/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/openstackclient/templates/docker-compose.yml.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/openstackclient/templates/openstack.j2",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/opentelemetry_collector/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/opentelemetry_collector/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/opentelemetry_collector/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/opentelemetry_collector/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/opentelemetry_collector/handlers/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/opentelemetry_collector/handlers/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/opentelemetry_collector/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/opentelemetry_collector/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/opentelemetry_collector/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/opentelemetry_collector/tasks/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/opentelemetry_collector/templates/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/opentelemetry_collector/templates/docker-compose.yml.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/opentelemetry_collector/templates/otel-collector-config.yml.j2",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/osquery/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/osquery/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/osquery/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/osquery/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/osquery/handlers/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/osquery/handlers/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/osquery/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/osquery/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/osquery/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/osquery/tasks/install-Debian-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/osquery/tasks/install-RedHat-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/osquery/tasks/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/phpmyadmin/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/phpmyadmin/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/phpmyadmin/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/phpmyadmin/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/phpmyadmin/handlers/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/phpmyadmin/handlers/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/phpmyadmin/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/phpmyadmin/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/phpmyadmin/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/phpmyadmin/tasks/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/phpmyadmin/templates/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/phpmyadmin/templates/docker-compose.yml.j2",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/rng/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/rng/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/rng/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/rng/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/rng/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/rng/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/rng/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/rng/tasks/install-Debian-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/rng/tasks/install-RedHat-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/rng/tasks/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/rsyslog/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/rsyslog/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/rsyslog/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/rsyslog/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/rsyslog/handlers/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/rsyslog/handlers/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/rsyslog/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/rsyslog/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/rsyslog/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/rsyslog/tasks/additional-log-server.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/rsyslog/tasks/fluentd.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/rsyslog/tasks/install-Debian-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/rsyslog/tasks/install-RedHat-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/rsyslog/tasks/logrotate.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/rsyslog/tasks/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/rsyslog/templates/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/rsyslog/templates/70-fluentd.conf.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/rsyslog/templates/71-additional-log-server.conf.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/rsyslog/templates/rsyslog-logrotate.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/rsyslog/templates/rsyslog.conf.j2",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/rsyslog/vars/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/rsyslog/vars/CentOS-dist.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/rsyslog/vars/Debian-dist.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/rsyslog/vars/Ubuntu-dist.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/scaphandre/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/scaphandre/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/scaphandre/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/scaphandre/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/scaphandre/handlers/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/scaphandre/handlers/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/scaphandre/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/scaphandre/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/scaphandre/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/scaphandre/tasks/config.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/scaphandre/tasks/main.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/scaphandre/tasks/service.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/scaphandre/templates/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/scaphandre/templates/docker-compose.yml.j2",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/smartd/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/smartd/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/smartd/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/smartd/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/smartd/files/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/smartd/files/smartmontools",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/smartd/handlers/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/smartd/handlers/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/smartd/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/smartd/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/smartd/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/smartd/tasks/install-Debian-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/smartd/tasks/install-RedHat-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/smartd/tasks/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/squid/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/squid/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/squid/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/squid/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/squid/handlers/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/squid/handlers/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/squid/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/squid/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/squid/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/squid/tasks/install-Debian-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/squid/tasks/install-RedHat-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/squid/tasks/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/squid/templates/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/squid/templates/docker-compose.yml.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/squid/templates/osism.conf.j2",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/sshd/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/sshd/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/sshd/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/sshd/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/sshd/handlers/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/sshd/handlers/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/sshd/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/sshd/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/sshd/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/sshd/tasks/install-Debian-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/sshd/tasks/install-RedHat-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/sshd/tasks/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/stepca/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/stepca/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/stepca/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/stepca/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/stepca/handlers/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/stepca/handlers/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/stepca/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/stepca/tasks/config-wrapper.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/stepca/tasks/config.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/stepca/tasks/main.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/stepca/tasks/provisioner-acme.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/stepca/tasks/service.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/stepca/templates/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/stepca/templates/docker-compose.yml.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/stepca/templates/stepca.env.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/stepca/templates/stepca.service.j2",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/stepca/templates/wrapper/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/stepca/templates/wrapper/step.j2",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/substation/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/substation/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/substation/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/substation/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/substation/files/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/substation/files/entrypoint.sh",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/substation/handlers/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/substation/handlers/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/substation/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/substation/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/substation/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/substation/tasks/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/substation/templates/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/substation/templates/docker-compose.yml.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/substation/templates/substation.j2",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/teleport/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/teleport/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/teleport/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/teleport/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/teleport/handlers/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/teleport/handlers/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/teleport/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/teleport/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/teleport/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/teleport/tasks/install-Debian-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/teleport/tasks/install-RedHat-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/teleport/tasks/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/teleport/templates/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/teleport/templates/teleport.yaml.j2",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/thanos_sidecar/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/thanos_sidecar/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/thanos_sidecar/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/thanos_sidecar/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/thanos_sidecar/handlers/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/thanos_sidecar/handlers/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/thanos_sidecar/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/thanos_sidecar/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/thanos_sidecar/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/thanos_sidecar/tasks/config.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/thanos_sidecar/tasks/main.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/thanos_sidecar/tasks/service.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/thanos_sidecar/templates/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/thanos_sidecar/templates/docker-compose.yml.j2",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/traefik/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/traefik/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/traefik/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/traefik/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/traefik/handlers/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/traefik/handlers/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/traefik/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/traefik/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/traefik/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/traefik/tasks/config.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/traefik/tasks/main.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/traefik/tasks/service.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/traefik/templates/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/traefik/templates/certificates.yml.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/traefik/templates/docker-compose.yml.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/traefik/templates/dynamic.yml.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/traefik/templates/traefik-cert.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/traefik/templates/traefik-key.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/traefik/templates/traefik.env.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/traefik/templates/traefik.yml.j2",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/tuned/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/tuned/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/tuned/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/tuned/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/tuned/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/tuned/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/tuned/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/tuned/tasks/install-Debian-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/tuned/tasks/install-RedHat-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/tuned/tasks/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/wazuh_agent/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/wazuh_agent/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/wazuh_agent/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/wazuh_agent/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/wazuh_agent/handlers/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/wazuh_agent/handlers/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/wazuh_agent/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/wazuh_agent/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/wazuh_agent/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/wazuh_agent/tasks/install-Debian-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/wazuh_agent/tasks/install-RedHat-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/wazuh_agent/tasks/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/wazuh_agent/templates/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/wazuh_agent/templates/client.keys.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/wazuh_agent/templates/ossec.conf.j2",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/wireguard/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/wireguard/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/wireguard/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/wireguard/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/wireguard/handlers/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/wireguard/handlers/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/wireguard/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/wireguard/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/wireguard/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/wireguard/tasks/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/wireguard/templates/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/wireguard/templates/client.conf.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/wireguard/templates/server.conf.j2",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/zabbix_agent/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/zabbix_agent/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/zabbix_agent/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/zabbix_agent/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/zabbix_agent/handlers/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/zabbix_agent/handlers/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/zabbix_agent/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/zabbix_agent/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/zabbix_agent/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/zabbix_agent/tasks/install-Debian-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/zabbix_agent/tasks/install-RedHat-family.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/zabbix_agent/tasks/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/zabbix_agent/templates/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/zabbix_agent/templates/zabbix_agent2.conf.j2",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/zuul/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/zuul/README.md",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/zuul/defaults/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/zuul/defaults/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/zuul/files/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/zuul/files/builder-logging.conf",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/zuul/files/executor-logging.conf",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/zuul/files/launcher-logging.conf",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/zuul/files/logging.conf",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/zuul/files/mime.types",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/zuul/files/openssl.cnf",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/zuul/files/web-logging.conf",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/zuul/files/zk-ca.sh",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/zuul/handlers/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/zuul/handlers/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/zuul/meta/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/zuul/meta/main.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/zuul/tasks/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/zuul/tasks/copy_tasks.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/zuul/tasks/generate_certs_and_keys.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/zuul/tasks/main.yml",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/zuul/tasks/template_tasks.yml",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/zuul/templates/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/zuul/templates/docker-compose.yaml.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/zuul/templates/httpd.conf.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/zuul/templates/main.yaml.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/zuul/templates/nodepool.yaml.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/zuul/templates/wait-to-start-gearman.sh.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/zuul/templates/wait-to-start.sh.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/zuul/templates/zoo.cfg.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/zuul/templates/zuul.conf.j2",
                                ".d..t...... src/github.com/osism/ansible-collection-services/roles/zuul/templates/env/",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/zuul/templates/env/executor.env.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/zuul/templates/env/mariadb.env.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/zuul/templates/env/nodepool_builder.env.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/zuul/templates/env/nodepool_launcher.env.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/zuul/templates/env/scheduler.env.j2",
                                "<f..t...... src/github.com/osism/ansible-collection-services/roles/zuul/templates/env/web.env.j2",
                                ".d..t...... src/github.com/osism/terraform-base/",
                                "<f..t...... src/github.com/osism/terraform-base/.yamllint.yml",
                                "<f..t...... src/github.com/osism/terraform-base/.zuul.yaml",
                                "<f..t...... src/github.com/osism/terraform-base/LICENSE",
                                "<f..t...... src/github.com/osism/terraform-base/README.md",
                                ".d..t...... src/github.com/osism/terraform-base/.git/",
                                "<f..t...... src/github.com/osism/terraform-base/.git/HEAD",
                                "<f..t...... src/github.com/osism/terraform-base/.git/ORIG_HEAD",
                                "<f..t...... src/github.com/osism/terraform-base/.git/config",
                                "<f..t...... src/github.com/osism/terraform-base/.git/config.worktree",
                                "<f..t...... src/github.com/osism/terraform-base/.git/description",
                                "<f..t...... src/github.com/osism/terraform-base/.git/index",
                                "<f..t...... src/github.com/osism/terraform-base/.git/packed-refs",
                                ".d..t...... src/github.com/osism/terraform-base/.git/branches/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/hooks/",
                                "<f..t...... src/github.com/osism/terraform-base/.git/hooks/applypatch-msg.sample",
                                "<f..t...... src/github.com/osism/terraform-base/.git/hooks/commit-msg.sample",
                                "<f..t...... src/github.com/osism/terraform-base/.git/hooks/fsmonitor-watchman.sample",
                                "<f..t...... src/github.com/osism/terraform-base/.git/hooks/post-update.sample",
                                "<f..t...... src/github.com/osism/terraform-base/.git/hooks/pre-applypatch.sample",
                                "<f..t...... src/github.com/osism/terraform-base/.git/hooks/pre-commit.sample",
                                "<f..t...... src/github.com/osism/terraform-base/.git/hooks/pre-merge-commit.sample",
                                "<f..t...... src/github.com/osism/terraform-base/.git/hooks/pre-push.sample",
                                "<f..t...... src/github.com/osism/terraform-base/.git/hooks/pre-rebase.sample",
                                "<f..t...... src/github.com/osism/terraform-base/.git/hooks/pre-receive.sample",
                                "<f..t...... src/github.com/osism/terraform-base/.git/hooks/prepare-commit-msg.sample",
                                "<f..t...... src/github.com/osism/terraform-base/.git/hooks/push-to-checkout.sample",
                                "<f..t...... src/github.com/osism/terraform-base/.git/hooks/update.sample",
                                ".d..t...... src/github.com/osism/terraform-base/.git/info/",
                                "<f..t...... src/github.com/osism/terraform-base/.git/info/exclude",
                                ".d..t...... src/github.com/osism/terraform-base/.git/logs/",
                                "<f..t...... src/github.com/osism/terraform-base/.git/logs/HEAD",
                                ".d..t...... src/github.com/osism/terraform-base/.git/logs/refs/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/logs/refs/heads/",
                                "<f..t...... src/github.com/osism/terraform-base/.git/logs/refs/heads/main",
                                ".d..t...... src/github.com/osism/terraform-base/.git/logs/refs/remotes/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/logs/refs/remotes/origin/",
                                "<f..t...... src/github.com/osism/terraform-base/.git/logs/refs/remotes/origin/HEAD",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/00/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/01/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/02/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/05/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/06/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/07/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/08/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/0a/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/0b/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/0d/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/0f/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/10/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/12/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/15/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/16/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/17/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/1b/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/1c/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/1f/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/20/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/21/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/24/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/26/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/29/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/2c/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/36/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/37/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/3a/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/3d/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/41/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/43/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/46/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/47/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/48/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/49/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/4b/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/4d/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/52/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/55/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/56/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/58/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/5d/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/61/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/63/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/66/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/67/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/68/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/6d/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/70/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/72/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/76/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/7c/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/7f/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/80/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/81/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/82/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/83/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/84/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/85/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/86/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/88/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/8c/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/8d/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/90/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/91/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/92/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/95/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/96/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/98/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/9b/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/9e/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/a2/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/a3/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/a6/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/a7/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/a9/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/aa/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/af/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/b0/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/b3/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/b4/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/b8/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/b9/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/ba/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/bc/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/bf/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/c0/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/c1/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/c4/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/c7/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/c8/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/cb/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/ce/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/cf/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/d2/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/d3/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/d6/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/d7/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/d9/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/da/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/db/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/dc/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/dd/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/de/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/df/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/e1/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/e2/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/e3/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/e4/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/e8/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/ea/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/ee/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/ef/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/f1/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/f2/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/f4/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/f9/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/fb/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/ff/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/info/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/objects/pack/",
                                ".d..t...... src/github.com/osism/terraform-base/.git/refs/",
                                ".d..t...... src/github.com/osism/terraform-base/.github/",
                                "<f..t...... src/github.com/osism/terraform-base/.github/renovate.json",
                                ".d..t...... src/github.com/osism/terraform-base/playbooks/",
                                "<f..t...... src/github.com/osism/terraform-base/playbooks/validate.yml",
                                ".d..t...... src/github.com/osism/terraform-base/testbed-default/",
                                "<f..t...... src/github.com/osism/terraform-base/testbed-default/data.tf",
                                "<f..t...... src/github.com/osism/terraform-base/testbed-default/main.tf",
                                "<f..t...... src/github.com/osism/terraform-base/testbed-default/manager.tf",
                                "<f..t...... src/github.com/osism/terraform-base/testbed-default/neutron.tf",
                                "<f..t...... src/github.com/osism/terraform-base/testbed-default/nodes.tf",
                                "<f..t...... src/github.com/osism/terraform-base/testbed-default/nova.tf",
                                "<f..t...... src/github.com/osism/terraform-base/testbed-default/outputs.tf",
                                "<f..t...... src/github.com/osism/terraform-base/testbed-default/provider.tf",
                                "<f..t...... src/github.com/osism/terraform-base/testbed-default/variables.tf",
                                ".d..t...... src/github.com/osism/terraform-base/testbed-default/customisations/",
                                "<f..t...... src/github.com/osism/terraform-base/testbed-default/customisations/access_floatingip_custom.tf",
                                "<f..t...... src/github.com/osism/terraform-base/testbed-default/customisations/access_ipv4_custom.tf",
                                "<f..t...... src/github.com/osism/terraform-base/testbed-default/customisations/access_ipv6_custom.tf",
                                "<f..t...... src/github.com/osism/terraform-base/testbed-default/customisations/default_custom.tf",
                                "<f..t...... src/github.com/osism/terraform-base/testbed-default/customisations/external_api_custom.tf",
                                "<f..t...... src/github.com/osism/terraform-base/testbed-default/customisations/neutron_floatingip_custom.tf",
                                ".d..t...... src/github.com/osism/terraform-base/testbed-default/environments/",
                                "<f..t...... src/github.com/osism/terraform-base/testbed-default/environments/default.tfvars",
                                ".d..t...... src/github.com/osism/terraform-base/testbed-default/overrides/",
                                "<f..t...... src/github.com/osism/terraform-base/testbed-default/overrides/manager_boot_from_image_override.tf",
                                "<f..t...... src/github.com/osism/terraform-base/testbed-default/overrides/manager_boot_from_volume_override.tf",
                                "<f..t...... src/github.com/osism/terraform-base/testbed-default/overrides/neutron_availability_zone_hints_network_override.tf",
                                "<f..t...... src/github.com/osism/terraform-base/testbed-default/overrides/neutron_availability_zone_hints_router_override.tf",
                                "<f..t...... src/github.com/osism/terraform-base/testbed-default/overrides/neutron_router_enable_snat_override.tf",
                                "<f..t...... src/github.com/osism/terraform-base/testbed-default/overrides/nodes_boot_from_image_override.tf",
                                "<f..t...... src/github.com/osism/terraform-base/testbed-default/overrides/nodes_boot_from_volume_override.tf",
                                "<f..t...... src/github.com/osism/terraform-base/testbed-default/overrides/nodes_use_ephemeral_storage_override.tf",
                                ".d..t...... src/github.com/osism/testbed/",
                                "<f..t...... src/github.com/osism/testbed/.ansible-lint",
                                "<f..t...... src/github.com/osism/testbed/.flake8",
                                "<f..t...... src/github.com/osism/testbed/.gitignore",
                                "<f..t...... src/github.com/osism/testbed/.yamllint.yml",
                                "<f..t...... src/github.com/osism/testbed/.zuul.yaml",
                                "<f..t...... src/github.com/osism/testbed/LICENSE",
                                "<f..t...... src/github.com/osism/testbed/Makefile",
                                "<f..t...... src/github.com/osism/testbed/README.md",
                                "<f..t...... src/github.com/osism/testbed/ansible.cfg",
                                "<f..t...... src/github.com/osism/testbed/gilt.yml",
                                "<f..t...... src/github.com/osism/testbed/requirements.txt",
                                "<f..t...... src/github.com/osism/testbed/requirements.yml",
                                "<f..t...... src/github.com/osism/testbed/tox.ini",
                                ".d..t...... src/github.com/osism/testbed/.git/",
                                "<f..t...... src/github.com/osism/testbed/.git/HEAD",
                                "<f..t...... src/github.com/osism/testbed/.git/ORIG_HEAD",
                                "<f.st...... src/github.com/osism/testbed/.git/config",
                                "<f..t...... src/github.com/osism/testbed/.git/config.worktree",
                                "<f..t...... src/github.com/osism/testbed/.git/description",
                                "<f..t...... src/github.com/osism/testbed/.git/index",
                                "<f..t...... src/github.com/osism/testbed/.git/packed-refs",
                                ".d..t...... src/github.com/osism/testbed/.git/branches/",
                                ".d..t...... src/github.com/osism/testbed/.git/hooks/",
                                "<f..t...... src/github.com/osism/testbed/.git/hooks/applypatch-msg.sample",
                                "<f..t...... src/github.com/osism/testbed/.git/hooks/commit-msg.sample",
                                "<f..t...... src/github.com/osism/testbed/.git/hooks/fsmonitor-watchman.sample",
                                "<f..t...... src/github.com/osism/testbed/.git/hooks/post-update.sample",
                                "<f..t...... src/github.com/osism/testbed/.git/hooks/pre-applypatch.sample",
                                "<f..t...... src/github.com/osism/testbed/.git/hooks/pre-commit.sample",
                                "<f..t...... src/github.com/osism/testbed/.git/hooks/pre-merge-commit.sample",
                                "<f..t...... src/github.com/osism/testbed/.git/hooks/pre-push.sample",
                                "<f..t...... src/github.com/osism/testbed/.git/hooks/pre-rebase.sample",
                                "<f..t...... src/github.com/osism/testbed/.git/hooks/pre-receive.sample",
                                "<f..t...... src/github.com/osism/testbed/.git/hooks/prepare-commit-msg.sample",
                                "<f..t...... src/github.com/osism/testbed/.git/hooks/push-to-checkout.sample",
                                "<f..t...... src/github.com/osism/testbed/.git/hooks/update.sample",
                                ".d..t...... src/github.com/osism/testbed/.git/info/",
                                "<f..t...... src/github.com/osism/testbed/.git/info/exclude",
                                ".d..t...... src/github.com/osism/testbed/.git/logs/",
                                "<f.st...... src/github.com/osism/testbed/.git/logs/HEAD",
                                "cd+++++++++ src/github.com/osism/testbed/.git/logs/refs/",
                                "cd+++++++++ src/github.com/osism/testbed/.git/logs/refs/heads/",
                                "<f+++++++++ src/github.com/osism/testbed/.git/logs/refs/heads/main",
                                "cd+++++++++ src/github.com/osism/testbed/.git/logs/refs/remotes/",
                                "cd+++++++++ src/github.com/osism/testbed/.git/logs/refs/remotes/origin/",
                                "<f+++++++++ src/github.com/osism/testbed/.git/logs/refs/remotes/origin/HEAD",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/00/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/01/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/02/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/03/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/04/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/05/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/06/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/07/",
                                "<f..t...... src/github.com/osism/testbed/.git/objects/07/1d016fb2341fa1b57d73e6c901348ad5a5cc85",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/08/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/09/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/0a/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/0b/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/0c/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/0d/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/0e/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/0f/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/10/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/11/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/12/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/13/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/14/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/15/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/16/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/17/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/18/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/19/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/1a/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/1b/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/1c/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/1d/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/1e/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/1f/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/20/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/21/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/22/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/23/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/24/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/25/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/26/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/27/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/28/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/29/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/2a/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/2b/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/2c/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/2d/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/2e/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/2f/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/30/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/31/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/32/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/33/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/34/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/35/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/36/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/37/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/38/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/39/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/3a/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/3b/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/3c/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/3d/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/3e/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/3f/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/40/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/41/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/42/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/43/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/44/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/45/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/46/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/47/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/48/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/49/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/4a/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/4c/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/4d/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/4e/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/4f/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/50/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/51/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/52/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/53/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/54/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/55/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/56/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/57/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/58/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/59/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/5a/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/5b/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/5c/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/5d/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/5e/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/5f/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/60/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/61/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/64/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/65/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/66/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/67/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/68/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/69/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/6a/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/6b/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/6c/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/6d/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/6e/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/6f/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/70/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/71/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/72/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/73/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/74/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/75/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/76/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/77/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/78/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/79/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/7a/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/7b/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/7c/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/7d/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/7e/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/7f/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/80/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/81/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/82/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/83/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/84/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/85/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/86/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/87/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/8a/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/8b/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/8c/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/8e/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/8f/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/90/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/91/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/92/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/93/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/94/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/95/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/96/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/97/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/98/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/9a/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/9c/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/9e/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/9f/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/a0/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/a1/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/a2/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/a3/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/a4/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/a5/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/a6/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/a7/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/a8/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/a9/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/aa/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/ab/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/ac/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/ad/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/ae/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/af/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/b0/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/b1/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/b2/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/b3/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/b4/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/b5/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/b6/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/b7/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/b8/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/b9/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/ba/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/bb/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/bc/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/bd/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/be/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/bf/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/c0/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/c1/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/c2/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/c3/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/c4/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/c5/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/c6/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/c7/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/c8/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/c9/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/ca/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/cb/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/cc/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/cd/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/ce/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/cf/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/d0/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/d2/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/d3/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/d5/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/d6/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/d7/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/d8/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/d9/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/da/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/db/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/dc/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/dd/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/de/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/df/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/e0/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/e2/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/e3/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/e4/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/e5/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/e6/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/e7/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/e8/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/e9/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/ea/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/eb/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/ec/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/ed/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/ee/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/ef/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/f0/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/f1/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/f2/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/f3/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/f4/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/f5/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/f6/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/f8/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/f9/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/fa/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/fb/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/fc/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/fd/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/fe/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/ff/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/info/",
                                ".d..t...... src/github.com/osism/testbed/.git/objects/pack/",
                                ".d..t...... src/github.com/osism/testbed/.git/refs/",
                                ".d..t...... src/github.com/osism/testbed/.github/",
                                "<f..t...... src/github.com/osism/testbed/.github/renovate.json",
                                ".d..t...... src/github.com/osism/testbed/.github/workflows/",
                                "<f..t...... src/github.com/osism/testbed/.github/workflows/update-manager-images.yml",
                                ".d..t...... src/github.com/osism/testbed/ansible/",
                                "<f..t...... src/github.com/osism/testbed/ansible/check-local-versions.yml",
                                "<f..t...... src/github.com/osism/testbed/ansible/localhost_inventory.yaml",
                                "<f..t...... src/github.com/osism/testbed/ansible/manager-part-0.yml",
                                "<f..t...... src/github.com/osism/testbed/ansible/manager-part-1.yml",
                                "<f..t...... src/github.com/osism/testbed/ansible/manager-part-2.yml",
                                "<f..t...... src/github.com/osism/testbed/ansible/manager-part-3.yml",
                                ".d..t...... src/github.com/osism/testbed/containerlab/",
                                "<f..t...... src/github.com/osism/testbed/containerlab/README.md",
                                "<f..t...... src/github.com/osism/testbed/containerlab/osism.clab.yml",
                                ".d..t...... src/github.com/osism/testbed/containerlab/files/",
                                "<f..t...... src/github.com/osism/testbed/containerlab/files/daemons",
                                "<f..t...... src/github.com/osism/testbed/containerlab/files/frr01.conf",
                                "<f..t...... src/github.com/osism/testbed/containerlab/files/frr02.conf",
                                "<f..t...... src/github.com/osism/testbed/containerlab/files/vtysh.conf",
                                ".d..t...... src/github.com/osism/testbed/contrib/",
                                "<f..t...... src/github.com/osism/testbed/contrib/analyze-tempest-logs.sh",
                                "<f..t...... src/github.com/osism/testbed/contrib/analyze-zuul-logs.py",
                                "<f..t...... src/github.com/osism/testbed/contrib/elasticsearch_random_data_load.sh",
                                "<f..t...... src/github.com/osism/testbed/contrib/mysql_random_data_load.sh",
                                "<f..t...... src/github.com/osism/testbed/contrib/mysql_random_data_load.sql",
                                "<f..t...... src/github.com/osism/testbed/contrib/semver2.sh",
                                "<f..t...... src/github.com/osism/testbed/contrib/setup-testbed.py",
                                ".d..t...... src/github.com/osism/testbed/contrib/ansible/",
                                "<f..t...... src/github.com/osism/testbed/contrib/ansible/clouds.yaml.sample",
                                "<f..t...... src/github.com/osism/testbed/contrib/ansible/dns.yml",
                                "<f..t...... src/github.com/osism/testbed/contrib/ansible/secure.yaml.sample",
                                ".d..t...... src/github.com/osism/testbed/contrib/assets/",
                                "<f..t...... src/github.com/osism/testbed/contrib/assets/operations-dashboard.png",
                                ".d..t...... src/github.com/osism/testbed/contrib/integration-tests/",
                                "<f..t...... src/github.com/osism/testbed/contrib/integration-tests/docker-compose.yaml",
                                "<f..t...... src/github.com/osism/testbed/contrib/integration-tests/run_test.sh",
                                ".d..t...... src/github.com/osism/testbed/contrib/integration-tests/tests/",
                                "<f..t...... src/github.com/osism/testbed/contrib/integration-tests/tests/new_user.robot",
                                ".d..t...... src/github.com/osism/testbed/contrib/nagios-plugins/",
                                "<f..t...... src/github.com/osism/testbed/contrib/nagios-plugins/check_elasticsearch",
                                "<f..t...... src/github.com/osism/testbed/contrib/nagios-plugins/check_galera_cluster",
                                "<f..t...... src/github.com/osism/testbed/contrib/nagios-plugins/check_rabbitmq_cluster",
                                ".d..t...... src/github.com/osism/testbed/contrib/openstack-cli/",
                                "<f..t...... src/github.com/osism/testbed/contrib/openstack-cli/README.md",
                                "<f..t...... src/github.com/osism/testbed/contrib/openstack-cli/clouds-public.yaml",
                                "<f..t...... src/github.com/osism/testbed/contrib/openstack-cli/clouds.yaml.example",
                                "<f..t...... src/github.com/osism/testbed/contrib/openstack-cli/secure.yaml.example",
                                ".d..t...... src/github.com/osism/testbed/contrib/ownca/",
                                "<f..t...... src/github.com/osism/testbed/contrib/ownca/README.md",
                                "<f..t...... src/github.com/osism/testbed/contrib/ownca/create_ca.yml",
                                "<f..t...... src/github.com/osism/testbed/contrib/ownca/create_manager.yml",
                                "<f..t...... src/github.com/osism/testbed/contrib/ownca/create_wildcard.yml",
                                ".d..t...... src/github.com/osism/testbed/environments/",
                                "<f..t...... src/github.com/osism/testbed/environments/.vault_pass",
                                "<f..t...... src/github.com/osism/testbed/environments/ansible.cfg",
                                "<f..t...... src/github.com/osism/testbed/environments/configuration.yml",
                                "<f..t...... src/github.com/osism/testbed/environments/images.yml",
                                "<f..t...... src/github.com/osism/testbed/environments/secrets.yml",
                                ".d..t...... src/github.com/osism/testbed/environments/ceph.test/",
                                "<f..t...... src/github.com/osism/testbed/environments/ceph.test/configuration.yml",
                                "<f..t...... src/github.com/osism/testbed/environments/ceph.test/images.yml",
                                "<f..t...... src/github.com/osism/testbed/environments/ceph.test/secrets.yml",
                                ".d..t...... src/github.com/osism/testbed/environments/ceph/",
                                "<f..t...... src/github.com/osism/testbed/environments/ceph/configuration.yml",
                                "<f..t...... src/github.com/osism/testbed/environments/ceph/images.yml",
                                "<f..t...... src/github.com/osism/testbed/environments/ceph/secrets.yml",
                                ".d..t...... src/github.com/osism/testbed/environments/custom/",
                                "<f..t...... src/github.com/osism/testbed/environments/custom/configuration.yml",
                                "<f..t...... src/github.com/osism/testbed/environments/custom/images.yml",
                                "<f..t...... src/github.com/osism/testbed/environments/custom/playbook-baremetal-bootstrap.yml",
                                "<f..t...... src/github.com/osism/testbed/environments/custom/playbook-baremetal-netbox.yml",
                                "<f..t...... src/github.com/osism/testbed/environments/custom/playbook-baremetal-pull-images.yml",
                                "<f..t...... src/github.com/osism/testbed/environments/custom/playbook-facts.yml",
                                "<f..t...... src/github.com/osism/testbed/environments/custom/playbook-migrate-proxysql.yml",
                                "<f..t...... src/github.com/osism/testbed/environments/custom/playbook-pull-images.yml",
                                "<f..t...... src/github.com/osism/testbed/environments/custom/playbook-wipe-partitions.yml",
                                "<f..t...... src/github.com/osism/testbed/environments/custom/playbook-workarounds-ceph.yml",
                                "<f..t...... src/github.com/osism/testbed/environments/custom/playbook-workarounds.yml",
                                "<f..t...... src/github.com/osism/testbed/environments/custom/secrets.yml",
                                ".d..t...... src/github.com/osism/testbed/environments/custom/files/",
                                "<f..t...... src/github.com/osism/testbed/environments/custom/files/baremetal-tenks-override.yml",
                                "<f..t...... src/github.com/osism/testbed/environments/custom/files/testbed_ceph_devices.fact",
                                "<f..t...... src/github.com/osism/testbed/environments/custom/files/testbed_ceph_devices_all.fact",
                                "<f..t...... src/github.com/osism/testbed/environments/custom/files/testbed_ceph_osd_devices.fact",
                                "<f..t...... src/github.com/osism/testbed/environments/custom/files/testbed_ceph_osd_devices_all.fact",
                                "<f..t...... src/github.com/osism/testbed/environments/custom/files/testbed_network_devices.fact",
                                ".d..t...... src/github.com/osism/testbed/environments/custom/templates/",
                                "<f..t...... src/github.com/osism/testbed/environments/custom/templates/baremetal-netbox-device.yml.j2",
                                "<f..t...... src/github.com/osism/testbed/environments/custom/templates/ironic-interface.service.j2",
                                "<f..t...... src/github.com/osism/testbed/environments/custom/templates/workarounds.service.j2",
                                "<f..t...... src/github.com/osism/testbed/environments/custom/templates/workarounds.sh.j2",
                                ".d..t...... src/github.com/osism/testbed/environments/generic/",
                                "<f..t...... src/github.com/osism/testbed/environments/generic/configuration.yml",
                                "<f..t...... src/github.com/osism/testbed/environments/generic/images.yml",
                                "<f..t...... src/github.com/osism/testbed/environments/generic/secrets.yml",
                                ".d..t...... src/github.com/osism/testbed/environments/infrastructure/",
                                "<f..t...... src/github.com/osism/testbed/environments/infrastructure/configuration.yml",
                                "<f..t...... src/github.com/osism/testbed/environments/infrastructure/images.yml",
                                "<f..t...... src/github.com/osism/testbed/environments/infrastructure/secrets.yml",
                                ".d..t...... src/github.com/osism/testbed/environments/infrastructure/files/",
                                ".d..t...... src/github.com/osism/testbed/environments/infrastructure/files/ceph/",
                                "<f..t...... src/github.com/osism/testbed/environments/infrastructure/files/ceph/.gitkeep",
                                ".d..t...... src/github.com/osism/testbed/environments/infrastructure/files/keycloak/",
                                "<f..t...... src/github.com/osism/testbed/environments/infrastructure/files/keycloak/cert.crt",
                                "<f..t...... src/github.com/osism/testbed/environments/infrastructure/files/keycloak/private_key.pem",
                                ".d..t...... src/github.com/osism/testbed/environments/kolla/",
                                "<f.st...... src/github.com/osism/testbed/environments/kolla/configuration.yml",
                                "<f..t...... src/github.com/osism/testbed/environments/kolla/images.yml",
                                "<f..t...... src/github.com/osism/testbed/environments/kolla/secrets.yml",
                                ".d..t...... src/github.com/osism/testbed/environments/kolla/certificates/",
                                "<f..t...... src/github.com/osism/testbed/environments/kolla/certificates/haproxy-internal.pem",
                                "<f..t...... src/github.com/osism/testbed/environments/kolla/certificates/haproxy.pem",
                                ".d..t...... src/github.com/osism/testbed/environments/kolla/certificates/ca/",
                                "<f..t...... src/github.com/osism/testbed/environments/kolla/certificates/ca/testbed.crt",
                                ".d..t...... src/github.com/osism/testbed/environments/kolla/files/",
                                ".d..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/",
                                "<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/cinder.conf",
                                "<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/galera.cnf",
                                "<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/global.conf",
                                "<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/grafana.ini",
                                "<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/keystone.conf",
                                "<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/magnum.conf",
                                "<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/neutron.conf",
                                ".d..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/ceilometer/",
                                "<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/ceilometer/event_pipeline.yaml",
                                "<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/ceilometer/pipeline.yaml",
                                ".d..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/cinder/",
                                "<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/cinder/ceph.conf",
                                "<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/cinder/cinder-scheduler.conf",
                                "<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/cinder/cinder-volume.conf",
                                ".d..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/cinder/cinder-backup/",
                                "<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/cinder/cinder-backup/.gitkeep",
                                ".d..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/cinder/cinder-volume/",
                                "<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/cinder/cinder-volume/.gitkeep",
                                ".d..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/glance/",
                                "<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/glance/ceph.conf",
                                "<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/glance/glance-api.conf",
                                ".d..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/gnocchi/",
                                "<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/gnocchi/ceph.conf",
                                ".d..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/grafana/",
                                "<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/grafana/provisioning.yaml",
                                ".d..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/grafana/dashboards/",
                                "<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/grafana/dashboards/.gitkeep",
                                ".d..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/haproxy/",
                                ".d..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/haproxy/services.d/",
                                "<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/haproxy/services.d/haproxy.cfg",
                                ".d..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/ironic/",
                                "<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/ironic/ironic-agent.initramfs",
                                "<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/ironic/ironic-agent.kernel",
                                "<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/ironic/ironic-conductor.conf",
                                "<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/ironic/ironic-dnsmasq.conf",
                                ".d..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/magnum/",
                                "<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/magnum/.gitkeep",
                                ".d..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/manila/",
                                "<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/manila/ceph.conf",
                                ".d..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/neutron/",
                                "<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/neutron/ml2_conf.ini",
                                ".d..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/nova/",
                                "<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/nova/ceph.conf",
                                "<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/nova/nova-compute.conf",
                                ".d..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/octavia/",
                                "<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/octavia/client.cert-and-key.pem",
                                "<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/octavia/client_ca.cert.pem",
                                "<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/octavia/server_ca.cert.pem",
                                "<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/octavia/server_ca.key.pem",
                                ".d..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/prometheus/",
                                "<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/prometheus/prometheus-alertmanager.yml",
                                ".d..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/prometheus/prometheus.yml.d/",
                                "<f..t...... src/github.com/osism/testbed/environments/kolla/files/overlays/prometheus/prometheus.yml.d/50-ceph.yml",
                                ".d..t...... src/github.com/osism/testbed/environments/kubernetes/",
                                "<f..t...... src/github.com/osism/testbed/environments/kubernetes/configuration.yml",
                                "<f..t...... src/github.com/osism/testbed/environments/kubernetes/images.yml",
                                "<f..t...... src/github.com/osism/testbed/environments/kubernetes/playbook-nginx.yml",
                                "<f..t...... src/github.com/osism/testbed/environments/kubernetes/secrets.yml",
                                ".d..t...... src/github.com/osism/testbed/environments/kubernetes/files/",
                                "<f..t...... src/github.com/osism/testbed/environments/kubernetes/files/nginx.yml",
                                ".d..t...... src/github.com/osism/testbed/environments/manager/",
                                "<f..t...... src/github.com/osism/testbed/environments/manager/configuration.yml",
                                "<f..t...... src/github.com/osism/testbed/environments/manager/hosts",
                                "<f..t...... src/github.com/osism/testbed/environments/manager/images.yml",
                                "<f..t...... src/github.com/osism/testbed/environments/manager/secrets.yml",
                                ".d..t...... src/github.com/osism/testbed/environments/manager/files/",
                                "<f..t...... src/github.com/osism/testbed/environments/manager/files/conductor.yml",
                                ".d..t...... src/github.com/osism/testbed/environments/monitoring/",
                                "<f..t...... src/github.com/osism/testbed/environments/monitoring/configuration.yml",
                                "<f..t...... src/github.com/osism/testbed/environments/monitoring/images.yml",
                                "<f..t...... src/github.com/osism/testbed/environments/monitoring/secrets.yml",
                                ".d..t...... src/github.com/osism/testbed/environments/openstack/",
                                "<f..t...... src/github.com/osism/testbed/environments/openstack/clouds.yml",
                                "<f..t...... src/github.com/osism/testbed/environments/openstack/configuration.yml",
                                "<f..t...... src/github.com/osism/testbed/environments/openstack/id_rsa.test",
                                "<f..t...... src/github.com/osism/testbed/environments/openstack/id_rsa.test.pub",
                                "<f..t...... src/github.com/osism/testbed/environments/openstack/images.yml",
                                "<f..t...... src/github.com/osism/testbed/environments/openstack/openrc",
                                "<f..t...... src/github.com/osism/testbed/environments/openstack/playbook-bootstrap-basic.yml",
                                "<f..t...... src/github.com/osism/testbed/environments/openstack/playbook-bootstrap-ceph-rgw.yml",
                                "<f..t...... src/github.com/osism/testbed/environments/openstack/playbook-bootstrap-flavors.yml",
                                "<f..t...... src/github.com/osism/testbed/environments/openstack/playbook-bootstrap-images.yml",
                                "<f..t...... src/github.com/osism/testbed/environments/openstack/playbook-test-identity.yml",
                                "<f..t...... src/github.com/osism/testbed/environments/openstack/playbook-test.yml",
                                "<f..t...... src/github.com/osism/testbed/environments/openstack/secrets.yml",
                                "<f..t...... src/github.com/osism/testbed/environments/openstack/secure.yml",
                                "<f..t...... src/github.com/osism/testbed/environments/openstack/testbed.pem",
                                ".d..t...... src/github.com/osism/testbed/environments/openstack/files/",
                                ".d..t...... src/github.com/osism/testbed/environments/openstack/files/tempest/",
                                "<f..t...... src/github.com/osism/testbed/environments/openstack/files/tempest/exclude.lst",
                                "<f..t...... src/github.com/osism/testbed/environments/openstack/files/tempest/include-scs-compatible.lst",
                                ".d..t...... src/github.com/osism/testbed/inventory/",
                                "<f..t...... src/github.com/osism/testbed/inventory/10-custom",
                                "<f..t...... src/github.com/osism/testbed/inventory/20-roles",
                                "<f..t...... src/github.com/osism/testbed/inventory/99-overwrite",
                                ".d..t...... src/github.com/osism/testbed/inventory/group_vars/",
                                "<f..t...... src/github.com/osism/testbed/inventory/group_vars/testbed-external-nodes.yml",
                                "<f..t...... src/github.com/osism/testbed/inventory/group_vars/testbed-managers.yml",
                                "<f..t...... src/github.com/osism/testbed/inventory/group_vars/testbed-nodes.yml",
                                ".d..t...... src/github.com/osism/testbed/inventory/group_vars/all/",
                                "<f..t...... src/github.com/osism/testbed/inventory/group_vars/all/ansible.yml",
                                "<f..t...... src/github.com/osism/testbed/inventory/group_vars/all/kolla.yml",
                                "<f..t...... src/github.com/osism/testbed/inventory/group_vars/all/registries.yml",
                                "<f..t...... src/github.com/osism/testbed/inventory/group_vars/all/vxlan.yml",
                                ".d..t...... src/github.com/osism/testbed/inventory/host_vars/",
                                ".d..t...... src/github.com/osism/testbed/inventory/host_vars/testbed-manager/",
                                "<f..t...... src/github.com/osism/testbed/inventory/host_vars/testbed-manager/vars.yml",
                                ".d..t...... src/github.com/osism/testbed/inventory/host_vars/testbed-node-0/",
                                "<f..t...... src/github.com/osism/testbed/inventory/host_vars/testbed-node-0/vars.yml",
                                ".d..t...... src/github.com/osism/testbed/inventory/host_vars/testbed-node-1/",
                                "<f..t...... src/github.com/osism/testbed/inventory/host_vars/testbed-node-1/vars.yml",
                                ".d..t...... src/github.com/osism/testbed/inventory/host_vars/testbed-node-2/",
                                "<f..t...... src/github.com/osism/testbed/inventory/host_vars/testbed-node-2/vars.yml",
                                ".d..t...... src/github.com/osism/testbed/inventory/host_vars/testbed-node-3/",
                                "<f..t...... src/github.com/osism/testbed/inventory/host_vars/testbed-node-3/vars.yml",
                                ".d..t...... src/github.com/osism/testbed/inventory/host_vars/testbed-node-4/",
                                "<f..t...... src/github.com/osism/testbed/inventory/host_vars/testbed-node-4/vars.yml",
                                ".d..t...... src/github.com/osism/testbed/inventory/host_vars/testbed-node-5/",
                                "<f..t...... src/github.com/osism/testbed/inventory/host_vars/testbed-node-5/vars.yml",
                                ".d..t...... src/github.com/osism/testbed/inventory/host_vars/testbed-node-6/",
                                "<f..t...... src/github.com/osism/testbed/inventory/host_vars/testbed-node-6/vars.yml",
                                ".d..t...... src/github.com/osism/testbed/inventory/host_vars/testbed-node-7/",
                                "<f..t...... src/github.com/osism/testbed/inventory/host_vars/testbed-node-7/vars.yml",
                                ".d..t...... src/github.com/osism/testbed/inventory/host_vars/testbed-node-8/",
                                "<f..t...... src/github.com/osism/testbed/inventory/host_vars/testbed-node-8/vars.yml",
                                ".d..t...... src/github.com/osism/testbed/inventory/host_vars/testbed-node-9/",
                                "<f..t...... src/github.com/osism/testbed/inventory/host_vars/testbed-node-9/vars.yml",
                                ".d..t...... src/github.com/osism/testbed/netbox/",
                                ".d..t...... src/github.com/osism/testbed/netbox/devicetypes/",
                                ".d..t...... src/github.com/osism/testbed/netbox/devicetypes/Edgecore/",
                                "<f..t...... src/github.com/osism/testbed/netbox/devicetypes/Edgecore/5835-54X-O-AC-F.yaml",
                                "<f..t...... src/github.com/osism/testbed/netbox/devicetypes/Edgecore/7726-32X-O-AC-F.yaml",
                                ".d..t...... src/github.com/osism/testbed/netbox/devicetypes/Other/",
                                "<f..t...... src/github.com/osism/testbed/netbox/devicetypes/Other/baremetal-device.yml",
                                "<f..t...... src/github.com/osism/testbed/netbox/devicetypes/Other/baremetal-housing.yml",
                                "<f..t...... src/github.com/osism/testbed/netbox/devicetypes/Other/manager.yml",
                                "<f..t...... src/github.com/osism/testbed/netbox/devicetypes/Other/node.yml",
                                ".d..t...... src/github.com/osism/testbed/netbox/moduletypes/",
                                "<f..t...... src/github.com/osism/testbed/netbox/moduletypes/.gitkeep",
                                ".d..t...... src/github.com/osism/testbed/netbox/resources/",
                                "<f..t...... src/github.com/osism/testbed/netbox/resources/100-initialise.yml",
                                "<f..t...... src/github.com/osism/testbed/netbox/resources/200-rack-1000.yml",
                                "<f..t...... src/github.com/osism/testbed/netbox/resources/300-testbed-manager.yml",
                                "<f..t...... src/github.com/osism/testbed/netbox/resources/300-testbed-node-0.yml",
                                "<f..t...... src/github.com/osism/testbed/netbox/resources/300-testbed-node-1.yml",
                                "<f..t...... src/github.com/osism/testbed/netbox/resources/300-testbed-node-2.yml",
                                "<f..t...... src/github.com/osism/testbed/netbox/resources/300-testbed-node-3.yml",
                                "<f..t...... src/github.com/osism/testbed/netbox/resources/300-testbed-node-4.yml",
                                "<f..t...... src/github.com/osism/testbed/netbox/resources/300-testbed-node-5.yml",
                                "<f..t...... src/github.com/osism/testbed/netbox/resources/300-testbed-node-6.yml",
                                "<f..t...... src/github.com/osism/testbed/netbox/resources/300-testbed-node-7.yml",
                                "<f..t...... src/github.com/osism/testbed/netbox/resources/300-testbed-node-8.yml",
                                "<f..t...... src/github.com/osism/testbed/netbox/resources/300-testbed-node-9.yml",
                                "<f..t...... src/github.com/osism/testbed/netbox/resources/300-testbed-switch-0.yml",
                                "<f..t...... src/github.com/osism/testbed/netbox/resources/300-testbed-switch-1.yml",
                                "<f..t...... src/github.com/osism/testbed/netbox/resources/300-testbed-switch-2.yml",
                                "<f..t...... src/github.com/osism/testbed/netbox/resources/300-testbed-switch-3.yml",
                                "<f..t...... src/github.com/osism/testbed/netbox/resources/300-testbed-switch-oob.yml",
                                ".d..t...... src/github.com/osism/testbed/network/",
                                "<f..t...... src/github.com/osism/testbed/network/iptables.sh",
                                "<f..t...... src/github.com/osism/testbed/network/vxlan.sh",
                                ".d..t...... src/github.com/osism/testbed/playbooks/",
                                "<f..t...... src/github.com/osism/testbed/playbooks/cleanup.yml",
                                "<f..t...... src/github.com/osism/testbed/playbooks/deploy.yml",
                                "<f..t...... src/github.com/osism/testbed/playbooks/post.yml",
                                "<f..t...... src/github.com/osism/testbed/playbooks/pre.yml",
                                "<f..t...... src/github.com/osism/testbed/playbooks/update-stable.yml",
                                "<f..t...... src/github.com/osism/testbed/playbooks/upgrade-stable.yml",
                                "<f..t...... src/github.com/osism/testbed/playbooks/upgrade.yml",
                                ".d..t...... src/github.com/osism/testbed/playbooks/templates/",
                                "<f..t...... src/github.com/osism/testbed/playbooks/templates/local.env.j2",
                                ".d..t...... src/github.com/osism/testbed/playbooks/vars/",
                                "<f..t...... src/github.com/osism/testbed/playbooks/vars/mappings.yml",
                                "<f..t...... src/github.com/osism/testbed/playbooks/vars/repositories.yml",
                                ".d..t...... src/github.com/osism/testbed/scripts/",
                                "<f..t...... src/github.com/osism/testbed/scripts/bootstrap-services.sh",
                                "<f..t...... src/github.com/osism/testbed/scripts/bootstrap.sh",
                                "<f..t...... src/github.com/osism/testbed/scripts/check-ceph.sh",
                                "<f..t...... src/github.com/osism/testbed/scripts/check-services.sh",
                                "<f..t...... src/github.com/osism/testbed/scripts/check.sh",
                                "<f..t...... src/github.com/osism/testbed/scripts/deploy-baremetal.sh",
                                "<f..t...... src/github.com/osism/testbed/scripts/deploy-ceph.sh",
                                "<f..t...... src/github.com/osism/testbed/scripts/deploy-in-a-nutshell.sh",
                                "<f..t...... src/github.com/osism/testbed/scripts/deploy-manager.sh",
                                "<f..t...... src/github.com/osism/testbed/scripts/deploy-services.sh",
                                "<f..t...... src/github.com/osism/testbed/scripts/deploy.sh",
                                "<f..t...... src/github.com/osism/testbed/scripts/disable-ara.sh",
                                "<f..t...... src/github.com/osism/testbed/scripts/disable-local-registry.sh",
                                "<f..t...... src/github.com/osism/testbed/scripts/enable-ara.sh",
                                "<f..t...... src/github.com/osism/testbed/scripts/enable-resource-nodes.sh",
                                "<f..t...... src/github.com/osism/testbed/scripts/include.sh",
                                "<f..t...... src/github.com/osism/testbed/scripts/manager-version.sh",
                                "<f..t...... src/github.com/osism/testbed/scripts/prepare-ceph-configuration.sh",
                                "<f..t...... src/github.com/osism/testbed/scripts/prepare-wireguard-configuration.sh",
                                "<f..t...... src/github.com/osism/testbed/scripts/pull-images.sh",
                                "<f..t...... src/github.com/osism/testbed/scripts/set-ceph-version.sh",
                                "<f..t...... src/github.com/osism/testbed/scripts/set-docker-registry.sh",
                                "<f..t...... src/github.com/osism/testbed/scripts/set-kolla-namespace.sh",
                                "<f..t...... src/github.com/osism/testbed/scripts/set-manager-version.sh",
                                "<f..t...... src/github.com/osism/testbed/scripts/set-openstack-version.sh",
                                "<f..t...... src/github.com/osism/testbed/scripts/sync-configuration-repository.sh",
                                "<f..t...... src/github.com/osism/testbed/scripts/upgrade-ceph.sh",
                                "<f..t...... src/github.com/osism/testbed/scripts/upgrade-manager.sh",
                                "<f..t...... src/github.com/osism/testbed/scripts/upgrade-services.sh",
                                "<f..t...... src/github.com/osism/testbed/scripts/upgrade.sh",
                                "<f..t...... src/github.com/osism/testbed/scripts/wrapper-gilt.sh",
                                ".d..t...... src/github.com/osism/testbed/scripts/bootstrap/",
                                "<f..t...... src/github.com/osism/testbed/scripts/bootstrap/000-netbox.sh",
                                "<f..t...... src/github.com/osism/testbed/scripts/bootstrap/300-openstack.sh",
                                "<f..t...... src/github.com/osism/testbed/scripts/bootstrap/301-openstack-octavia-amhpora-image.sh",
                                "<f..t...... src/github.com/osism/testbed/scripts/bootstrap/302-openstack-k8s-clusterapi-images.sh",
                                ".d..t...... src/github.com/osism/testbed/scripts/check/",
                                "<f..t...... src/github.com/osism/testbed/scripts/check/100-ceph-with-ansible.sh",
                                "<f..t...... src/github.com/osism/testbed/scripts/check/100-ceph-with-rook.sh",
                                "<f..t...... src/github.com/osism/testbed/scripts/check/200-infrastructure.sh",
                                "<f..t...... src/github.com/osism/testbed/scripts/check/300-openstack.sh",
                                "<f..t...... src/github.com/osism/testbed/scripts/check/302-openstack-scs-compatible.sh",
                                "<f..t...... src/github.com/osism/testbed/scripts/check/302-openstack-with-tempest.sh",
                                "<f..t...... src/github.com/osism/testbed/scripts/check/303-prometheus-alert-status.sh",
                                "<f..t...... src/github.com/osism/testbed/scripts/check/500-kubernetes.sh",
                                ".d..t...... src/github.com/osism/testbed/scripts/customisations/",
                                "<f..t...... src/github.com/osism/testbed/scripts/customisations/external-api.sh",
                                ".d..t...... src/github.com/osism/testbed/scripts/deploy/",
                                "<f..t...... src/github.com/osism/testbed/scripts/deploy/000-manager.sh",
                                "<f..t...... src/github.com/osism/testbed/scripts/deploy/001-helpers.sh",
                                "<f..t...... src/github.com/osism/testbed/scripts/deploy/100-ceph-with-ansible.sh",
                                "<f..t...... src/github.com/osism/testbed/scripts/deploy/100-ceph-with-rook.sh",
                                "<f..t...... src/github.com/osism/testbed/scripts/deploy/200-infrastructure.sh",
                                "<f..t...... src/github.com/osism/testbed/scripts/deploy/300-openstack.sh",
                                "<f..t...... src/github.com/osism/testbed/scripts/deploy/310-openstack-extended.sh",
                                "<f..t...... src/github.com/osism/testbed/scripts/deploy/320-openstack-minimal.sh",
                                "<f..t...... src/github.com/osism/testbed/scripts/deploy/400-monitoring.sh",
                                "<f..t...... src/github.com/osism/testbed/scripts/deploy/500-kubernetes.sh",
                                "<f..t...... src/github.com/osism/testbed/scripts/deploy/510-clusterapi.sh",
                                ".d..t...... src/github.com/osism/testbed/scripts/upgrade/",
                                "<f..t...... src/github.com/osism/testbed/scripts/upgrade/100-ceph-with-ansible.sh",
                                "<f..t...... src/github.com/osism/testbed/scripts/upgrade/100-ceph-with-rook.sh",
                                "<f..t...... src/github.com/osism/testbed/scripts/upgrade/200-infrastructure.sh",
                                "<f..t...... src/github.com/osism/testbed/scripts/upgrade/300-openstack.sh",
                                "<f..t...... src/github.com/osism/testbed/scripts/upgrade/310-openstack-extended.sh",
                                "<f..t...... src/github.com/osism/testbed/scripts/upgrade/320-openstack-minimal.sh",
                                "<f..t...... src/github.com/osism/testbed/scripts/upgrade/400-monitoring.sh",
                                "<f..t...... src/github.com/osism/testbed/scripts/upgrade/500-kubernetes.sh",
                                "<f..t...... src/github.com/osism/testbed/scripts/upgrade/510-clusterapi.sh",
                                ".d..t...... src/github.com/osism/testbed/terraform/",
                                "<f..t...... src/github.com/osism/testbed/terraform/Makefile",
                                "<f..t...... src/github.com/osism/testbed/terraform/README.md",
                                "<f..t...... src/github.com/osism/testbed/terraform/backend.env.sample",
                                "<f..t...... src/github.com/osism/testbed/terraform/backend.tf.sample",
                                "<f..t...... src/github.com/osism/testbed/terraform/clouds-public.yaml",
                                "<f..t...... src/github.com/osism/testbed/terraform/clouds.yaml.sample",
                                "<f..t...... src/github.com/osism/testbed/terraform/local.env.sample",
                                "<f..t...... src/github.com/osism/testbed/terraform/secure.yaml.sample",
                                "<f..t...... src/github.com/osism/testbed/terraform/terraformrc.sample",
                                ".d..t...... src/github.com/osism/testbed/terraform/environments/",
                                "<f..t...... src/github.com/osism/testbed/terraform/environments/artcodix.tfvars",
                                "<f..t...... src/github.com/osism/testbed/terraform/environments/cah-f1a.tfvars",
                                "<f..t...... src/github.com/osism/testbed/terraform/environments/ci-centos-stream-9.tfvars",
                                "<f..t...... src/github.com/osism/testbed/terraform/environments/ci-debian-12.tfvars",
                                "<f..t...... src/github.com/osism/testbed/terraform/environments/ci-ubuntu-24.04.tfvars",
                                "<f..t...... src/github.com/osism/testbed/terraform/environments/ci.tfvars",
                                "<f..t...... src/github.com/osism/testbed/terraform/environments/cleura.tfvars",
                                "<f..t...... src/github.com/osism/testbed/terraform/environments/cloudpod.tfvars",
                                "<f..t...... src/github.com/osism/testbed/terraform/environments/default.tfvars",
                                "<f..t...... src/github.com/osism/testbed/terraform/environments/exigo.tfvars",
                                "<f..t...... src/github.com/osism/testbed/terraform/environments/gx-scs.tfvars",
                                "<f..t...... src/github.com/osism/testbed/terraform/environments/huaweicloud.tfvars",
                                "<f..t...... src/github.com/osism/testbed/terraform/environments/otc.tfvars",
                                "<f..t...... src/github.com/osism/testbed/terraform/environments/ovh.tfvars",
                                "<f..t...... src/github.com/osism/testbed/terraform/environments/pluscloudopen.tfvars",
                                "<f..t...... src/github.com/osism/testbed/terraform/environments/regiocloud.tfvars",
                                "<f..t...... src/github.com/osism/testbed/terraform/environments/wavestack.tfvars",
                                ".d..t...... src/github.com/osism/testbed/terraform/scripts/",
                                "<f..t...... src/github.com/osism/testbed/terraform/scripts/cleanup.py",
                                "*deleting   src/github.com/osism/testbed/.git/FETCH_HEAD",
                                "*deleting   src/github.com/osism/testbed/.git/COMMIT_EDITMSG",
                                "*deleting   src/github.com/osism/testbed/.git/refs/zuul/fetch",
                                "*deleting   src/github.com/osism/testbed/.git/refs/zuul/b28b7af69320201d1cf206ebf28373980add1451",
                                "*deleting   src/github.com/osism/testbed/.git/refs/zuul/",
                                "*deleting   src/github.com/osism/testbed/.git/refs/remotes/origin/main",
                                "*deleting   src/github.com/osism/testbed/.git/refs/remotes/origin/",
                                "*deleting   src/github.com/osism/testbed/.git/refs/remotes/",
                                "*deleting   src/github.com/osism/testbed/.git/refs/heads/main",
                                "*deleting   src/github.com/osism/testbed/.git/refs/heads/",
                                "*deleting   src/github.com/osism/testbed/ansible/ansible-manager-part-1.log",
                                "*deleting   src/github.com/osism/testbed/ansible/ansible-manager-part-0.log",
                                "*deleting   src/github.com/osism/testbed/ansible/.id_rsa.ci.pub",
                                "*deleting   src/github.com/osism/testbed/ansible/.id_rsa.ci",
                                "*deleting   src/github.com/osism/testbed/terraform/terraform.tfstate.d/ci/terraform.tfstate",
                                "*deleting   src/github.com/osism/testbed/terraform/terraform.tfstate.d/ci/",
                                "*deleting   src/github.com/osism/testbed/terraform/terraform.tfstate.d/",
                                "*deleting   src/github.com/osism/testbed/terraform/overrides/nodes_use_ephemeral_storage_override.tf",
                                "*deleting   src/github.com/osism/testbed/terraform/overrides/nodes_boot_from_volume_override.tf",
                                "*deleting   src/github.com/osism/testbed/terraform/overrides/nodes_boot_from_image_override.tf",
                                "*deleting   src/github.com/osism/testbed/terraform/overrides/neutron_router_enable_snat_override.tf",
                                "*deleting   src/github.com/osism/testbed/terraform/overrides/neutron_availability_zone_hints_router_override.tf",
                                "*deleting   src/github.com/osism/testbed/terraform/overrides/neutron_availability_zone_hints_network_override.tf",
                                "*deleting   src/github.com/osism/testbed/terraform/overrides/manager_boot_from_volume_override.tf",
                                "*deleting   src/github.com/osism/testbed/terraform/overrides/manager_boot_from_image_override.tf",
                                "*deleting   src/github.com/osism/testbed/terraform/overrides/",
                                "*deleting   src/github.com/osism/testbed/terraform/customisations/neutron_floatingip_custom.tf",
                                "*deleting   src/github.com/osism/testbed/terraform/customisations/external_api_custom.tf",
                                "*deleting   src/github.com/osism/testbed/terraform/customisations/default_custom.tf",
                                "*deleting   src/github.com/osism/testbed/terraform/customisations/access_ipv6_custom.tf",
                                "*deleting   src/github.com/osism/testbed/terraform/customisations/access_ipv4_custom.tf",
                                "*deleting   src/github.com/osism/testbed/terraform/customisations/access_floatingip_custom.tf",
                                "*deleting   src/github.com/osism/testbed/terraform/customisations/",
                                "*deleting   src/github.com/osism/testbed/terraform/.terraform/providers/registry.opentofu.org/terraform-provider-openstack/openstack/3.4.0/linux_amd64/terraform-provider-openstack_v3.4.0",
                                "*deleting   src/github.com/osism/testbed/terraform/.terraform/providers/registry.opentofu.org/terraform-provider-openstack/openstack/3.4.0/linux_amd64/README.md",
                                "*deleting   src/github.com/osism/testbed/terraform/.terraform/providers/registry.opentofu.org/terraform-provider-openstack/openstack/3.4.0/linux_amd64/LICENSE",
                                "*deleting   src/github.com/osism/testbed/terraform/.terraform/providers/registry.opentofu.org/terraform-provider-openstack/openstack/3.4.0/linux_amd64/CHANGELOG.md",
                                "*deleting   src/github.com/osism/testbed/terraform/.terraform/providers/registry.opentofu.org/terraform-provider-openstack/openstack/3.4.0/linux_amd64/",
                                "*deleting   src/github.com/osism/testbed/terraform/.terraform/providers/registry.opentofu.org/terraform-provider-openstack/openstack/3.4.0/linux_amd64.lock",
                                "*deleting   src/github.com/osism/testbed/terraform/.terraform/providers/registry.opentofu.org/terraform-provider-openstack/openstack/3.4.0/",
                                "*deleting   src/github.com/osism/testbed/terraform/.terraform/providers/registry.opentofu.org/terraform-provider-openstack/openstack/",
                                "*deleting   src/github.com/osism/testbed/terraform/.terraform/providers/registry.opentofu.org/terraform-provider-openstack/",
                                "*deleting   src/github.com/osism/testbed/terraform/.terraform/providers/registry.opentofu.org/hashicorp/null/3.2.4/linux_amd64/terraform-provider-null",
                                "*deleting   src/github.com/osism/testbed/terraform/.terraform/providers/registry.opentofu.org/hashicorp/null/3.2.4/linux_amd64/README.md",
                                "*deleting   src/github.com/osism/testbed/terraform/.terraform/providers/registry.opentofu.org/hashicorp/null/3.2.4/linux_amd64/LICENSE",
                                "*deleting   src/github.com/osism/testbed/terraform/.terraform/providers/registry.opentofu.org/hashicorp/null/3.2.4/linux_amd64/CHANGELOG.md",
                                "*deleting   src/github.com/osism/testbed/terraform/.terraform/providers/registry.opentofu.org/hashicorp/null/3.2.4/linux_amd64/",
                                "*deleting   src/github.com/osism/testbed/terraform/.terraform/providers/registry.opentofu.org/hashicorp/null/3.2.4/linux_amd64.lock",
                                "*deleting   src/github.com/osism/testbed/terraform/.terraform/providers/registry.opentofu.org/hashicorp/null/3.2.4/",
                                "*deleting   src/github.com/osism/testbed/terraform/.terraform/providers/registry.opentofu.org/hashicorp/null/",
                                "*deleting   src/github.com/osism/testbed/terraform/.terraform/providers/registry.opentofu.org/hashicorp/local/2.7.0/linux_amd64/terraform-provider-local",
                                "*deleting   src/github.com/osism/testbed/terraform/.terraform/providers/registry.opentofu.org/hashicorp/local/2.7.0/linux_amd64/README.md",
                                "*deleting   src/github.com/osism/testbed/terraform/.terraform/providers/registry.opentofu.org/hashicorp/local/2.7.0/linux_amd64/LICENSE",
                                "*deleting   src/github.com/osism/testbed/terraform/.terraform/providers/registry.opentofu.org/hashicorp/local/2.7.0/linux_amd64/CHANGELOG.md",
                                "*deleting   src/github.com/osism/testbed/terraform/.terraform/providers/registry.opentofu.org/hashicorp/local/2.7.0/linux_amd64/",
                                "*deleting   src/github.com/osism/testbed/terraform/.terraform/providers/registry.opentofu.org/hashicorp/local/2.7.0/linux_amd64.lock",
                                "*deleting   src/github.com/osism/testbed/terraform/.terraform/providers/registry.opentofu.org/hashicorp/local/2.7.0/",
                                "*deleting   src/github.com/osism/testbed/terraform/.terraform/providers/registry.opentofu.org/hashicorp/local/",
                                "*deleting   src/github.com/osism/testbed/terraform/.terraform/providers/registry.opentofu.org/hashicorp/",
                                "*deleting   src/github.com/osism/testbed/terraform/.terraform/providers/registry.opentofu.org/",
                                "*deleting   src/github.com/osism/testbed/terraform/.terraform/providers/",
                                "*deleting   src/github.com/osism/testbed/terraform/.terraform/environment",
                                "*deleting   src/github.com/osism/testbed/terraform/.terraform/",
                                "*deleting   src/github.com/osism/testbed/terraform/variables.tf",
                                "*deleting   src/github.com/osism/testbed/terraform/provider.tf",
                                "*deleting   src/github.com/osism/testbed/terraform/outputs.tf",
                                "*deleting   src/github.com/osism/testbed/terraform/nova.tf",
                                "*deleting   src/github.com/osism/testbed/terraform/nodes_boot_from_volume_override.tf",
                                "*deleting   src/github.com/osism/testbed/terraform/nodes.tf",
                                "*deleting   src/github.com/osism/testbed/terraform/neutron_floatingip_custom.tf",
                                "*deleting   src/github.com/osism/testbed/terraform/neutron_availability_zone_hints_router_override.tf",
                                "*deleting   src/github.com/osism/testbed/terraform/neutron_availability_zone_hints_network_override.tf",
                                "*deleting   src/github.com/osism/testbed/terraform/neutron.tf",
                                "*deleting   src/github.com/osism/testbed/terraform/manager_boot_from_volume_override.tf",
                                "*deleting   src/github.com/osism/testbed/terraform/manager.tf",
                                "*deleting   src/github.com/osism/testbed/terraform/main.tf",
                                "*deleting   src/github.com/osism/testbed/terraform/local.env",
                                "*deleting   src/github.com/osism/testbed/terraform/inventory.ci",
                                "*deleting   src/github.com/osism/testbed/terraform/default_custom.tf",
                                "*deleting   src/github.com/osism/testbed/terraform/data.tf",
                                "*deleting   src/github.com/osism/testbed/terraform/ci.auto.tfvars",
                                "*deleting   src/github.com/osism/testbed/terraform/access_floatingip_custom.tf",
                                "*deleting   src/github.com/osism/testbed/terraform/.terraform.lock.hcl",
                                "*deleting   src/github.com/osism/testbed/terraform/.id_rsa.ci.pub",
                                "*deleting   src/github.com/osism/testbed/terraform/.id_rsa.ci",
                                "*deleting   src/github.com/osism/testbed/terraform/.deploy.ci",
                                "*deleting   src/github.com/osism/testbed/terraform/.MANAGER_ADDRESS.ci"
                            ]
                        }
                    },
                    "role": {
                        "id": "0242ac1f-6606-c045-40cc-00000000002a",
                        "name": "prepare-workspace-log",
                        "path": "/var/lib/zuul/builds/7573dac089bc49b5b355588f880c7711/trusted/project_0/github.com/osism/zuul-config/roles/prepare-workspace-log"
                    },
                    "task": {
                        "duration": {
                            "end": "2026-02-24T00:01:09.090551Z",
                            "start": "2026-02-24T00:00:36.796083Z"
                        },
                        "id": "0242ac1f-6606-c045-40cc-00000000002d",
                        "name": "Synchronize src repos to workspace directory."
                    }
                }
            ]
        }
    ],
    "stats": {
        "orchestrator": {
            "changed": 9,
            "failures": 0,
            "ignored": 0,
            "ok": 26,
            "rescued": 0,
            "skipped": 3,
            "unreachable": 0
        }
    },
    "trusted": true
},
{
    "branch": "main",
    "index": "1",
    "phase": "pre",
    "playbook": "github.com/osism/testbed/playbooks/pre.yml",
    "plays": [
        {
            "play": {
                "duration": {
                    "end": "2026-02-24T00:01:19.331926Z",
                    "start": "2026-02-24T00:01:10.237172Z"
                },
                "id": "0242ac1f-6606-2f35-8944-000000000002",
                "name": "Pre play"
            },
            "tasks": [
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "ansible.builtin.set_fact",
                            "ansible_facts": {
                                "cloud": "ci"
                            },
                            "changed": false
                        }
                    },
                    "task": {
                        "duration": {
                            "end": "2026-02-24T00:01:10.312549Z",
                            "start": "2026-02-24T00:01:10.259734Z"
                        },
                        "id": "0242ac1f-6606-2f35-8944-000000000004",
                        "name": "Set cloud fact (Zuul deployment)"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "ansible.builtin.set_fact",
                            "changed": false,
                            "false_condition": "'nodepool' not in hostvars[groups['all'][0]]",
                            "skip_reason": "Conditional result was False",
                            "skipped": true
                        }
                    },
                    "task": {
                        "duration": {
                            "end": "2026-02-24T00:01:10.384712Z",
                            "start": "2026-02-24T00:01:10.329710Z"
                        },
                        "id": "0242ac1f-6606-2f35-8944-000000000005",
                        "name": "Set cloud fact (local deployment)"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "ansible.builtin.shell",
                            "changed": true,
                            "cmd": "OS_CLOUD=ci ~/venv/bin/python3 cleanup.py\n",
                            "delta": "0:00:01.540882",
                            "end": "2026-02-24 00:01:12.660529",
                            "failed_when_result": false,
                            "invocation": {
                                "module_args": {
                                    "_raw_params": "OS_CLOUD=ci ~/venv/bin/python3 cleanup.py\n",
                                    "_uses_shell": true,
                                    "argv": null,
                                    "chdir": "/home/zuul-testbed02/src/github.com/osism/testbed/terraform/scripts",
                                    "creates": null,
                                    "executable": null,
                                    "expand_argument_vars": true,
                                    "removes": null,
                                    "stdin": null,
                                    "stdin_add_newline": true,
                                    "strip_empty_ends": true,
                                    "zuul_ansible_split_streams": false,
                                    "zuul_log_id": "0242ac1f-6606-2f35-8944-000000000006-1-orchestrator",
                                    "zuul_output_max_bytes": 1073741824
                                }
                            },
                            "msg": "",
                            "rc": 0,
                            "start": "2026-02-24 00:01:11.119647",
                            "stderr": "",
                            "stderr_lines": [],
                            "stdout": "2026-02-24 00:01:11 - clean up servers\n2026-02-24 00:01:12 - clean up keypairs\n2026-02-24 00:01:12 - wait for servers to be gone\n2026-02-24 00:01:12 - clean up ports\n2026-02-24 00:01:12 - clean up volumes\n2026-02-24 00:01:12 - disconnect routers\n2026-02-24 00:01:12 - clean up subnets\n2026-02-24 00:01:12 - clean up networks\n2026-02-24 00:01:12 - clean up security groups\n2026-02-24 00:01:12 - clean up floating ips\n2026-02-24 00:01:12 - clean up routers",
                            "stdout_lines": [
                                "2026-02-24 00:01:11 - clean up servers",
                                "2026-02-24 00:01:12 - clean up keypairs",
                                "2026-02-24 00:01:12 - wait for servers to be gone",
                                "2026-02-24 00:01:12 - clean up ports",
                                "2026-02-24 00:01:12 - clean up volumes",
                                "2026-02-24 00:01:12 - disconnect routers",
                                "2026-02-24 00:01:12 - clean up subnets",
                                "2026-02-24 00:01:12 - clean up networks",
                                "2026-02-24 00:01:12 - clean up security groups",
                                "2026-02-24 00:01:12 - clean up floating ips",
                                "2026-02-24 00:01:12 - clean up routers"
                            ],
                            "zuul_log_id": "0242ac1f-6606-2f35-8944-000000000006-1-orchestrator"
                        }
                    },
                    "task": {
                        "duration": {
                            "end": "2026-02-24T00:01:12.978156Z",
                            "start": "2026-02-24T00:01:10.461873Z"
                        },
                        "id": "0242ac1f-6606-2f35-8944-000000000006",
                        "name": "Clean the cloud environment"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "ansible.builtin.unarchive",
                            "changed": false,
                            "dest": "/home/zuul-testbed02",
                            "gid": 1003,
                            "group": "zuul-testbed02",
                            "handler": "ZipArchive",
                            "invocation": {
                                "module_args": {
                                    "attributes": null,
                                    "copy": true,
                                    "creates": null,
                                    "decrypt": true,
                                    "dest": "/home/zuul-testbed02",
                                    "exclude": [],
                                    "extra_opts": [],
                                    "group": null,
                                    "include": [
                                        "tofu"
                                    ],
                                    "io_buffer_size": 65536,
                                    "keep_newer": false,
                                    "list_files": false,
                                    "mode": null,
                                    "owner": null,
                                    "remote_src": true,
                                    "selevel": null,
                                    "serole": null,
                                    "setype": null,
                                    "seuser": null,
                                    "src": "https://github.com/opentofu/opentofu/releases/download/v1.11.5/tofu_1.11.5_linux_amd64.zip",
                                    "unsafe_writes": false,
                                    "validate_certs": true
                                }
                            },
                            "mode": "0755",
                            "owner": "zuul-testbed02",
                            "size": 4096,
                            "src": "/home/zuul-testbed02/.ansible/tmp/ansible-tmp-1771891273.0484061-10-60648607901170/tofu_1.11.5_linux_amd644odiui6n.zip",
                            "state": "directory",
                            "uid": 1002
                        }
                    },
                    "task": {
                        "duration": {
                            "end": "2026-02-24T00:01:17.009582Z",
                            "start": "2026-02-24T00:01:12.993510Z"
                        },
                        "id": "0242ac1f-6606-2f35-8944-000000000007",
                        "name": "Extract tofu binary"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "ansible.builtin.command",
                            "changed": true,
                            "cmd": [
                                "rsync",
                                "-avz",
                                "/home/zuul-testbed02/src/github.com/osism/terraform-base/testbed-default/",
                                "/home/zuul-testbed02/src/github.com/osism/testbed/terraform"
                            ],
                            "delta": "0:00:00.067453",
                            "end": "2026-02-24 00:01:17.580025",
                            "invocation": {
                                "module_args": {
                                    "_raw_params": "rsync -avz /home/zuul-testbed02/src/github.com/osism/terraform-base/testbed-default/ /home/zuul-testbed02/src/github.com/osism/testbed/terraform",
                                    "_uses_shell": false,
                                    "argv": null,
                                    "chdir": null,
                                    "creates": null,
                                    "executable": null,
                                    "expand_argument_vars": true,
                                    "removes": null,
                                    "stdin": null,
                                    "stdin_add_newline": true,
                                    "strip_empty_ends": true,
                                    "zuul_ansible_split_streams": false,
                                    "zuul_log_id": "0242ac1f-6606-2f35-8944-000000000008-1-orchestrator",
                                    "zuul_output_max_bytes": 1073741824
                                }
                            },
                            "msg": "",
                            "rc": 0,
                            "start": "2026-02-24 00:01:17.512572",
                            "stderr": "",
                            "stderr_lines": [],
                            "stdout": "sending incremental file list\ndata.tf\nmain.tf\nmanager.tf\nneutron.tf\nnodes.tf\nnova.tf\noutputs.tf\nprovider.tf\nvariables.tf\ncustomisations/\ncustomisations/access_floatingip_custom.tf\ncustomisations/access_ipv4_custom.tf\ncustomisations/access_ipv6_custom.tf\ncustomisations/default_custom.tf\ncustomisations/external_api_custom.tf\ncustomisations/neutron_floatingip_custom.tf\noverrides/\noverrides/manager_boot_from_image_override.tf\noverrides/manager_boot_from_volume_override.tf\noverrides/neutron_availability_zone_hints_network_override.tf\noverrides/neutron_availability_zone_hints_router_override.tf\noverrides/neutron_router_enable_snat_override.tf\noverrides/nodes_boot_from_image_override.tf\noverrides/nodes_boot_from_volume_override.tf\noverrides/nodes_use_ephemeral_storage_override.tf\n\nsent 6,924 bytes  received 466 bytes  14,780.00 bytes/sec\ntotal size is 26,532  speedup is 3.59",
                            "stdout_lines": [
                                "sending incremental file list",
                                "data.tf",
                                "main.tf",
                                "manager.tf",
                                "neutron.tf",
                                "nodes.tf",
                                "nova.tf",
                                "outputs.tf",
                                "provider.tf",
                                "variables.tf",
                                "customisations/",
                                "customisations/access_floatingip_custom.tf",
                                "customisations/access_ipv4_custom.tf",
                                "customisations/access_ipv6_custom.tf",
                                "customisations/default_custom.tf",
                                "customisations/external_api_custom.tf",
                                "customisations/neutron_floatingip_custom.tf",
                                "overrides/",
                                "overrides/manager_boot_from_image_override.tf",
                                "overrides/manager_boot_from_volume_override.tf",
                                "overrides/neutron_availability_zone_hints_network_override.tf",
                                "overrides/neutron_availability_zone_hints_router_override.tf",
                                "overrides/neutron_router_enable_snat_override.tf",
                                "overrides/nodes_boot_from_image_override.tf",
                                "overrides/nodes_boot_from_volume_override.tf",
                                "overrides/nodes_use_ephemeral_storage_override.tf",
                                "",
                                "sent 6,924 bytes  received 466 bytes  14,780.00 bytes/sec",
                                "total size is 26,532  speedup is 3.59"
                            ],
                            "zuul_log_id": "0242ac1f-6606-2f35-8944-000000000008-1-orchestrator"
                        }
                    },
                    "task": {
                        "duration": {
                            "end": "2026-02-24T00:01:18.074529Z",
                            "start": "2026-02-24T00:01:17.063974Z"
                        },
                        "id": "0242ac1f-6606-2f35-8944-000000000008",
                        "name": "Sync terraform blueprint"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "ansible.builtin.template",
                            "changed": true,
                            "checksum": "0486e4f954d1035566c9d518cd9e10adbaddcf54",
                            "dest": "/home/zuul-testbed02/src/github.com/osism/testbed/terraform/local.env",
                            "diff": [],
                            "gid": 1003,
                            "group": "zuul-testbed02",
                            "invocation": {
                                "module_args": {
                                    "_original_basename": "local.env.j2",
                                    "attributes": null,
                                    "backup": false,
                                    "checksum": "0486e4f954d1035566c9d518cd9e10adbaddcf54",
                                    "content": null,
                                    "dest": "/home/zuul-testbed02/src/github.com/osism/testbed/terraform/local.env",
                                    "directory_mode": null,
                                    "follow": false,
                                    "force": true,
                                    "group": null,
                                    "local_follow": null,
                                    "mode": "0644",
                                    "owner": null,
                                    "remote_src": null,
                                    "selevel": null,
                                    "serole": null,
                                    "setype": null,
                                    "seuser": null,
                                    "src": "/home/zuul-testbed02/.ansible/tmp/ansible-tmp-1771891278.178983-19-186152541128658/source",
                                    "unsafe_writes": false,
                                    "validate": null
                                }
                            },
                            "md5sum": "9d6b790c9266c88e68834b29cbcdcbf8",
                            "mode": "0644",
                            "owner": "zuul-testbed02",
                            "size": 42,
                            "src": "/home/zuul-testbed02/.ansible/tmp/ansible-tmp-1771891278.178983-19-186152541128658/source",
                            "state": "file",
                            "uid": 1002
                        }
                    },
                    "task": {
                        "duration": {
                            "end": "2026-02-24T00:01:19.331926Z",
                            "start": "2026-02-24T00:01:18.084930Z"
                        },
                        "id": "0242ac1f-6606-2f35-8944-000000000009",
                        "name": "Create local.env file"
                    }
                }
            ]
        }
    ],
    "stats": {
        "orchestrator": {
            "changed": 3,
            "failures": 0,
            "ignored": 0,
            "ok": 5,
            "rescued": 0,
            "skipped": 1,
            "unreachable": 0
        }
    },
    "trusted": false
},
{
    "branch": "main",
    "index": "0",
    "phase": "run",
    "playbook": "github.com/osism/testbed/playbooks/deploy.yml",
    "plays": [
        {
            "play": {
                "duration": {
                    "end": "2026-02-24T01:41:35.591931Z",
                    "start": "2026-02-24T00:01:21.232858Z"
                },
                "id": "0242ac1f-6606-1b2b-a63a-000000000002",
                "name": "Deploy testbed"
            },
            "tasks": [
                {
                    "hosts": {
                        "orchestrator": {
                            "_ceph_version": "reef",
                            "action": "ansible.builtin.debug",
                            "changed": false
                        }
                    },
                    "task": {
                        "duration": {
                            "end": "2026-02-24T00:01:21.385918Z",
                            "start": "2026-02-24T00:01:21.272335Z"
                        },
                        "id": "0242ac1f-6606-1b2b-a63a-000000000004",
                        "name": "Print used ceph version"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "_openstack_version": "2024.2",
                            "action": "ansible.builtin.debug",
                            "changed": false
                        }
                    },
                    "task": {
                        "duration": {
                            "end": "2026-02-24T00:01:21.484963Z",
                            "start": "2026-02-24T00:01:21.399425Z"
                        },
                        "id": "0242ac1f-6606-1b2b-a63a-000000000005",
                        "name": "Print used openstack version"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "_manager_version": "latest",
                            "action": "ansible.builtin.debug",
                            "changed": false
                        }
                    },
                    "task": {
                        "duration": {
                            "end": "2026-02-24T00:01:21.614499Z",
                            "start": "2026-02-24T00:01:21.501048Z"
                        },
                        "id": "0242ac1f-6606-1b2b-a63a-000000000006",
                        "name": "Print used manager version"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "ansible.builtin.set_fact",
                            "ansible_facts": {
                                "cloud": "ci",
                                "image_username": "ubuntu"
                            },
                            "changed": false
                        }
                    },
                    "task": {
                        "duration": {
                            "end": "2026-02-24T00:01:21.761796Z",
                            "start": "2026-02-24T00:01:21.622365Z"
                        },
                        "id": "0242ac1f-6606-1b2b-a63a-000000000007",
                        "name": "Set facts (Zuul deployment)"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "ansible.builtin.set_fact",
                            "changed": false,
                            "false_condition": "'nodepool' not in hostvars[groups['all'][0]]",
                            "skip_reason": "Conditional result was False",
                            "skipped": true
                        }
                    },
                    "task": {
                        "duration": {
                            "end": "2026-02-24T00:01:21.837068Z",
                            "start": "2026-02-24T00:01:21.781615Z"
                        },
                        "id": "0242ac1f-6606-1b2b-a63a-000000000008",
                        "name": "Set facts (local deployment)"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "ansible.builtin.command",
                            "changed": true,
                            "cmd": [
                                "make",
                                "ENVIRONMENT=ci-ubuntu-24.04",
                                "CLOUD=ci",
                                "IMAGE_USERNAME=ubuntu",
                                "TERRAFORM=/home/zuul-testbed02/tofu",
                                "VERSION_CEPH=reef",
                                "VERSION_MANAGER=latest",
                                "VERSION_OPENSTACK=2024.2",
                                "TEMPEST=true",
                                "CEPH_STACK=ceph-ansible",
                                "create"
                            ],
                            "delta": "0:01:09.296729",
                            "end": "2026-02-24 00:02:31.892227",
                            "invocation": {
                                "module_args": {
                                    "_raw_params": "make\n  ENVIRONMENT=ci-ubuntu-24.04\n  CLOUD=ci\n  IMAGE_USERNAME=ubuntu\n  TERRAFORM=/home/zuul-testbed02/tofu\n  VERSION_CEPH=reef\n  VERSION_MANAGER=latest\n  VERSION_OPENSTACK=2024.2\n  TEMPEST=true\n  CEPH_STACK=ceph-ansible\n  create\n",
                                    "_uses_shell": false,
                                    "argv": null,
                                    "chdir": "/home/zuul-testbed02/src/github.com/osism/testbed/terraform",
                                    "creates": null,
                                    "executable": null,
                                    "expand_argument_vars": true,
                                    "removes": null,
                                    "stdin": null,
                                    "stdin_add_newline": true,
                                    "strip_empty_ends": true,
                                    "zuul_ansible_split_streams": false,
                                    "zuul_log_id": "0242ac1f-6606-1b2b-a63a-000000000009-1-orchestrator",
                                    "zuul_output_max_bytes": 1073741824
                                }
                            },
                            "msg": "",
                            "rc": 0,
                            "start": "2026-02-24 00:01:22.595498",
                            "stderr": "",
                            "stderr_lines": [],
                            "stdout": "\nInitializing the backend...\n\nInitializing provider plugins...\n- terraform.io/builtin/terraform is built in to OpenTofu\n- Finding hashicorp/local versions matching \">= 2.2.0\"...\n- Finding latest version of hashicorp/null...\n- Finding terraform-provider-openstack/openstack versions matching \">= 1.53.0\"...\n- Installing hashicorp/local v2.7.0...\n- Installed hashicorp/local v2.7.0 (signed, key ID 0C0AF313E5FD9F80)\n- Installing hashicorp/null v3.2.4...\n- Installed hashicorp/null v3.2.4 (signed, key ID 0C0AF313E5FD9F80)\n- Installing terraform-provider-openstack/openstack v3.4.0...\n- Installed terraform-provider-openstack/openstack v3.4.0 (signed, key ID 4F80527A391BEFD2)\n\nProviders are signed by their developers.\nIf you'd like to know more about provider signing, you can read about it here:\nhttps://opentofu.org/docs/cli/plugins/signing/\n\nOpenTofu has created a lock file .terraform.lock.hcl to record the provider\nselections it made above. Include this file in your version control repository\nso that OpenTofu can guarantee to make the same selections by default when\nyou run \"tofu init\" in the future.\n\nOpenTofu has been successfully initialized!\n\nYou may now begin working with OpenTofu. Try running \"tofu plan\" to see\nany changes that are required for your infrastructure. All OpenTofu commands\nshould now work.\n\nIf you ever set or change modules or backend configuration for OpenTofu,\nrerun this command to reinitialize your working directory. If you forget, other\ncommands will detect it and remind you to do so if necessary.\nCreated and switched to workspace \"ci\"!\n\nYou're now on a new, empty workspace. Workspaces isolate their state,\nso if you run \"tofu plan\" OpenTofu will not see any existing state\nfor this configuration.\nci.auto.tfvars\ndefault_custom.tf\ndata.openstack_networking_network_v2.public: Reading...\ndata.openstack_networking_network_v2.public: Read complete after 1s [id=e6be7364-bfd8-4de7-8120-8f41c69a139a]\n\nOpenTofu used the selected providers to generate the following execution\nplan. Resource actions are indicated with the following symbols:\n  + create\n <= read (data resources)\n\nOpenTofu will perform the following actions:\n\n  # data.openstack_images_image_v2.image will be read during apply\n  # (config refers to values not yet known)\n <= data \"openstack_images_image_v2\" \"image\" {\n      + checksum    = (known after apply)\n      + created_at  = (known after apply)\n      + file        = (known after apply)\n      + id          = (known after apply)\n      + metadata    = (known after apply)\n      + min_disk_gb = (known after apply)\n      + min_ram_mb  = (known after apply)\n      + most_recent = true\n      + name        = (known after apply)\n      + protected   = (known after apply)\n      + region      = (known after apply)\n      + schema      = (known after apply)\n      + size_bytes  = (known after apply)\n      + tags        = (known after apply)\n      + updated_at  = (known after apply)\n    }\n\n  # data.openstack_images_image_v2.image_node will be read during apply\n  # (config refers to values not yet known)\n <= data \"openstack_images_image_v2\" \"image_node\" {\n      + checksum    = (known after apply)\n      + created_at  = (known after apply)\n      + file        = (known after apply)\n      + id          = (known after apply)\n      + metadata    = (known after apply)\n      + min_disk_gb = (known after apply)\n      + min_ram_mb  = (known after apply)\n      + most_recent = true\n      + name        = (known after apply)\n      + protected   = (known after apply)\n      + region      = (known after apply)\n      + schema      = (known after apply)\n      + size_bytes  = (known after apply)\n      + tags        = (known after apply)\n      + updated_at  = (known after apply)\n    }\n\n  # local_file.MANAGER_ADDRESS will be created\n  + resource \"local_file\" \"MANAGER_ADDRESS\" {\n      + content              = (known after apply)\n      + content_base64sha256 = (known after apply)\n      + content_base64sha512 = (known after apply)\n      + content_md5          = (known after apply)\n      + content_sha1         = (known after apply)\n      + content_sha256       = (known after apply)\n      + content_sha512       = (known after apply)\n      + directory_permission = \"0777\"\n      + file_permission      = \"0644\"\n      + filename             = \".MANAGER_ADDRESS.ci\"\n      + id                   = (known after apply)\n    }\n\n  # local_file.id_rsa_pub will be created\n  + resource \"local_file\" \"id_rsa_pub\" {\n      + content              = (known after apply)\n      + content_base64sha256 = (known after apply)\n      + content_base64sha512 = (known after apply)\n      + content_md5          = (known after apply)\n      + content_sha1         = (known after apply)\n      + content_sha256       = (known after apply)\n      + content_sha512       = (known after apply)\n      + directory_permission = \"0777\"\n      + file_permission      = \"0644\"\n      + filename             = \".id_rsa.ci.pub\"\n      + id                   = (known after apply)\n    }\n\n  # local_file.inventory will be created\n  + resource \"local_file\" \"inventory\" {\n      + content              = (known after apply)\n      + content_base64sha256 = (known after apply)\n      + content_base64sha512 = (known after apply)\n      + content_md5          = (known after apply)\n      + content_sha1         = (known after apply)\n      + content_sha256       = (known after apply)\n      + content_sha512       = (known after apply)\n      + directory_permission = \"0777\"\n      + file_permission      = \"0644\"\n      + filename             = \"inventory.ci\"\n      + id                   = (known after apply)\n    }\n\n  # local_sensitive_file.id_rsa will be created\n  + resource \"local_sensitive_file\" \"id_rsa\" {\n      + content              = (sensitive value)\n      + content_base64sha256 = (known after apply)\n      + content_base64sha512 = (known after apply)\n      + content_md5          = (known after apply)\n      + content_sha1         = (known after apply)\n      + content_sha256       = (known after apply)\n      + content_sha512       = (known after apply)\n      + directory_permission = \"0700\"\n      + file_permission      = \"0600\"\n      + filename             = \".id_rsa.ci\"\n      + id                   = (known after apply)\n    }\n\n  # null_resource.node_semaphore will be created\n  + resource \"null_resource\" \"node_semaphore\" {\n      + id = (known after apply)\n    }\n\n  # openstack_blockstorage_volume_v3.manager_base_volume[0] will be created\n  + resource \"openstack_blockstorage_volume_v3\" \"manager_base_volume\" {\n      + attachment           = (known after apply)\n      + availability_zone    = \"nova\"\n      + id                   = (known after apply)\n      + image_id             = (known after apply)\n      + metadata             = (known after apply)\n      + name                 = \"testbed-volume-manager-base\"\n      + region               = (known after apply)\n      + size                 = 80\n      + volume_retype_policy = \"never\"\n      + volume_type          = \"ssd\"\n    }\n\n  # openstack_blockstorage_volume_v3.node_base_volume[0] will be created\n  + resource \"openstack_blockstorage_volume_v3\" \"node_base_volume\" {\n      + attachment           = (known after apply)\n      + availability_zone    = \"nova\"\n      + id                   = (known after apply)\n      + image_id             = (known after apply)\n      + metadata             = (known after apply)\n      + name                 = \"testbed-volume-0-node-base\"\n      + region               = (known after apply)\n      + size                 = 80\n      + volume_retype_policy = \"never\"\n      + volume_type          = \"ssd\"\n    }\n\n  # openstack_blockstorage_volume_v3.node_base_volume[1] will be created\n  + resource \"openstack_blockstorage_volume_v3\" \"node_base_volume\" {\n      + attachment           = (known after apply)\n      + availability_zone    = \"nova\"\n      + id                   = (known after apply)\n      + image_id             = (known after apply)\n      + metadata             = (known after apply)\n      + name                 = \"testbed-volume-1-node-base\"\n      + region               = (known after apply)\n      + size                 = 80\n      + volume_retype_policy = \"never\"\n      + volume_type          = \"ssd\"\n    }\n\n  # openstack_blockstorage_volume_v3.node_base_volume[2] will be created\n  + resource \"openstack_blockstorage_volume_v3\" \"node_base_volume\" {\n      + attachment           = (known after apply)\n      + availability_zone    = \"nova\"\n      + id                   = (known after apply)\n      + image_id             = (known after apply)\n      + metadata             = (known after apply)\n      + name                 = \"testbed-volume-2-node-base\"\n      + region               = (known after apply)\n      + size                 = 80\n      + volume_retype_policy = \"never\"\n      + volume_type          = \"ssd\"\n    }\n\n  # openstack_blockstorage_volume_v3.node_base_volume[3] will be created\n  + resource \"openstack_blockstorage_volume_v3\" \"node_base_volume\" {\n      + attachment           = (known after apply)\n      + availability_zone    = \"nova\"\n      + id                   = (known after apply)\n      + image_id             = (known after apply)\n      + metadata             = (known after apply)\n      + name                 = \"testbed-volume-3-node-base\"\n      + region               = (known after apply)\n      + size                 = 80\n      + volume_retype_policy = \"never\"\n      + volume_type          = \"ssd\"\n    }\n\n  # openstack_blockstorage_volume_v3.node_base_volume[4] will be created\n  + resource \"openstack_blockstorage_volume_v3\" \"node_base_volume\" {\n      + attachment           = (known after apply)\n      + availability_zone    = \"nova\"\n      + id                   = (known after apply)\n      + image_id             = (known after apply)\n      + metadata             = (known after apply)\n      + name                 = \"testbed-volume-4-node-base\"\n      + region               = (known after apply)\n      + size                 = 80\n      + volume_retype_policy = \"never\"\n      + volume_type          = \"ssd\"\n    }\n\n  # openstack_blockstorage_volume_v3.node_base_volume[5] will be created\n  + resource \"openstack_blockstorage_volume_v3\" \"node_base_volume\" {\n      + attachment           = (known after apply)\n      + availability_zone    = \"nova\"\n      + id                   = (known after apply)\n      + image_id             = (known after apply)\n      + metadata             = (known after apply)\n      + name                 = \"testbed-volume-5-node-base\"\n      + region               = (known after apply)\n      + size                 = 80\n      + volume_retype_policy = \"never\"\n      + volume_type          = \"ssd\"\n    }\n\n  # openstack_blockstorage_volume_v3.node_volume[0] will be created\n  + resource \"openstack_blockstorage_volume_v3\" \"node_volume\" {\n      + attachment           = (known after apply)\n      + availability_zone    = \"nova\"\n      + id                   = (known after apply)\n      + metadata             = (known after apply)\n      + name                 = \"testbed-volume-0-node-3\"\n      + region               = (known after apply)\n      + size                 = 20\n      + volume_retype_policy = \"never\"\n      + volume_type          = \"ssd\"\n    }\n\n  # openstack_blockstorage_volume_v3.node_volume[1] will be created\n  + resource \"openstack_blockstorage_volume_v3\" \"node_volume\" {\n      + attachment           = (known after apply)\n      + availability_zone    = \"nova\"\n      + id                   = (known after apply)\n      + metadata             = (known after apply)\n      + name                 = \"testbed-volume-1-node-4\"\n      + region               = (known after apply)\n      + size                 = 20\n      + volume_retype_policy = \"never\"\n      + volume_type          = \"ssd\"\n    }\n\n  # openstack_blockstorage_volume_v3.node_volume[2] will be created\n  + resource \"openstack_blockstorage_volume_v3\" \"node_volume\" {\n      + attachment           = (known after apply)\n      + availability_zone    = \"nova\"\n      + id                   = (known after apply)\n      + metadata             = (known after apply)\n      + name                 = \"testbed-volume-2-node-5\"\n      + region               = (known after apply)\n      + size                 = 20\n      + volume_retype_policy = \"never\"\n      + volume_type          = \"ssd\"\n    }\n\n  # openstack_blockstorage_volume_v3.node_volume[3] will be created\n  + resource \"openstack_blockstorage_volume_v3\" \"node_volume\" {\n      + attachment           = (known after apply)\n      + availability_zone    = \"nova\"\n      + id                   = (known after apply)\n      + metadata             = (known after apply)\n      + name                 = \"testbed-volume-3-node-3\"\n      + region               = (known after apply)\n      + size                 = 20\n      + volume_retype_policy = \"never\"\n      + volume_type          = \"ssd\"\n    }\n\n  # openstack_blockstorage_volume_v3.node_volume[4] will be created\n  + resource \"openstack_blockstorage_volume_v3\" \"node_volume\" {\n      + attachment           = (known after apply)\n      + availability_zone    = \"nova\"\n      + id                   = (known after apply)\n      + metadata             = (known after apply)\n      + name                 = \"testbed-volume-4-node-4\"\n      + region               = (known after apply)\n      + size                 = 20\n      + volume_retype_policy = \"never\"\n      + volume_type          = \"ssd\"\n    }\n\n  # openstack_blockstorage_volume_v3.node_volume[5] will be created\n  + resource \"openstack_blockstorage_volume_v3\" \"node_volume\" {\n      + attachment           = (known after apply)\n      + availability_zone    = \"nova\"\n      + id                   = (known after apply)\n      + metadata             = (known after apply)\n      + name                 = \"testbed-volume-5-node-5\"\n      + region               = (known after apply)\n      + size                 = 20\n      + volume_retype_policy = \"never\"\n      + volume_type          = \"ssd\"\n    }\n\n  # openstack_blockstorage_volume_v3.node_volume[6] will be created\n  + resource \"openstack_blockstorage_volume_v3\" \"node_volume\" {\n      + attachment           = (known after apply)\n      + availability_zone    = \"nova\"\n      + id                   = (known after apply)\n      + metadata             = (known after apply)\n      + name                 = \"testbed-volume-6-node-3\"\n      + region               = (known after apply)\n      + size                 = 20\n      + volume_retype_policy = \"never\"\n      + volume_type          = \"ssd\"\n    }\n\n  # openstack_blockstorage_volume_v3.node_volume[7] will be created\n  + resource \"openstack_blockstorage_volume_v3\" \"node_volume\" {\n      + attachment           = (known after apply)\n      + availability_zone    = \"nova\"\n      + id                   = (known after apply)\n      + metadata             = (known after apply)\n      + name                 = \"testbed-volume-7-node-4\"\n      + region               = (known after apply)\n      + size                 = 20\n      + volume_retype_policy = \"never\"\n      + volume_type          = \"ssd\"\n    }\n\n  # openstack_blockstorage_volume_v3.node_volume[8] will be created\n  + resource \"openstack_blockstorage_volume_v3\" \"node_volume\" {\n      + attachment           = (known after apply)\n      + availability_zone    = \"nova\"\n      + id                   = (known after apply)\n      + metadata             = (known after apply)\n      + name                 = \"testbed-volume-8-node-5\"\n      + region               = (known after apply)\n      + size                 = 20\n      + volume_retype_policy = \"never\"\n      + volume_type          = \"ssd\"\n    }\n\n  # openstack_compute_instance_v2.manager_server will be created\n  + resource \"openstack_compute_instance_v2\" \"manager_server\" {\n      + access_ip_v4        = (known after apply)\n      + access_ip_v6        = (known after apply)\n      + all_metadata        = (known after apply)\n      + all_tags            = (known after apply)\n      + availability_zone   = \"nova\"\n      + config_drive        = true\n      + created             = (known after apply)\n      + flavor_id           = (known after apply)\n      + flavor_name         = \"OSISM-4V-16\"\n      + force_delete        = false\n      + hypervisor_hostname = (known after apply)\n      + id                  = (known after apply)\n      + image_id            = (known after apply)\n      + image_name          = (known after apply)\n      + key_pair            = \"testbed\"\n      + name                = \"testbed-manager\"\n      + power_state         = \"active\"\n      + region              = (known after apply)\n      + security_groups     = (known after apply)\n      + stop_before_destroy = false\n      + updated             = (known after apply)\n      + user_data           = (sensitive value)\n\n      + block_device {\n          + boot_index            = 0\n          + delete_on_termination = false\n          + destination_type      = \"volume\"\n          + multiattach           = false\n          + source_type           = \"volume\"\n          + uuid                  = (known after apply)\n        }\n\n      + network {\n          + access_network = false\n          + fixed_ip_v4    = (known after apply)\n          + fixed_ip_v6    = (known after apply)\n          + mac            = (known after apply)\n          + name           = (known after apply)\n          + port           = (known after apply)\n          + uuid           = (known after apply)\n        }\n    }\n\n  # openstack_compute_instance_v2.node_server[0] will be created\n  + resource \"openstack_compute_instance_v2\" \"node_server\" {\n      + access_ip_v4        = (known after apply)\n      + access_ip_v6        = (known after apply)\n      + all_metadata        = (known after apply)\n      + all_tags            = (known after apply)\n      + availability_zone   = \"nova\"\n      + config_drive        = true\n      + created             = (known after apply)\n      + flavor_id           = (known after apply)\n      + flavor_name         = \"OSISM-8V-32\"\n      + force_delete        = false\n      + hypervisor_hostname = (known after apply)\n      + id                  = (known after apply)\n      + image_id            = (known after apply)\n      + image_name          = (known after apply)\n      + key_pair            = \"testbed\"\n      + name                = \"testbed-node-0\"\n      + power_state         = \"active\"\n      + region              = (known after apply)\n      + security_groups     = (known after apply)\n      + stop_before_destroy = false\n      + updated             = (known after apply)\n      + user_data           = \"ae09e46b224a6ca206a9ed4f8f8a4f8520827854\"\n\n      + block_device {\n          + boot_index            = 0\n          + delete_on_termination = false\n          + destination_type      = \"volume\"\n          + multiattach           = false\n          + source_type           = \"volume\"\n          + uuid                  = (known after apply)\n        }\n\n      + network {\n          + access_network = false\n          + fixed_ip_v4    = (known after apply)\n          + fixed_ip_v6    = (known after apply)\n          + mac            = (known after apply)\n          + name           = (known after apply)\n          + port           = (known after apply)\n          + uuid           = (known after apply)\n        }\n    }\n\n  # openstack_compute_instance_v2.node_server[1] will be created\n  + resource \"openstack_compute_instance_v2\" \"node_server\" {\n      + access_ip_v4        = (known after apply)\n      + access_ip_v6        = (known after apply)\n      + all_metadata        = (known after apply)\n      + all_tags            = (known after apply)\n      + availability_zone   = \"nova\"\n      + config_drive        = true\n      + created             = (known after apply)\n      + flavor_id           = (known after apply)\n      + flavor_name         = \"OSISM-8V-32\"\n      + force_delete        = false\n      + hypervisor_hostname = (known after apply)\n      + id                  = (known after apply)\n      + image_id            = (known after apply)\n      + image_name          = (known after apply)\n      + key_pair            = \"testbed\"\n      + name                = \"testbed-node-1\"\n      + power_state         = \"active\"\n      + region              = (known after apply)\n      + security_groups     = (known after apply)\n      + stop_before_destroy = false\n      + updated             = (known after apply)\n      + user_data           = \"ae09e46b224a6ca206a9ed4f8f8a4f8520827854\"\n\n      + block_device {\n          + boot_index            = 0\n          + delete_on_termination = false\n          + destination_type      = \"volume\"\n          + multiattach           = false\n          + source_type           = \"volume\"\n          + uuid                  = (known after apply)\n        }\n\n      + network {\n          + access_network = false\n          + fixed_ip_v4    = (known after apply)\n          + fixed_ip_v6    = (known after apply)\n          + mac            = (known after apply)\n          + name           = (known after apply)\n          + port           = (known after apply)\n          + uuid           = (known after apply)\n        }\n    }\n\n  # openstack_compute_instance_v2.node_server[2] will be created\n  + resource \"openstack_compute_instance_v2\" \"node_server\" {\n      + access_ip_v4        = (known after apply)\n      + access_ip_v6        = (known after apply)\n      + all_metadata        = (known after apply)\n      + all_tags            = (known after apply)\n      + availability_zone   = \"nova\"\n      + config_drive        = true\n      + created             = (known after apply)\n      + flavor_id           = (known after apply)\n      + flavor_name         = \"OSISM-8V-32\"\n      + force_delete        = false\n      + hypervisor_hostname = (known after apply)\n      + id                  = (known after apply)\n      + image_id            = (known after apply)\n      + image_name          = (known after apply)\n      + key_pair            = \"testbed\"\n      + name                = \"testbed-node-2\"\n      + power_state         = \"active\"\n      + region              = (known after apply)\n      + security_groups     = (known after apply)\n      + stop_before_destroy = false\n      + updated             = (known after apply)\n      + user_data           = \"ae09e46b224a6ca206a9ed4f8f8a4f8520827854\"\n\n      + block_device {\n          + boot_index            = 0\n          + delete_on_termination = false\n          + destination_type      = \"volume\"\n          + multiattach           = false\n          + source_type           = \"volume\"\n          + uuid                  = (known after apply)\n        }\n\n      + network {\n          + access_network = false\n          + fixed_ip_v4    = (known after apply)\n          + fixed_ip_v6    = (known after apply)\n          + mac            = (known after apply)\n          + name           = (known after apply)\n          + port           = (known after apply)\n          + uuid           = (known after apply)\n        }\n    }\n\n  # openstack_compute_instance_v2.node_server[3] will be created\n  + resource \"openstack_compute_instance_v2\" \"node_server\" {\n      + access_ip_v4        = (known after apply)\n      + access_ip_v6        = (known after apply)\n      + all_metadata        = (known after apply)\n      + all_tags            = (known after apply)\n      + availability_zone   = \"nova\"\n      + config_drive        = true\n      + created             = (known after apply)\n      + flavor_id           = (known after apply)\n      + flavor_name         = \"OSISM-8V-32\"\n      + force_delete        = false\n      + hypervisor_hostname = (known after apply)\n      + id                  = (known after apply)\n      + image_id            = (known after apply)\n      + image_name          = (known after apply)\n      + key_pair            = \"testbed\"\n      + name                = \"testbed-node-3\"\n      + power_state         = \"active\"\n      + region              = (known after apply)\n      + security_groups     = (known after apply)\n      + stop_before_destroy = false\n      + updated             = (known after apply)\n      + user_data           = \"ae09e46b224a6ca206a9ed4f8f8a4f8520827854\"\n\n      + block_device {\n          + boot_index            = 0\n          + delete_on_termination = false\n          + destination_type      = \"volume\"\n          + multiattach           = false\n          + source_type           = \"volume\"\n          + uuid                  = (known after apply)\n        }\n\n      + network {\n          + access_network = false\n          + fixed_ip_v4    = (known after apply)\n          + fixed_ip_v6    = (known after apply)\n          + mac            = (known after apply)\n          + name           = (known after apply)\n          + port           = (known after apply)\n          + uuid           = (known after apply)\n        }\n    }\n\n  # openstack_compute_instance_v2.node_server[4] will be created\n  + resource \"openstack_compute_instance_v2\" \"node_server\" {\n      + access_ip_v4        = (known after apply)\n      + access_ip_v6        = (known after apply)\n      + all_metadata        = (known after apply)\n      + all_tags            = (known after apply)\n      + availability_zone   = \"nova\"\n      + config_drive        = true\n      + created             = (known after apply)\n      + flavor_id           = (known after apply)\n      + flavor_name         = \"OSISM-8V-32\"\n      + force_delete        = false\n      + hypervisor_hostname = (known after apply)\n      + id                  = (known after apply)\n      + image_id            = (known after apply)\n      + image_name          = (known after apply)\n      + key_pair            = \"testbed\"\n      + name                = \"testbed-node-4\"\n      + power_state         = \"active\"\n      + region              = (known after apply)\n      + security_groups     = (known after apply)\n      + stop_before_destroy = false\n      + updated             = (known after apply)\n      + user_data           = \"ae09e46b224a6ca206a9ed4f8f8a4f8520827854\"\n\n      + block_device {\n          + boot_index            = 0\n          + delete_on_termination = false\n          + destination_type      = \"volume\"\n          + multiattach           = false\n          + source_type           = \"volume\"\n          + uuid                  = (known after apply)\n        }\n\n      + network {\n          + access_network = false\n          + fixed_ip_v4    = (known after apply)\n          + fixed_ip_v6    = (known after apply)\n          + mac            = (known after apply)\n          + name           = (known after apply)\n          + port           = (known after apply)\n          + uuid           = (known after apply)\n        }\n    }\n\n  # openstack_compute_instance_v2.node_server[5] will be created\n  + resource \"openstack_compute_instance_v2\" \"node_server\" {\n      + access_ip_v4        = (known after apply)\n      + access_ip_v6        = (known after apply)\n      + all_metadata        = (known after apply)\n      + all_tags            = (known after apply)\n      + availability_zone   = \"nova\"\n      + config_drive        = true\n      + created             = (known after apply)\n      + flavor_id           = (known after apply)\n      + flavor_name         = \"OSISM-8V-32\"\n      + force_delete        = false\n      + hypervisor_hostname = (known after apply)\n      + id                  = (known after apply)\n      + image_id            = (known after apply)\n      + image_name          = (known after apply)\n      + key_pair            = \"testbed\"\n      + name                = \"testbed-node-5\"\n      + power_state         = \"active\"\n      + region              = (known after apply)\n      + security_groups     = (known after apply)\n      + stop_before_destroy = false\n      + updated             = (known after apply)\n      + user_data           = \"ae09e46b224a6ca206a9ed4f8f8a4f8520827854\"\n\n      + block_device {\n          + boot_index            = 0\n          + delete_on_termination = false\n          + destination_type      = \"volume\"\n          + multiattach           = false\n          + source_type           = \"volume\"\n          + uuid                  = (known after apply)\n        }\n\n      + network {\n          + access_network = false\n          + fixed_ip_v4    = (known after apply)\n          + fixed_ip_v6    = (known after apply)\n          + mac            = (known after apply)\n          + name           = (known after apply)\n          + port           = (known after apply)\n          + uuid           = (known after apply)\n        }\n    }\n\n  # openstack_compute_keypair_v2.key will be created\n  + resource \"openstack_compute_keypair_v2\" \"key\" {\n      + fingerprint = (known after apply)\n      + id          = (known after apply)\n      + name        = \"testbed\"\n      + private_key = (sensitive value)\n      + public_key  = (known after apply)\n      + region      = (known after apply)\n      + user_id     = (known after apply)\n    }\n\n  # openstack_compute_volume_attach_v2.node_volume_attachment[0] will be created\n  + resource \"openstack_compute_volume_attach_v2\" \"node_volume_attachment\" {\n      + device      = (known after apply)\n      + id          = (known after apply)\n      + instance_id = (known after apply)\n      + region      = (known after apply)\n      + volume_id   = (known after apply)\n    }\n\n  # openstack_compute_volume_attach_v2.node_volume_attachment[1] will be created\n  + resource \"openstack_compute_volume_attach_v2\" \"node_volume_attachment\" {\n      + device      = (known after apply)\n      + id          = (known after apply)\n      + instance_id = (known after apply)\n      + region      = (known after apply)\n      + volume_id   = (known after apply)\n    }\n\n  # openstack_compute_volume_attach_v2.node_volume_attachment[2] will be created\n  + resource \"openstack_compute_volume_attach_v2\" \"node_volume_attachment\" {\n      + device      = (known after apply)\n      + id          = (known after apply)\n      + instance_id = (known after apply)\n      + region      = (known after apply)\n      + volume_id   = (known after apply)\n    }\n\n  # openstack_compute_volume_attach_v2.node_volume_attachment[3] will be created\n  + resource \"openstack_compute_volume_attach_v2\" \"node_volume_attachment\" {\n      + device      = (known after apply)\n      + id          = (known after apply)\n      + instance_id = (known after apply)\n      + region      = (known after apply)\n      + volume_id   = (known after apply)\n    }\n\n  # openstack_compute_volume_attach_v2.node_volume_attachment[4] will be created\n  + resource \"openstack_compute_volume_attach_v2\" \"node_volume_attachment\" {\n      + device      = (known after apply)\n      + id          = (known after apply)\n      + instance_id = (known after apply)\n      + region      = (known after apply)\n      + volume_id   = (known after apply)\n    }\n\n  # openstack_compute_volume_attach_v2.node_volume_attachment[5] will be created\n  + resource \"openstack_compute_volume_attach_v2\" \"node_volume_attachment\" {\n      + device      = (known after apply)\n      + id          = (known after apply)\n      + instance_id = (known after apply)\n      + region      = (known after apply)\n      + volume_id   = (known after apply)\n    }\n\n  # openstack_compute_volume_attach_v2.node_volume_attachment[6] will be created\n  + resource \"openstack_compute_volume_attach_v2\" \"node_volume_attachment\" {\n      + device      = (known after apply)\n      + id          = (known after apply)\n      + instance_id = (known after apply)\n      + region      = (known after apply)\n      + volume_id   = (known after apply)\n    }\n\n  # openstack_compute_volume_attach_v2.node_volume_attachment[7] will be created\n  + resource \"openstack_compute_volume_attach_v2\" \"node_volume_attachment\" {\n      + device      = (known after apply)\n      + id          = (known after apply)\n      + instance_id = (known after apply)\n      + region      = (known after apply)\n      + volume_id   = (known after apply)\n    }\n\n  # openstack_compute_volume_attach_v2.node_volume_attachment[8] will be created\n  + resource \"openstack_compute_volume_attach_v2\" \"node_volume_attachment\" {\n      + device      = (known after apply)\n      + id          = (known after apply)\n      + instance_id = (known after apply)\n      + region      = (known after apply)\n      + volume_id   = (known after apply)\n    }\n\n  # openstack_networking_floatingip_associate_v2.manager_floating_ip_association will be created\n  + resource \"openstack_networking_floatingip_associate_v2\" \"manager_floating_ip_association\" {\n      + fixed_ip    = (known after apply)\n      + floating_ip = (known after apply)\n      + id          = (known after apply)\n      + port_id     = (known after apply)\n      + region      = (known after apply)\n    }\n\n  # openstack_networking_floatingip_v2.manager_floating_ip will be created\n  + resource \"openstack_networking_floatingip_v2\" \"manager_floating_ip\" {\n      + address    = (known after apply)\n      + all_tags   = (known after apply)\n      + dns_domain = (known after apply)\n      + dns_name   = (known after apply)\n      + fixed_ip   = (known after apply)\n      + id         = (known after apply)\n      + pool       = \"public\"\n      + port_id    = (known after apply)\n      + region     = (known after apply)\n      + subnet_id  = (known after apply)\n      + tenant_id  = (known after apply)\n    }\n\n  # openstack_networking_network_v2.net_management will be created\n  + resource \"openstack_networking_network_v2\" \"net_management\" {\n      + admin_state_up          = (known after apply)\n      + all_tags                = (known after apply)\n      + availability_zone_hints = [\n          + \"nova\",\n        ]\n      + dns_domain              = (known after apply)\n      + external                = (known after apply)\n      + id                      = (known after apply)\n      + mtu                     = (known after apply)\n      + name                    = \"net-testbed-management\"\n      + port_security_enabled   = (known after apply)\n      + qos_policy_id           = (known after apply)\n      + region                  = (known after apply)\n      + shared                  = (known after apply)\n      + tenant_id               = (known after apply)\n      + transparent_vlan        = (known after apply)\n\n      + segments (known after apply)\n    }\n\n  # openstack_networking_port_v2.manager_port_management will be created\n  + resource \"openstack_networking_port_v2\" \"manager_port_management\" {\n      + admin_state_up         = (known after apply)\n      + all_fixed_ips          = (known after apply)\n      + all_security_group_ids = (known after apply)\n      + all_tags               = (known after apply)\n      + device_id              = (known after apply)\n      + device_owner           = (known after apply)\n      + dns_assignment         = (known after apply)\n      + dns_name               = (known after apply)\n      + id                     = (known after apply)\n      + mac_address            = (known after apply)\n      + network_id             = (known after apply)\n      + port_security_enabled  = (known after apply)\n      + qos_policy_id          = (known after apply)\n      + region                 = (known after apply)\n      + security_group_ids     = (known after apply)\n      + tenant_id              = (known after apply)\n\n      + allowed_address_pairs {\n          + ip_address = \"192.168.16.8/32\"\n        }\n\n      + binding (known after apply)\n\n      + fixed_ip {\n          + ip_address = \"192.168.16.5\"\n          + subnet_id  = (known after apply)\n        }\n    }\n\n  # openstack_networking_port_v2.node_port_management[0] will be created\n  + resource \"openstack_networking_port_v2\" \"node_port_management\" {\n      + admin_state_up         = (known after apply)\n      + all_fixed_ips          = (known after apply)\n      + all_security_group_ids = (known after apply)\n      + all_tags               = (known after apply)\n      + device_id              = (known after apply)\n      + device_owner           = (known after apply)\n      + dns_assignment         = (known after apply)\n      + dns_name               = (known after apply)\n      + id                     = (known after apply)\n      + mac_address            = (known after apply)\n      + network_id             = (known after apply)\n      + port_security_enabled  = (known after apply)\n      + qos_policy_id          = (known after apply)\n      + region                 = (known after apply)\n      + security_group_ids     = (known after apply)\n      + tenant_id              = (known after apply)\n\n      + allowed_address_pairs {\n          + ip_address = \"192.168.16.254/32\"\n        }\n      + allowed_address_pairs {\n          + ip_address = \"192.168.16.8/32\"\n        }\n      + allowed_address_pairs {\n          + ip_address = \"192.168.16.9/32\"\n        }\n\n      + binding (known after apply)\n\n      + fixed_ip {\n          + ip_address = \"192.168.16.10\"\n          + subnet_id  = (known after apply)\n        }\n    }\n\n  # openstack_networking_port_v2.node_port_management[1] will be created\n  + resource \"openstack_networking_port_v2\" \"node_port_management\" {\n      + admin_state_up         = (known after apply)\n      + all_fixed_ips          = (known after apply)\n      + all_security_group_ids = (known after apply)\n      + all_tags               = (known after apply)\n      + device_id              = (known after apply)\n      + device_owner           = (known after apply)\n      + dns_assignment         = (known after apply)\n      + dns_name               = (known after apply)\n      + id                     = (known after apply)\n      + mac_address            = (known after apply)\n      + network_id             = (known after apply)\n      + port_security_enabled  = (known after apply)\n      + qos_policy_id          = (known after apply)\n      + region                 = (known after apply)\n      + security_group_ids     = (known after apply)\n      + tenant_id              = (known after apply)\n\n      + allowed_address_pairs {\n          + ip_address = \"192.168.16.254/32\"\n        }\n      + allowed_address_pairs {\n          + ip_address = \"192.168.16.8/32\"\n        }\n      + allowed_address_pairs {\n          + ip_address = \"192.168.16.9/32\"\n        }\n\n      + binding (known after apply)\n\n      + fixed_ip {\n          + ip_address = \"192.168.16.11\"\n          + subnet_id  = (known after apply)\n        }\n    }\n\n  # openstack_networking_port_v2.node_port_management[2] will be created\n  + resource \"openstack_networking_port_v2\" \"node_port_management\" {\n      + admin_state_up         = (known after apply)\n      + all_fixed_ips          = (known after apply)\n      + all_security_group_ids = (known after apply)\n      + all_tags               = (known after apply)\n      + device_id              = (known after apply)\n      + device_owner           = (known after apply)\n      + dns_assignment         = (known after apply)\n      + dns_name               = (known after apply)\n      + id                     = (known after apply)\n      + mac_address            = (known after apply)\n      + network_id             = (known after apply)\n      + port_security_enabled  = (known after apply)\n      + qos_policy_id          = (known after apply)\n      + region                 = (known after apply)\n      + security_group_ids     = (known after apply)\n      + tenant_id              = (known after apply)\n\n      + allowed_address_pairs {\n          + ip_address = \"192.168.16.254/32\"\n        }\n      + allowed_address_pairs {\n          + ip_address = \"192.168.16.8/32\"\n        }\n      + allowed_address_pairs {\n          + ip_address = \"192.168.16.9/32\"\n        }\n\n      + binding (known after apply)\n\n      + fixed_ip {\n          + ip_address = \"192.168.16.12\"\n          + subnet_id  = (known after apply)\n        }\n    }\n\n  # openstack_networking_port_v2.node_port_management[3] will be created\n  + resource \"openstack_networking_port_v2\" \"node_port_management\" {\n      + admin_state_up         = (known after apply)\n      + all_fixed_ips          = (known after apply)\n      + all_security_group_ids = (known after apply)\n      + all_tags               = (known after apply)\n      + device_id              = (known after apply)\n      + device_owner           = (known after apply)\n      + dns_assignment         = (known after apply)\n      + dns_name               = (known after apply)\n      + id                     = (known after apply)\n      + mac_address            = (known after apply)\n      + network_id             = (known after apply)\n      + port_security_enabled  = (known after apply)\n      + qos_policy_id          = (known after apply)\n      + region                 = (known after apply)\n      + security_group_ids     = (known after apply)\n      + tenant_id              = (known after apply)\n\n      + allowed_address_pairs {\n          + ip_address = \"192.168.16.254/32\"\n        }\n      + allowed_address_pairs {\n          + ip_address = \"192.168.16.8/32\"\n        }\n      + allowed_address_pairs {\n          + ip_address = \"192.168.16.9/32\"\n        }\n\n      + binding (known after apply)\n\n      + fixed_ip {\n          + ip_address = \"192.168.16.13\"\n          + subnet_id  = (known after apply)\n        }\n    }\n\n  # openstack_networking_port_v2.node_port_management[4] will be created\n  + resource \"openstack_networking_port_v2\" \"node_port_management\" {\n      + admin_state_up         = (known after apply)\n      + all_fixed_ips          = (known after apply)\n      + all_security_group_ids = (known after apply)\n      + all_tags               = (known after apply)\n      + device_id              = (known after apply)\n      + device_owner           = (known after apply)\n      + dns_assignment         = (known after apply)\n      + dns_name               = (known after apply)\n      + id                     = (known after apply)\n      + mac_address            = (known after apply)\n      + network_id             = (known after apply)\n      + port_security_enabled  = (known after apply)\n      + qos_policy_id          = (known after apply)\n      + region                 = (known after apply)\n      + security_group_ids     = (known after apply)\n      + tenant_id              = (known after apply)\n\n      + allowed_address_pairs {\n          + ip_address = \"192.168.16.254/32\"\n        }\n      + allowed_address_pairs {\n          + ip_address = \"192.168.16.8/32\"\n        }\n      + allowed_address_pairs {\n          + ip_address = \"192.168.16.9/32\"\n        }\n\n      + binding (known after apply)\n\n      + fixed_ip {\n          + ip_address = \"192.168.16.14\"\n          + subnet_id  = (known after apply)\n        }\n    }\n\n  # openstack_networking_port_v2.node_port_management[5] will be created\n  + resource \"openstack_networking_port_v2\" \"node_port_management\" {\n      + admin_state_up         = (known after apply)\n      + all_fixed_ips          = (known after apply)\n      + all_security_group_ids = (known after apply)\n      + all_tags               = (known after apply)\n      + device_id              = (known after apply)\n      + device_owner           = (known after apply)\n      + dns_assignment         = (known after apply)\n      + dns_name               = (known after apply)\n      + id                     = (known after apply)\n      + mac_address            = (known after apply)\n      + network_id             = (known after apply)\n      + port_security_enabled  = (known after apply)\n      + qos_policy_id          = (known after apply)\n      + region                 = (known after apply)\n      + security_group_ids     = (known after apply)\n      + tenant_id              = (known after apply)\n\n      + allowed_address_pairs {\n          + ip_address = \"192.168.16.254/32\"\n        }\n      + allowed_address_pairs {\n          + ip_address = \"192.168.16.8/32\"\n        }\n      + allowed_address_pairs {\n          + ip_address = \"192.168.16.9/32\"\n        }\n\n      + binding (known after apply)\n\n      + fixed_ip {\n          + ip_address = \"192.168.16.15\"\n          + subnet_id  = (known after apply)\n        }\n    }\n\n  # openstack_networking_router_interface_v2.router_interface will be created\n  + resource \"openstack_networking_router_interface_v2\" \"router_interface\" {\n      + force_destroy = false\n      + id            = (known after apply)\n      + port_id       = (known after apply)\n      + region        = (known after apply)\n      + router_id     = (known after apply)\n      + subnet_id     = (known after apply)\n    }\n\n  # openstack_networking_router_v2.router will be created\n  + resource \"openstack_networking_router_v2\" \"router\" {\n      + admin_state_up          = (known after apply)\n      + all_tags                = (known after apply)\n      + availability_zone_hints = [\n          + \"nova\",\n        ]\n      + distributed             = (known after apply)\n      + enable_snat             = (known after apply)\n      + external_network_id     = \"e6be7364-bfd8-4de7-8120-8f41c69a139a\"\n      + external_qos_policy_id  = (known after apply)\n      + id                      = (known after apply)\n      + name                    = \"testbed\"\n      + region                  = (known after apply)\n      + tenant_id               = (known after apply)\n\n      + external_fixed_ip (known after apply)\n    }\n\n  # openstack_networking_secgroup_rule_v2.security_group_management_rule1 will be created\n  + resource \"openstack_networking_secgroup_rule_v2\" \"security_group_management_rule1\" {\n      + description             = \"ssh\"\n      + direction               = \"ingress\"\n      + ethertype               = \"IPv4\"\n      + id                      = (known after apply)\n      + port_range_max          = 22\n      + port_range_min          = 22\n      + protocol                = \"tcp\"\n      + region                  = (known after apply)\n      + remote_address_group_id = (known after apply)\n      + remote_group_id         = (known after apply)\n      + remote_ip_prefix        = \"0.0.0.0/0\"\n      + security_group_id       = (known after apply)\n      + tenant_id               = (known after apply)\n    }\n\n  # openstack_networking_secgroup_rule_v2.security_group_management_rule2 will be created\n  + resource \"openstack_networking_secgroup_rule_v2\" \"security_group_management_rule2\" {\n      + description             = \"wireguard\"\n      + direction               = \"ingress\"\n      + ethertype               = \"IPv4\"\n      + id                      = (known after apply)\n      + port_range_max          = 51820\n      + port_range_min          = 51820\n      + protocol                = \"udp\"\n      + region                  = (known after apply)\n      + remote_address_group_id = (known after apply)\n      + remote_group_id         = (known after apply)\n      + remote_ip_prefix        = \"0.0.0.0/0\"\n      + security_group_id       = (known after apply)\n      + tenant_id               = (known after apply)\n    }\n\n  # openstack_networking_secgroup_rule_v2.security_group_management_rule3 will be created\n  + resource \"openstack_networking_secgroup_rule_v2\" \"security_group_management_rule3\" {\n      + direction               = \"ingress\"\n      + ethertype               = \"IPv4\"\n      + id                      = (known after apply)\n      + protocol                = \"tcp\"\n      + region                  = (known after apply)\n      + remote_address_group_id = (known after apply)\n      + remote_group_id         = (known after apply)\n      + remote_ip_prefix        = \"192.168.16.0/20\"\n      + security_group_id       = (known after apply)\n      + tenant_id               = (known after apply)\n    }\n\n  # openstack_networking_secgroup_rule_v2.security_group_management_rule4 will be created\n  + resource \"openstack_networking_secgroup_rule_v2\" \"security_group_management_rule4\" {\n      + direction               = \"ingress\"\n      + ethertype               = \"IPv4\"\n      + id                      = (known after apply)\n      + protocol                = \"udp\"\n      + region                  = (known after apply)\n      + remote_address_group_id = (known after apply)\n      + remote_group_id         = (known after apply)\n      + remote_ip_prefix        = \"192.168.16.0/20\"\n      + security_group_id       = (known after apply)\n      + tenant_id               = (known after apply)\n    }\n\n  # openstack_networking_secgroup_rule_v2.security_group_management_rule5 will be created\n  + resource \"openstack_networking_secgroup_rule_v2\" \"security_group_management_rule5\" {\n      + direction               = \"ingress\"\n      + ethertype               = \"IPv4\"\n      + id                      = (known after apply)\n      + protocol                = \"icmp\"\n      + region                  = (known after apply)\n      + remote_address_group_id = (known after apply)\n      + remote_group_id         = (known after apply)\n      + remote_ip_prefix        = \"0.0.0.0/0\"\n      + security_group_id       = (known after apply)\n      + tenant_id               = (known after apply)\n    }\n\n  # openstack_networking_secgroup_rule_v2.security_group_node_rule1 will be created\n  + resource \"openstack_networking_secgroup_rule_v2\" \"security_group_node_rule1\" {\n      + direction               = \"ingress\"\n      + ethertype               = \"IPv4\"\n      + id                      = (known after apply)\n      + protocol                = \"tcp\"\n      + region                  = (known after apply)\n      + remote_address_group_id = (known after apply)\n      + remote_group_id         = (known after apply)\n      + remote_ip_prefix        = \"0.0.0.0/0\"\n      + security_group_id       = (known after apply)\n      + tenant_id               = (known after apply)\n    }\n\n  # openstack_networking_secgroup_rule_v2.security_group_node_rule2 will be created\n  + resource \"openstack_networking_secgroup_rule_v2\" \"security_group_node_rule2\" {\n      + direction               = \"ingress\"\n      + ethertype               = \"IPv4\"\n      + id                      = (known after apply)\n      + protocol                = \"udp\"\n      + region                  = (known after apply)\n      + remote_address_group_id = (known after apply)\n      + remote_group_id         = (known after apply)\n      + remote_ip_prefix        = \"0.0.0.0/0\"\n      + security_group_id       = (known after apply)\n      + tenant_id               = (known after apply)\n    }\n\n  # openstack_networking_secgroup_rule_v2.security_group_node_rule3 will be created\n  + resource \"openstack_networking_secgroup_rule_v2\" \"security_group_node_rule3\" {\n      + direction               = \"ingress\"\n      + ethertype               = \"IPv4\"\n      + id                      = (known after apply)\n      + protocol                = \"icmp\"\n      + region                  = (known after apply)\n      + remote_address_group_id = (known after apply)\n      + remote_group_id         = (known after apply)\n      + remote_ip_prefix        = \"0.0.0.0/0\"\n      + security_group_id       = (known after apply)\n      + tenant_id               = (known after apply)\n    }\n\n  # openstack_networking_secgroup_rule_v2.security_group_rule_vrrp will be created\n  + resource \"openstack_networking_secgroup_rule_v2\" \"security_group_rule_vrrp\" {\n      + description             = \"vrrp\"\n      + direction               = \"ingress\"\n      + ethertype               = \"IPv4\"\n      + id                      = (known after apply)\n      + protocol                = \"112\"\n      + region                  = (known after apply)\n      + remote_address_group_id = (known after apply)\n      + remote_group_id         = (known after apply)\n      + remote_ip_prefix        = \"0.0.0.0/0\"\n      + security_group_id       = (known after apply)\n      + tenant_id               = (known after apply)\n    }\n\n  # openstack_networking_secgroup_v2.security_group_management will be created\n  + resource \"openstack_networking_secgroup_v2\" \"security_group_management\" {\n      + all_tags    = (known after apply)\n      + description = \"management security group\"\n      + id          = (known after apply)\n      + name        = \"testbed-management\"\n      + region      = (known after apply)\n      + stateful    = (known after apply)\n      + tenant_id   = (known after apply)\n    }\n\n  # openstack_networking_secgroup_v2.security_group_node will be created\n  + resource \"openstack_networking_secgroup_v2\" \"security_group_node\" {\n      + all_tags    = (known after apply)\n      + description = \"node security group\"\n      + id          = (known after apply)\n      + name        = \"testbed-node\"\n      + region      = (known after apply)\n      + stateful    = (known after apply)\n      + tenant_id   = (known after apply)\n    }\n\n  # openstack_networking_subnet_v2.subnet_management will be created\n  + resource \"openstack_networking_subnet_v2\" \"subnet_management\" {\n      + all_tags          = (known after apply)\n      + cidr              = \"192.168.16.0/20\"\n      + dns_nameservers   = [\n          + \"8.8.8.8\",\n          + \"9.9.9.9\",\n        ]\n      + enable_dhcp       = true\n      + gateway_ip        = (known after apply)\n      + id                = (known after apply)\n      + ip_version        = 4\n      + ipv6_address_mode = (known after apply)\n      + ipv6_ra_mode      = (known after apply)\n      + name              = \"subnet-testbed-management\"\n      + network_id        = (known after apply)\n      + no_gateway        = false\n      + region            = (known after apply)\n      + service_types     = (known after apply)\n      + tenant_id         = (known after apply)\n\n      + allocation_pool {\n          + end   = \"192.168.31.250\"\n          + start = \"192.168.31.200\"\n        }\n    }\n\n  # terraform_data.image will be created\n  + resource \"terraform_data\" \"image\" {\n      + id     = (known after apply)\n      + input  = \"Ubuntu 24.04\"\n      + output = (known after apply)\n    }\n\n  # terraform_data.image_node will be created\n  + resource \"terraform_data\" \"image_node\" {\n      + id     = (known after apply)\n      + input  = \"Ubuntu 24.04\"\n      + output = (known after apply)\n    }\n\nPlan: 64 to add, 0 to change, 0 to destroy.\n\nChanges to Outputs:\n  + manager_address = (sensitive value)\n  + private_key     = (sensitive value)\nterraform_data.image_node: Creating...\nterraform_data.image: Creating...\nterraform_data.image_node: Creation complete after 0s [id=15076263-b98f-8a1e-6728-793c2a492a4f]\nterraform_data.image: Creation complete after 0s [id=529e8ade-efc7-96ff-a4ad-79bfc4eed7eb]\ndata.openstack_images_image_v2.image: Reading...\ndata.openstack_images_image_v2.image_node: Reading...\nopenstack_blockstorage_volume_v3.node_volume[8]: Creating...\nopenstack_blockstorage_volume_v3.node_volume[3]: Creating...\nopenstack_blockstorage_volume_v3.node_volume[1]: Creating...\nopenstack_blockstorage_volume_v3.node_volume[4]: Creating...\nopenstack_blockstorage_volume_v3.node_volume[2]: Creating...\nopenstack_blockstorage_volume_v3.node_volume[0]: Creating...\nopenstack_compute_keypair_v2.key: Creating...\nopenstack_networking_network_v2.net_management: Creating...\ndata.openstack_images_image_v2.image: Read complete after 1s [id=846820b2-039e-4b42-adad-daf72e0f8ea4]\nopenstack_blockstorage_volume_v3.node_volume[6]: Creating...\nopenstack_compute_keypair_v2.key: Creation complete after 1s [id=testbed]\ndata.openstack_images_image_v2.image_node: Read complete after 1s [id=846820b2-039e-4b42-adad-daf72e0f8ea4]\nopenstack_blockstorage_volume_v3.node_volume[7]: Creating...\nopenstack_blockstorage_volume_v3.node_volume[5]: Creating...\nopenstack_networking_network_v2.net_management: Creation complete after 1s [id=efb372d3-0e12-4094-ab09-ea954e656478]\nopenstack_blockstorage_volume_v3.manager_base_volume[0]: Creating...\nopenstack_blockstorage_volume_v3.node_volume[2]: Creation complete after 4s [id=6f4643f0-329b-4ae2-97c7-d348648fd970]\nlocal_sensitive_file.id_rsa: Creating...\nlocal_sensitive_file.id_rsa: Creation complete after 0s [id=d4355e1f232bca0bfbf7b62060f5199acdefa892]\nopenstack_blockstorage_volume_v3.node_volume[0]: Creation complete after 4s [id=9c6b6763-9704-41ec-9b6d-97eb80a5145c]\nlocal_file.id_rsa_pub: Creating...\nopenstack_blockstorage_volume_v3.node_volume[8]: Creation complete after 4s [id=5f47789f-2a1a-442a-8a3c-644a4f92e5a8]\nopenstack_blockstorage_volume_v3.node_volume[4]: Creation complete after 4s [id=fb58bf16-a1c6-4abb-925b-29c6a6a6f100]\nopenstack_blockstorage_volume_v3.node_base_volume[1]: Creating...\nopenstack_networking_subnet_v2.subnet_management: Creating...\nopenstack_blockstorage_volume_v3.node_volume[1]: Creation complete after 4s [id=5660bb62-e2c9-45b2-ae11-8ab0f829b194]\nlocal_file.id_rsa_pub: Creation complete after 0s [id=d54a83a615742cbaaac15d6ca6f05a14c6d0c5ab]\nopenstack_blockstorage_volume_v3.node_base_volume[0]: Creating...\nopenstack_blockstorage_volume_v3.node_base_volume[4]: Creating...\nopenstack_blockstorage_volume_v3.node_base_volume[2]: Creating...\nopenstack_blockstorage_volume_v3.node_volume[3]: Creation complete after 4s [id=6644f191-69e9-4f3a-baff-cffe4152595b]\nopenstack_blockstorage_volume_v3.node_base_volume[3]: Creating...\nopenstack_blockstorage_volume_v3.node_volume[6]: Creation complete after 3s [id=4534650b-23a0-46c8-8bf1-44520caee637]\nopenstack_blockstorage_volume_v3.node_volume[7]: Creation complete after 3s [id=f659b61f-c39f-426f-a640-6a753951661d]\nopenstack_blockstorage_volume_v3.node_base_volume[5]: Creating...\nopenstack_blockstorage_volume_v3.node_volume[5]: Creation complete after 3s [id=1558935b-4c66-4614-97e5-cda14abd2641]\nopenstack_blockstorage_volume_v3.manager_base_volume[0]: Creation complete after 4s [id=65266dda-9f19-479d-a42c-1592584b970f]\nopenstack_networking_subnet_v2.subnet_management: Creation complete after 1s [id=ee6d269f-6556-408a-a662-603f9f99eb43]\nopenstack_networking_router_v2.router: Creating...\nopenstack_blockstorage_volume_v3.node_base_volume[1]: Creation complete after 3s [id=6fef90f9-4ce5-456f-a170-152290f19313]\nopenstack_blockstorage_volume_v3.node_base_volume[0]: Creation complete after 3s [id=2d259cae-7b79-479e-b978-3fffee75aa51]\nopenstack_blockstorage_volume_v3.node_base_volume[4]: Creation complete after 3s [id=47e600ec-05ca-44a6-93d0-5ca3104e3639]\nopenstack_blockstorage_volume_v3.node_base_volume[2]: Creation complete after 4s [id=35b8d51e-716e-455d-9a95-995756ae5900]\nopenstack_blockstorage_volume_v3.node_base_volume[3]: Creation complete after 4s [id=f90a4e80-1f36-4d14-b7da-de7b20e9fbca]\nopenstack_blockstorage_volume_v3.node_base_volume[5]: Creation complete after 4s [id=7a9487cb-ec9e-4508-9c5a-f692700fc5a4]\nopenstack_networking_router_v2.router: Creation complete after 3s [id=78bdffa7-7274-45bc-a222-3c8067a8b67a]\nopenstack_networking_secgroup_v2.security_group_management: Creating...\nopenstack_networking_secgroup_v2.security_group_node: Creating...\nopenstack_networking_router_interface_v2.router_interface: Creating...\nopenstack_networking_secgroup_v2.security_group_node: Creation complete after 0s [id=0e28c216-6c91-4be2-bd8a-e93830fffbeb]\nopenstack_networking_secgroup_rule_v2.security_group_node_rule3: Creating...\nopenstack_networking_secgroup_rule_v2.security_group_node_rule2: Creating...\nopenstack_networking_port_v2.node_port_management[5]: Creating...\nopenstack_networking_secgroup_rule_v2.security_group_node_rule1: Creating...\nopenstack_networking_secgroup_rule_v2.security_group_rule_vrrp: Creating...\nopenstack_networking_port_v2.node_port_management[3]: Creating...\nopenstack_networking_port_v2.node_port_management[1]: Creating...\nopenstack_networking_port_v2.node_port_management[0]: Creating...\nopenstack_networking_secgroup_rule_v2.security_group_node_rule3: Creation complete after 0s [id=387089e4-f760-4928-88f5-a915100e55d6]\nopenstack_networking_port_v2.node_port_management[2]: Creating...\nopenstack_networking_secgroup_rule_v2.security_group_node_rule2: Creation complete after 0s [id=b73046cc-2cee-4e90-9574-a7fc0413601b]\nopenstack_networking_port_v2.node_port_management[4]: Creating...\nopenstack_networking_secgroup_rule_v2.security_group_rule_vrrp: Creation complete after 0s [id=1d63a497-952f-4017-a681-67dbf229b9f0]\nopenstack_networking_port_v2.node_port_management[1]: Creation complete after 1s [id=4e0612ba-f5fc-449f-bdf0-52cf3f9e0ff4]\nopenstack_networking_port_v2.node_port_management[5]: Creation complete after 1s [id=b963b7df-f3b4-4efc-a2f3-fc1a7ac620e0]\nopenstack_networking_port_v2.node_port_management[0]: Creation complete after 1s [id=52a531cd-5463-4412-a687-49e3322e1d3a]\nopenstack_networking_secgroup_rule_v2.security_group_node_rule1: Creation complete after 1s [id=ab75e10b-5a7d-4465-a77b-a2d27ac85a87]\nopenstack_networking_secgroup_v2.security_group_management: Creation complete after 1s [id=4576d5df-b4e8-453a-9f85-e9033140afde]\nopenstack_networking_secgroup_rule_v2.security_group_management_rule5: Creating...\nopenstack_networking_secgroup_rule_v2.security_group_management_rule1: Creating...\nopenstack_networking_port_v2.manager_port_management: Creating...\nopenstack_networking_secgroup_rule_v2.security_group_management_rule3: Creating...\nopenstack_networking_secgroup_rule_v2.security_group_management_rule2: Creating...\nopenstack_networking_secgroup_rule_v2.security_group_management_rule4: Creating...\nopenstack_networking_port_v2.node_port_management[2]: Creation complete after 1s [id=3efcac55-5642-4046-a9a6-94f0970dacfb]\nopenstack_networking_secgroup_rule_v2.security_group_management_rule5: Creation complete after 0s [id=791e61d3-0ebb-4c2f-b01f-4a99d7a46541]\nopenstack_networking_port_v2.node_port_management[4]: Creation complete after 1s [id=fd0bab54-e5d7-4115-8ac6-6d5e24c663da]\nopenstack_networking_port_v2.node_port_management[3]: Creation complete after 1s [id=b86f1b60-f237-4c8b-a846-4b44402e937f]\nopenstack_networking_secgroup_rule_v2.security_group_management_rule1: Creation complete after 0s [id=f8964514-2650-45f3-9372-ee021595be3f]\nopenstack_networking_secgroup_rule_v2.security_group_management_rule3: Creation complete after 0s [id=4dd1cb5f-5d60-43d2-bcd1-f648f83f847e]\nopenstack_networking_secgroup_rule_v2.security_group_management_rule2: Creation complete after 0s [id=c8b4c78c-fd27-4b8c-98e4-38b549c80b55]\nopenstack_networking_secgroup_rule_v2.security_group_management_rule4: Creation complete after 1s [id=e6ef18d5-506a-44c4-be35-c43969468703]\nopenstack_networking_port_v2.manager_port_management: Creation complete after 1s [id=6419139a-cf37-4920-a5ce-6fbaf77a4267]\nopenstack_networking_router_interface_v2.router_interface: Creation complete after 4s [id=e6f7b8e7-2ad3-4d1c-a6b4-399b8439e436]\nopenstack_networking_floatingip_v2.manager_floating_ip: Creating...\nopenstack_compute_instance_v2.node_server[1]: Creating...\nopenstack_compute_instance_v2.node_server[0]: Creating...\nopenstack_compute_instance_v2.node_server[5]: Creating...\nopenstack_compute_instance_v2.node_server[4]: Creating...\nopenstack_compute_instance_v2.node_server[2]: Creating...\nopenstack_compute_instance_v2.node_server[3]: Creating...\nopenstack_networking_floatingip_v2.manager_floating_ip: Creation complete after 3s [id=4b75e324-3592-41b1-8c39-d8b21170c67a]\nlocal_file.MANAGER_ADDRESS: Creating...\nlocal_file.inventory: Creating...\nopenstack_networking_floatingip_associate_v2.manager_floating_ip_association: Creating...\nlocal_file.inventory: Creation complete after 0s [id=5cef27c57f3cca0ffaf63b133cf6c70a2ab823cb]\nlocal_file.MANAGER_ADDRESS: Creation complete after 0s [id=2a52d35ad08ef77d309e2f503983de36521d64f8]\nopenstack_networking_floatingip_associate_v2.manager_floating_ip_association: Creation complete after 1s [id=4b75e324-3592-41b1-8c39-d8b21170c67a]\nopenstack_compute_instance_v2.node_server[1]: Still creating... [10s elapsed]\nopenstack_compute_instance_v2.node_server[0]: Still creating... [10s elapsed]\nopenstack_compute_instance_v2.node_server[5]: Still creating... [10s elapsed]\nopenstack_compute_instance_v2.node_server[4]: Still creating... [10s elapsed]\nopenstack_compute_instance_v2.node_server[2]: Still creating... [10s elapsed]\nopenstack_compute_instance_v2.node_server[3]: Still creating... [10s elapsed]\nopenstack_compute_instance_v2.node_server[0]: Still creating... [20s elapsed]\nopenstack_compute_instance_v2.node_server[1]: Still creating... [20s elapsed]\nopenstack_compute_instance_v2.node_server[5]: Still creating... [20s elapsed]\nopenstack_compute_instance_v2.node_server[4]: Still creating... [20s elapsed]\nopenstack_compute_instance_v2.node_server[2]: Still creating... [20s elapsed]\nopenstack_compute_instance_v2.node_server[3]: Still creating... [20s elapsed]\nopenstack_compute_instance_v2.node_server[3]: Creation complete after 21s [id=05d0fb61-2d0a-4090-b8f0-bd104996720e]\nopenstack_compute_instance_v2.node_server[0]: Creation complete after 21s [id=aaa02988-9362-4770-8564-8f17ae7318b2]\nopenstack_compute_instance_v2.node_server[1]: Still creating... [30s elapsed]\nopenstack_compute_instance_v2.node_server[5]: Still creating... [30s elapsed]\nopenstack_compute_instance_v2.node_server[4]: Still creating... [30s elapsed]\nopenstack_compute_instance_v2.node_server[2]: Still creating... [30s elapsed]\nopenstack_compute_instance_v2.node_server[5]: Creation complete after 31s [id=f7128b88-877f-46fc-9d1d-3f4c1596413d]\nopenstack_compute_instance_v2.node_server[2]: Creation complete after 31s [id=84c66398-f55f-4354-b3cf-775d88768d57]\nopenstack_compute_instance_v2.node_server[1]: Creation complete after 31s [id=45422329-13ac-46a3-a475-2d2bacc66157]\nopenstack_compute_instance_v2.node_server[4]: Creation complete after 31s [id=e22977cc-084b-4b2a-bbb9-fad0560981cd]\nnull_resource.node_semaphore: Creating...\nnull_resource.node_semaphore: Creation complete after 0s [id=45525888549353150]\nopenstack_compute_volume_attach_v2.node_volume_attachment[4]: Creating...\nopenstack_compute_volume_attach_v2.node_volume_attachment[1]: Creating...\nopenstack_compute_volume_attach_v2.node_volume_attachment[7]: Creating...\nopenstack_compute_volume_attach_v2.node_volume_attachment[6]: Creating...\nopenstack_compute_volume_attach_v2.node_volume_attachment[3]: Creating...\nopenstack_compute_volume_attach_v2.node_volume_attachment[8]: Creating...\nopenstack_compute_volume_attach_v2.node_volume_attachment[2]: Creating...\nopenstack_compute_volume_attach_v2.node_volume_attachment[5]: Creating...\nopenstack_compute_instance_v2.manager_server: Creating...\nopenstack_compute_volume_attach_v2.node_volume_attachment[0]: Creating...\nopenstack_compute_volume_attach_v2.node_volume_attachment[6]: Creation complete after 4s [id=05d0fb61-2d0a-4090-b8f0-bd104996720e/4534650b-23a0-46c8-8bf1-44520caee637]\nopenstack_compute_volume_attach_v2.node_volume_attachment[1]: Creation complete after 4s [id=e22977cc-084b-4b2a-bbb9-fad0560981cd/5660bb62-e2c9-45b2-ae11-8ab0f829b194]\nopenstack_compute_volume_attach_v2.node_volume_attachment[8]: Creation complete after 4s [id=f7128b88-877f-46fc-9d1d-3f4c1596413d/5f47789f-2a1a-442a-8a3c-644a4f92e5a8]\nopenstack_compute_volume_attach_v2.node_volume_attachment[4]: Creation complete after 10s [id=e22977cc-084b-4b2a-bbb9-fad0560981cd/fb58bf16-a1c6-4abb-925b-29c6a6a6f100]\nopenstack_compute_volume_attach_v2.node_volume_attachment[3]: Creation complete after 10s [id=05d0fb61-2d0a-4090-b8f0-bd104996720e/6644f191-69e9-4f3a-baff-cffe4152595b]\nopenstack_compute_volume_attach_v2.node_volume_attachment[2]: Creation complete after 10s [id=f7128b88-877f-46fc-9d1d-3f4c1596413d/6f4643f0-329b-4ae2-97c7-d348648fd970]\nopenstack_compute_volume_attach_v2.node_volume_attachment[0]: Creation complete after 10s [id=05d0fb61-2d0a-4090-b8f0-bd104996720e/9c6b6763-9704-41ec-9b6d-97eb80a5145c]\nopenstack_compute_volume_attach_v2.node_volume_attachment[7]: Creation complete after 10s [id=e22977cc-084b-4b2a-bbb9-fad0560981cd/f659b61f-c39f-426f-a640-6a753951661d]\nopenstack_compute_volume_attach_v2.node_volume_attachment[5]: Creation complete after 10s [id=f7128b88-877f-46fc-9d1d-3f4c1596413d/1558935b-4c66-4614-97e5-cda14abd2641]\nopenstack_compute_instance_v2.manager_server: Still creating... [10s elapsed]\nopenstack_compute_instance_v2.manager_server: Still creating... [20s elapsed]\nopenstack_compute_instance_v2.manager_server: Creation complete after 21s [id=17506fa8-3f0b-4351-b562-9644c43fa35e]\n\nApply complete! Resources: 64 added, 0 changed, 0 destroyed.\n\nOutputs:\n\nmanager_address = <sensitive>\nprivate_key = <sensitive>",
                            "stdout_lines": [
                                "",
                                "Initializing the backend...",
                                "",
                                "Initializing provider plugins...",
                                "- terraform.io/builtin/terraform is built in to OpenTofu",
                                "- Finding hashicorp/local versions matching \">= 2.2.0\"...",
                                "- Finding latest version of hashicorp/null...",
                                "- Finding terraform-provider-openstack/openstack versions matching \">= 1.53.0\"...",
                                "- Installing hashicorp/local v2.7.0...",
                                "- Installed hashicorp/local v2.7.0 (signed, key ID 0C0AF313E5FD9F80)",
                                "- Installing hashicorp/null v3.2.4...",
                                "- Installed hashicorp/null v3.2.4 (signed, key ID 0C0AF313E5FD9F80)",
                                "- Installing terraform-provider-openstack/openstack v3.4.0...",
                                "- Installed terraform-provider-openstack/openstack v3.4.0 (signed, key ID 4F80527A391BEFD2)",
                                "",
                                "Providers are signed by their developers.",
                                "If you'd like to know more about provider signing, you can read about it here:",
                                "https://opentofu.org/docs/cli/plugins/signing/",
                                "",
                                "OpenTofu has created a lock file .terraform.lock.hcl to record the provider",
                                "selections it made above. Include this file in your version control repository",
                                "so that OpenTofu can guarantee to make the same selections by default when",
                                "you run \"tofu init\" in the future.",
                                "",
                                "OpenTofu has been successfully initialized!",
                                "",
                                "You may now begin working with OpenTofu. Try running \"tofu plan\" to see",
                                "any changes that are required for your infrastructure. All OpenTofu commands",
                                "should now work.",
                                "",
                                "If you ever set or change modules or backend configuration for OpenTofu,",
                                "rerun this command to reinitialize your working directory. If you forget, other",
                                "commands will detect it and remind you to do so if necessary.",
                                "Created and switched to workspace \"ci\"!",
                                "",
                                "You're now on a new, empty workspace. Workspaces isolate their state,",
                                "so if you run \"tofu plan\" OpenTofu will not see any existing state",
                                "for this configuration.",
                                "ci.auto.tfvars",
                                "default_custom.tf",
                                "data.openstack_networking_network_v2.public: Reading...",
                                "data.openstack_networking_network_v2.public: Read complete after 1s [id=e6be7364-bfd8-4de7-8120-8f41c69a139a]",
                                "",
                                "OpenTofu used the selected providers to generate the following execution",
                                "plan. Resource actions are indicated with the following symbols:",
                                "  + create",
                                " <= read (data resources)",
                                "",
                                "OpenTofu will perform the following actions:",
                                "",
                                "  # data.openstack_images_image_v2.image will be read during apply",
                                "  # (config refers to values not yet known)",
                                " <= data \"openstack_images_image_v2\" \"image\" {",
                                "      + checksum    = (known after apply)",
                                "      + created_at  = (known after apply)",
                                "      + file        = (known after apply)",
                                "      + id          = (known after apply)",
                                "      + metadata    = (known after apply)",
                                "      + min_disk_gb = (known after apply)",
                                "      + min_ram_mb  = (known after apply)",
                                "      + most_recent = true",
                                "      + name        = (known after apply)",
                                "      + protected   = (known after apply)",
                                "      + region      = (known after apply)",
                                "      + schema      = (known after apply)",
                                "      + size_bytes  = (known after apply)",
                                "      + tags        = (known after apply)",
                                "      + updated_at  = (known after apply)",
                                "    }",
                                "",
                                "  # data.openstack_images_image_v2.image_node will be read during apply",
                                "  # (config refers to values not yet known)",
                                " <= data \"openstack_images_image_v2\" \"image_node\" {",
                                "      + checksum    = (known after apply)",
                                "      + created_at  = (known after apply)",
                                "      + file        = (known after apply)",
                                "      + id          = (known after apply)",
                                "      + metadata    = (known after apply)",
                                "      + min_disk_gb = (known after apply)",
                                "      + min_ram_mb  = (known after apply)",
                                "      + most_recent = true",
                                "      + name        = (known after apply)",
                                "      + protected   = (known after apply)",
                                "      + region      = (known after apply)",
                                "      + schema      = (known after apply)",
                                "      + size_bytes  = (known after apply)",
                                "      + tags        = (known after apply)",
                                "      + updated_at  = (known after apply)",
                                "    }",
                                "",
                                "  # local_file.MANAGER_ADDRESS will be created",
                                "  + resource \"local_file\" \"MANAGER_ADDRESS\" {",
                                "      + content              = (known after apply)",
                                "      + content_base64sha256 = (known after apply)",
                                "      + content_base64sha512 = (known after apply)",
                                "      + content_md5          = (known after apply)",
                                "      + content_sha1         = (known after apply)",
                                "      + content_sha256       = (known after apply)",
                                "      + content_sha512       = (known after apply)",
                                "      + directory_permission = \"0777\"",
                                "      + file_permission      = \"0644\"",
                                "      + filename             = \".MANAGER_ADDRESS.ci\"",
                                "      + id                   = (known after apply)",
                                "    }",
                                "",
                                "  # local_file.id_rsa_pub will be created",
                                "  + resource \"local_file\" \"id_rsa_pub\" {",
                                "      + content              = (known after apply)",
                                "      + content_base64sha256 = (known after apply)",
                                "      + content_base64sha512 = (known after apply)",
                                "      + content_md5          = (known after apply)",
                                "      + content_sha1         = (known after apply)",
                                "      + content_sha256       = (known after apply)",
                                "      + content_sha512       = (known after apply)",
                                "      + directory_permission = \"0777\"",
                                "      + file_permission      = \"0644\"",
                                "      + filename             = \".id_rsa.ci.pub\"",
                                "      + id                   = (known after apply)",
                                "    }",
                                "",
                                "  # local_file.inventory will be created",
                                "  + resource \"local_file\" \"inventory\" {",
                                "      + content              = (known after apply)",
                                "      + content_base64sha256 = (known after apply)",
                                "      + content_base64sha512 = (known after apply)",
                                "      + content_md5          = (known after apply)",
                                "      + content_sha1         = (known after apply)",
                                "      + content_sha256       = (known after apply)",
                                "      + content_sha512       = (known after apply)",
                                "      + directory_permission = \"0777\"",
                                "      + file_permission      = \"0644\"",
                                "      + filename             = \"inventory.ci\"",
                                "      + id                   = (known after apply)",
                                "    }",
                                "",
                                "  # local_sensitive_file.id_rsa will be created",
                                "  + resource \"local_sensitive_file\" \"id_rsa\" {",
                                "      + content              = (sensitive value)",
                                "      + content_base64sha256 = (known after apply)",
                                "      + content_base64sha512 = (known after apply)",
                                "      + content_md5          = (known after apply)",
                                "      + content_sha1         = (known after apply)",
                                "      + content_sha256       = (known after apply)",
                                "      + content_sha512       = (known after apply)",
                                "      + directory_permission = \"0700\"",
                                "      + file_permission      = \"0600\"",
                                "      + filename             = \".id_rsa.ci\"",
                                "      + id                   = (known after apply)",
                                "    }",
                                "",
                                "  # null_resource.node_semaphore will be created",
                                "  + resource \"null_resource\" \"node_semaphore\" {",
                                "      + id = (known after apply)",
                                "    }",
                                "",
                                "  # openstack_blockstorage_volume_v3.manager_base_volume[0] will be created",
                                "  + resource \"openstack_blockstorage_volume_v3\" \"manager_base_volume\" {",
                                "      + attachment           = (known after apply)",
                                "      + availability_zone    = \"nova\"",
                                "      + id                   = (known after apply)",
                                "      + image_id             = (known after apply)",
                                "      + metadata             = (known after apply)",
                                "      + name                 = \"testbed-volume-manager-base\"",
                                "      + region               = (known after apply)",
                                "      + size                 = 80",
                                "      + volume_retype_policy = \"never\"",
                                "      + volume_type          = \"ssd\"",
                                "    }",
                                "",
                                "  # openstack_blockstorage_volume_v3.node_base_volume[0] will be created",
                                "  + resource \"openstack_blockstorage_volume_v3\" \"node_base_volume\" {",
                                "      + attachment           = (known after apply)",
                                "      + availability_zone    = \"nova\"",
                                "      + id                   = (known after apply)",
                                "      + image_id             = (known after apply)",
                                "      + metadata             = (known after apply)",
                                "      + name                 = \"testbed-volume-0-node-base\"",
                                "      + region               = (known after apply)",
                                "      + size                 = 80",
                                "      + volume_retype_policy = \"never\"",
                                "      + volume_type          = \"ssd\"",
                                "    }",
                                "",
                                "  # openstack_blockstorage_volume_v3.node_base_volume[1] will be created",
                                "  + resource \"openstack_blockstorage_volume_v3\" \"node_base_volume\" {",
                                "      + attachment           = (known after apply)",
                                "      + availability_zone    = \"nova\"",
                                "      + id                   = (known after apply)",
                                "      + image_id             = (known after apply)",
                                "      + metadata             = (known after apply)",
                                "      + name                 = \"testbed-volume-1-node-base\"",
                                "      + region               = (known after apply)",
                                "      + size                 = 80",
                                "      + volume_retype_policy = \"never\"",
                                "      + volume_type          = \"ssd\"",
                                "    }",
                                "",
                                "  # openstack_blockstorage_volume_v3.node_base_volume[2] will be created",
                                "  + resource \"openstack_blockstorage_volume_v3\" \"node_base_volume\" {",
                                "      + attachment           = (known after apply)",
                                "      + availability_zone    = \"nova\"",
                                "      + id                   = (known after apply)",
                                "      + image_id             = (known after apply)",
                                "      + metadata             = (known after apply)",
                                "      + name                 = \"testbed-volume-2-node-base\"",
                                "      + region               = (known after apply)",
                                "      + size                 = 80",
                                "      + volume_retype_policy = \"never\"",
                                "      + volume_type          = \"ssd\"",
                                "    }",
                                "",
                                "  # openstack_blockstorage_volume_v3.node_base_volume[3] will be created",
                                "  + resource \"openstack_blockstorage_volume_v3\" \"node_base_volume\" {",
                                "      + attachment           = (known after apply)",
                                "      + availability_zone    = \"nova\"",
                                "      + id                   = (known after apply)",
                                "      + image_id             = (known after apply)",
                                "      + metadata             = (known after apply)",
                                "      + name                 = \"testbed-volume-3-node-base\"",
                                "      + region               = (known after apply)",
                                "      + size                 = 80",
                                "      + volume_retype_policy = \"never\"",
                                "      + volume_type          = \"ssd\"",
                                "    }",
                                "",
                                "  # openstack_blockstorage_volume_v3.node_base_volume[4] will be created",
                                "  + resource \"openstack_blockstorage_volume_v3\" \"node_base_volume\" {",
                                "      + attachment           = (known after apply)",
                                "      + availability_zone    = \"nova\"",
                                "      + id                   = (known after apply)",
                                "      + image_id             = (known after apply)",
                                "      + metadata             = (known after apply)",
                                "      + name                 = \"testbed-volume-4-node-base\"",
                                "      + region               = (known after apply)",
                                "      + size                 = 80",
                                "      + volume_retype_policy = \"never\"",
                                "      + volume_type          = \"ssd\"",
                                "    }",
                                "",
                                "  # openstack_blockstorage_volume_v3.node_base_volume[5] will be created",
                                "  + resource \"openstack_blockstorage_volume_v3\" \"node_base_volume\" {",
                                "      + attachment           = (known after apply)",
                                "      + availability_zone    = \"nova\"",
                                "      + id                   = (known after apply)",
                                "      + image_id             = (known after apply)",
                                "      + metadata             = (known after apply)",
                                "      + name                 = \"testbed-volume-5-node-base\"",
                                "      + region               = (known after apply)",
                                "      + size                 = 80",
                                "      + volume_retype_policy = \"never\"",
                                "      + volume_type          = \"ssd\"",
                                "    }",
                                "",
                                "  # openstack_blockstorage_volume_v3.node_volume[0] will be created",
                                "  + resource \"openstack_blockstorage_volume_v3\" \"node_volume\" {",
                                "      + attachment           = (known after apply)",
                                "      + availability_zone    = \"nova\"",
                                "      + id                   = (known after apply)",
                                "      + metadata             = (known after apply)",
                                "      + name                 = \"testbed-volume-0-node-3\"",
                                "      + region               = (known after apply)",
                                "      + size                 = 20",
                                "      + volume_retype_policy = \"never\"",
                                "      + volume_type          = \"ssd\"",
                                "    }",
                                "",
                                "  # openstack_blockstorage_volume_v3.node_volume[1] will be created",
                                "  + resource \"openstack_blockstorage_volume_v3\" \"node_volume\" {",
                                "      + attachment           = (known after apply)",
                                "      + availability_zone    = \"nova\"",
                                "      + id                   = (known after apply)",
                                "      + metadata             = (known after apply)",
                                "      + name                 = \"testbed-volume-1-node-4\"",
                                "      + region               = (known after apply)",
                                "      + size                 = 20",
                                "      + volume_retype_policy = \"never\"",
                                "      + volume_type          = \"ssd\"",
                                "    }",
                                "",
                                "  # openstack_blockstorage_volume_v3.node_volume[2] will be created",
                                "  + resource \"openstack_blockstorage_volume_v3\" \"node_volume\" {",
                                "      + attachment           = (known after apply)",
                                "      + availability_zone    = \"nova\"",
                                "      + id                   = (known after apply)",
                                "      + metadata             = (known after apply)",
                                "      + name                 = \"testbed-volume-2-node-5\"",
                                "      + region               = (known after apply)",
                                "      + size                 = 20",
                                "      + volume_retype_policy = \"never\"",
                                "      + volume_type          = \"ssd\"",
                                "    }",
                                "",
                                "  # openstack_blockstorage_volume_v3.node_volume[3] will be created",
                                "  + resource \"openstack_blockstorage_volume_v3\" \"node_volume\" {",
                                "      + attachment           = (known after apply)",
                                "      + availability_zone    = \"nova\"",
                                "      + id                   = (known after apply)",
                                "      + metadata             = (known after apply)",
                                "      + name                 = \"testbed-volume-3-node-3\"",
                                "      + region               = (known after apply)",
                                "      + size                 = 20",
                                "      + volume_retype_policy = \"never\"",
                                "      + volume_type          = \"ssd\"",
                                "    }",
                                "",
                                "  # openstack_blockstorage_volume_v3.node_volume[4] will be created",
                                "  + resource \"openstack_blockstorage_volume_v3\" \"node_volume\" {",
                                "      + attachment           = (known after apply)",
                                "      + availability_zone    = \"nova\"",
                                "      + id                   = (known after apply)",
                                "      + metadata             = (known after apply)",
                                "      + name                 = \"testbed-volume-4-node-4\"",
                                "      + region               = (known after apply)",
                                "      + size                 = 20",
                                "      + volume_retype_policy = \"never\"",
                                "      + volume_type          = \"ssd\"",
                                "    }",
                                "",
                                "  # openstack_blockstorage_volume_v3.node_volume[5] will be created",
                                "  + resource \"openstack_blockstorage_volume_v3\" \"node_volume\" {",
                                "      + attachment           = (known after apply)",
                                "      + availability_zone    = \"nova\"",
                                "      + id                   = (known after apply)",
                                "      + metadata             = (known after apply)",
                                "      + name                 = \"testbed-volume-5-node-5\"",
                                "      + region               = (known after apply)",
                                "      + size                 = 20",
                                "      + volume_retype_policy = \"never\"",
                                "      + volume_type          = \"ssd\"",
                                "    }",
                                "",
                                "  # openstack_blockstorage_volume_v3.node_volume[6] will be created",
                                "  + resource \"openstack_blockstorage_volume_v3\" \"node_volume\" {",
                                "      + attachment           = (known after apply)",
                                "      + availability_zone    = \"nova\"",
                                "      + id                   = (known after apply)",
                                "      + metadata             = (known after apply)",
                                "      + name                 = \"testbed-volume-6-node-3\"",
                                "      + region               = (known after apply)",
                                "      + size                 = 20",
                                "      + volume_retype_policy = \"never\"",
                                "      + volume_type          = \"ssd\"",
                                "    }",
                                "",
                                "  # openstack_blockstorage_volume_v3.node_volume[7] will be created",
                                "  + resource \"openstack_blockstorage_volume_v3\" \"node_volume\" {",
                                "      + attachment           = (known after apply)",
                                "      + availability_zone    = \"nova\"",
                                "      + id                   = (known after apply)",
                                "      + metadata             = (known after apply)",
                                "      + name                 = \"testbed-volume-7-node-4\"",
                                "      + region               = (known after apply)",
                                "      + size                 = 20",
                                "      + volume_retype_policy = \"never\"",
                                "      + volume_type          = \"ssd\"",
                                "    }",
                                "",
                                "  # openstack_blockstorage_volume_v3.node_volume[8] will be created",
                                "  + resource \"openstack_blockstorage_volume_v3\" \"node_volume\" {",
                                "      + attachment           = (known after apply)",
                                "      + availability_zone    = \"nova\"",
                                "      + id                   = (known after apply)",
                                "      + metadata             = (known after apply)",
                                "      + name                 = \"testbed-volume-8-node-5\"",
                                "      + region               = (known after apply)",
                                "      + size                 = 20",
                                "      + volume_retype_policy = \"never\"",
                                "      + volume_type          = \"ssd\"",
                                "    }",
                                "",
                                "  # openstack_compute_instance_v2.manager_server will be created",
                                "  + resource \"openstack_compute_instance_v2\" \"manager_server\" {",
                                "      + access_ip_v4        = (known after apply)",
                                "      + access_ip_v6        = (known after apply)",
                                "      + all_metadata        = (known after apply)",
                                "      + all_tags            = (known after apply)",
                                "      + availability_zone   = \"nova\"",
                                "      + config_drive        = true",
                                "      + created             = (known after apply)",
                                "      + flavor_id           = (known after apply)",
                                "      + flavor_name         = \"OSISM-4V-16\"",
                                "      + force_delete        = false",
                                "      + hypervisor_hostname = (known after apply)",
                                "      + id                  = (known after apply)",
                                "      + image_id            = (known after apply)",
                                "      + image_name          = (known after apply)",
                                "      + key_pair            = \"testbed\"",
                                "      + name                = \"testbed-manager\"",
                                "      + power_state         = \"active\"",
                                "      + region              = (known after apply)",
                                "      + security_groups     = (known after apply)",
                                "      + stop_before_destroy = false",
                                "      + updated             = (known after apply)",
                                "      + user_data           = (sensitive value)",
                                "",
                                "      + block_device {",
                                "          + boot_index            = 0",
                                "          + delete_on_termination = false",
                                "          + destination_type      = \"volume\"",
                                "          + multiattach           = false",
                                "          + source_type           = \"volume\"",
                                "          + uuid                  = (known after apply)",
                                "        }",
                                "",
                                "      + network {",
                                "          + access_network = false",
                                "          + fixed_ip_v4    = (known after apply)",
                                "          + fixed_ip_v6    = (known after apply)",
                                "          + mac            = (known after apply)",
                                "          + name           = (known after apply)",
                                "          + port           = (known after apply)",
                                "          + uuid           = (known after apply)",
                                "        }",
                                "    }",
                                "",
                                "  # openstack_compute_instance_v2.node_server[0] will be created",
                                "  + resource \"openstack_compute_instance_v2\" \"node_server\" {",
                                "      + access_ip_v4        = (known after apply)",
                                "      + access_ip_v6        = (known after apply)",
                                "      + all_metadata        = (known after apply)",
                                "      + all_tags            = (known after apply)",
                                "      + availability_zone   = \"nova\"",
                                "      + config_drive        = true",
                                "      + created             = (known after apply)",
                                "      + flavor_id           = (known after apply)",
                                "      + flavor_name         = \"OSISM-8V-32\"",
                                "      + force_delete        = false",
                                "      + hypervisor_hostname = (known after apply)",
                                "      + id                  = (known after apply)",
                                "      + image_id            = (known after apply)",
                                "      + image_name          = (known after apply)",
                                "      + key_pair            = \"testbed\"",
                                "      + name                = \"testbed-node-0\"",
                                "      + power_state         = \"active\"",
                                "      + region              = (known after apply)",
                                "      + security_groups     = (known after apply)",
                                "      + stop_before_destroy = false",
                                "      + updated             = (known after apply)",
                                "      + user_data           = \"ae09e46b224a6ca206a9ed4f8f8a4f8520827854\"",
                                "",
                                "      + block_device {",
                                "          + boot_index            = 0",
                                "          + delete_on_termination = false",
                                "          + destination_type      = \"volume\"",
                                "          + multiattach           = false",
                                "          + source_type           = \"volume\"",
                                "          + uuid                  = (known after apply)",
                                "        }",
                                "",
                                "      + network {",
                                "          + access_network = false",
                                "          + fixed_ip_v4    = (known after apply)",
                                "          + fixed_ip_v6    = (known after apply)",
                                "          + mac            = (known after apply)",
                                "          + name           = (known after apply)",
                                "          + port           = (known after apply)",
                                "          + uuid           = (known after apply)",
                                "        }",
                                "    }",
                                "",
                                "  # openstack_compute_instance_v2.node_server[1] will be created",
                                "  + resource \"openstack_compute_instance_v2\" \"node_server\" {",
                                "      + access_ip_v4        = (known after apply)",
                                "      + access_ip_v6        = (known after apply)",
                                "      + all_metadata        = (known after apply)",
                                "      + all_tags            = (known after apply)",
                                "      + availability_zone   = \"nova\"",
                                "      + config_drive        = true",
                                "      + created             = (known after apply)",
                                "      + flavor_id           = (known after apply)",
                                "      + flavor_name         = \"OSISM-8V-32\"",
                                "      + force_delete        = false",
                                "      + hypervisor_hostname = (known after apply)",
                                "      + id                  = (known after apply)",
                                "      + image_id            = (known after apply)",
                                "      + image_name          = (known after apply)",
                                "      + key_pair            = \"testbed\"",
                                "      + name                = \"testbed-node-1\"",
                                "      + power_state         = \"active\"",
                                "      + region              = (known after apply)",
                                "      + security_groups     = (known after apply)",
                                "      + stop_before_destroy = false",
                                "      + updated             = (known after apply)",
                                "      + user_data           = \"ae09e46b224a6ca206a9ed4f8f8a4f8520827854\"",
                                "",
                                "      + block_device {",
                                "          + boot_index            = 0",
                                "          + delete_on_termination = false",
                                "          + destination_type      = \"volume\"",
                                "          + multiattach           = false",
                                "          + source_type           = \"volume\"",
                                "          + uuid                  = (known after apply)",
                                "        }",
                                "",
                                "      + network {",
                                "          + access_network = false",
                                "          + fixed_ip_v4    = (known after apply)",
                                "          + fixed_ip_v6    = (known after apply)",
                                "          + mac            = (known after apply)",
                                "          + name           = (known after apply)",
                                "          + port           = (known after apply)",
                                "          + uuid           = (known after apply)",
                                "        }",
                                "    }",
                                "",
                                "  # openstack_compute_instance_v2.node_server[2] will be created",
                                "  + resource \"openstack_compute_instance_v2\" \"node_server\" {",
                                "      + access_ip_v4        = (known after apply)",
                                "      + access_ip_v6        = (known after apply)",
                                "      + all_metadata        = (known after apply)",
                                "      + all_tags            = (known after apply)",
                                "      + availability_zone   = \"nova\"",
                                "      + config_drive        = true",
                                "      + created             = (known after apply)",
                                "      + flavor_id           = (known after apply)",
                                "      + flavor_name         = \"OSISM-8V-32\"",
                                "      + force_delete        = false",
                                "      + hypervisor_hostname = (known after apply)",
                                "      + id                  = (known after apply)",
                                "      + image_id            = (known after apply)",
                                "      + image_name          = (known after apply)",
                                "      + key_pair            = \"testbed\"",
                                "      + name                = \"testbed-node-2\"",
                                "      + power_state         = \"active\"",
                                "      + region              = (known after apply)",
                                "      + security_groups     = (known after apply)",
                                "      + stop_before_destroy = false",
                                "      + updated             = (known after apply)",
                                "      + user_data           = \"ae09e46b224a6ca206a9ed4f8f8a4f8520827854\"",
                                "",
                                "      + block_device {",
                                "          + boot_index            = 0",
                                "          + delete_on_termination = false",
                                "          + destination_type      = \"volume\"",
                                "          + multiattach           = false",
                                "          + source_type           = \"volume\"",
                                "          + uuid                  = (known after apply)",
                                "        }",
                                "",
                                "      + network {",
                                "          + access_network = false",
                                "          + fixed_ip_v4    = (known after apply)",
                                "          + fixed_ip_v6    = (known after apply)",
                                "          + mac            = (known after apply)",
                                "          + name           = (known after apply)",
                                "          + port           = (known after apply)",
                                "          + uuid           = (known after apply)",
                                "        }",
                                "    }",
                                "",
                                "  # openstack_compute_instance_v2.node_server[3] will be created",
                                "  + resource \"openstack_compute_instance_v2\" \"node_server\" {",
                                "      + access_ip_v4        = (known after apply)",
                                "      + access_ip_v6        = (known after apply)",
                                "      + all_metadata        = (known after apply)",
                                "      + all_tags            = (known after apply)",
                                "      + availability_zone   = \"nova\"",
                                "      + config_drive        = true",
                                "      + created             = (known after apply)",
                                "      + flavor_id           = (known after apply)",
                                "      + flavor_name         = \"OSISM-8V-32\"",
                                "      + force_delete        = false",
                                "      + hypervisor_hostname = (known after apply)",
                                "      + id                  = (known after apply)",
                                "      + image_id            = (known after apply)",
                                "      + image_name          = (known after apply)",
                                "      + key_pair            = \"testbed\"",
                                "      + name                = \"testbed-node-3\"",
                                "      + power_state         = \"active\"",
                                "      + region              = (known after apply)",
                                "      + security_groups     = (known after apply)",
                                "      + stop_before_destroy = false",
                                "      + updated             = (known after apply)",
                                "      + user_data           = \"ae09e46b224a6ca206a9ed4f8f8a4f8520827854\"",
                                "",
                                "      + block_device {",
                                "          + boot_index            = 0",
                                "          + delete_on_termination = false",
                                "          + destination_type      = \"volume\"",
                                "          + multiattach           = false",
                                "          + source_type           = \"volume\"",
                                "          + uuid                  = (known after apply)",
                                "        }",
                                "",
                                "      + network {",
                                "          + access_network = false",
                                "          + fixed_ip_v4    = (known after apply)",
                                "          + fixed_ip_v6    = (known after apply)",
                                "          + mac            = (known after apply)",
                                "          + name           = (known after apply)",
                                "          + port           = (known after apply)",
                                "          + uuid           = (known after apply)",
                                "        }",
                                "    }",
                                "",
                                "  # openstack_compute_instance_v2.node_server[4] will be created",
                                "  + resource \"openstack_compute_instance_v2\" \"node_server\" {",
                                "      + access_ip_v4        = (known after apply)",
                                "      + access_ip_v6        = (known after apply)",
                                "      + all_metadata        = (known after apply)",
                                "      + all_tags            = (known after apply)",
                                "      + availability_zone   = \"nova\"",
                                "      + config_drive        = true",
                                "      + created             = (known after apply)",
                                "      + flavor_id           = (known after apply)",
                                "      + flavor_name         = \"OSISM-8V-32\"",
                                "      + force_delete        = false",
                                "      + hypervisor_hostname = (known after apply)",
                                "      + id                  = (known after apply)",
                                "      + image_id            = (known after apply)",
                                "      + image_name          = (known after apply)",
                                "      + key_pair            = \"testbed\"",
                                "      + name                = \"testbed-node-4\"",
                                "      + power_state         = \"active\"",
                                "      + region              = (known after apply)",
                                "      + security_groups     = (known after apply)",
                                "      + stop_before_destroy = false",
                                "      + updated             = (known after apply)",
                                "      + user_data           = \"ae09e46b224a6ca206a9ed4f8f8a4f8520827854\"",
                                "",
                                "      + block_device {",
                                "          + boot_index            = 0",
                                "          + delete_on_termination = false",
                                "          + destination_type      = \"volume\"",
                                "          + multiattach           = false",
                                "          + source_type           = \"volume\"",
                                "          + uuid                  = (known after apply)",
                                "        }",
                                "",
                                "      + network {",
                                "          + access_network = false",
                                "          + fixed_ip_v4    = (known after apply)",
                                "          + fixed_ip_v6    = (known after apply)",
                                "          + mac            = (known after apply)",
                                "          + name           = (known after apply)",
                                "          + port           = (known after apply)",
                                "          + uuid           = (known after apply)",
                                "        }",
                                "    }",
                                "",
                                "  # openstack_compute_instance_v2.node_server[5] will be created",
                                "  + resource \"openstack_compute_instance_v2\" \"node_server\" {",
                                "      + access_ip_v4        = (known after apply)",
                                "      + access_ip_v6        = (known after apply)",
                                "      + all_metadata        = (known after apply)",
                                "      + all_tags            = (known after apply)",
                                "      + availability_zone   = \"nova\"",
                                "      + config_drive        = true",
                                "      + created             = (known after apply)",
                                "      + flavor_id           = (known after apply)",
                                "      + flavor_name         = \"OSISM-8V-32\"",
                                "      + force_delete        = false",
                                "      + hypervisor_hostname = (known after apply)",
                                "      + id                  = (known after apply)",
                                "      + image_id            = (known after apply)",
                                "      + image_name          = (known after apply)",
                                "      + key_pair            = \"testbed\"",
                                "      + name                = \"testbed-node-5\"",
                                "      + power_state         = \"active\"",
                                "      + region              = (known after apply)",
                                "      + security_groups     = (known after apply)",
                                "      + stop_before_destroy = false",
                                "      + updated             = (known after apply)",
                                "      + user_data           = \"ae09e46b224a6ca206a9ed4f8f8a4f8520827854\"",
                                "",
                                "      + block_device {",
                                "          + boot_index            = 0",
                                "          + delete_on_termination = false",
                                "          + destination_type      = \"volume\"",
                                "          + multiattach           = false",
                                "          + source_type           = \"volume\"",
                                "          + uuid                  = (known after apply)",
                                "        }",
                                "",
                                "      + network {",
                                "          + access_network = false",
                                "          + fixed_ip_v4    = (known after apply)",
                                "          + fixed_ip_v6    = (known after apply)",
                                "          + mac            = (known after apply)",
                                "          + name           = (known after apply)",
                                "          + port           = (known after apply)",
                                "          + uuid           = (known after apply)",
                                "        }",
                                "    }",
                                "",
                                "  # openstack_compute_keypair_v2.key will be created",
                                "  + resource \"openstack_compute_keypair_v2\" \"key\" {",
                                "      + fingerprint = (known after apply)",
                                "      + id          = (known after apply)",
                                "      + name        = \"testbed\"",
                                "      + private_key = (sensitive value)",
                                "      + public_key  = (known after apply)",
                                "      + region      = (known after apply)",
                                "      + user_id     = (known after apply)",
                                "    }",
                                "",
                                "  # openstack_compute_volume_attach_v2.node_volume_attachment[0] will be created",
                                "  + resource \"openstack_compute_volume_attach_v2\" \"node_volume_attachment\" {",
                                "      + device      = (known after apply)",
                                "      + id          = (known after apply)",
                                "      + instance_id = (known after apply)",
                                "      + region      = (known after apply)",
                                "      + volume_id   = (known after apply)",
                                "    }",
                                "",
                                "  # openstack_compute_volume_attach_v2.node_volume_attachment[1] will be created",
                                "  + resource \"openstack_compute_volume_attach_v2\" \"node_volume_attachment\" {",
                                "      + device      = (known after apply)",
                                "      + id          = (known after apply)",
                                "      + instance_id = (known after apply)",
                                "      + region      = (known after apply)",
                                "      + volume_id   = (known after apply)",
                                "    }",
                                "",
                                "  # openstack_compute_volume_attach_v2.node_volume_attachment[2] will be created",
                                "  + resource \"openstack_compute_volume_attach_v2\" \"node_volume_attachment\" {",
                                "      + device      = (known after apply)",
                                "      + id          = (known after apply)",
                                "      + instance_id = (known after apply)",
                                "      + region      = (known after apply)",
                                "      + volume_id   = (known after apply)",
                                "    }",
                                "",
                                "  # openstack_compute_volume_attach_v2.node_volume_attachment[3] will be created",
                                "  + resource \"openstack_compute_volume_attach_v2\" \"node_volume_attachment\" {",
                                "      + device      = (known after apply)",
                                "      + id          = (known after apply)",
                                "      + instance_id = (known after apply)",
                                "      + region      = (known after apply)",
                                "      + volume_id   = (known after apply)",
                                "    }",
                                "",
                                "  # openstack_compute_volume_attach_v2.node_volume_attachment[4] will be created",
                                "  + resource \"openstack_compute_volume_attach_v2\" \"node_volume_attachment\" {",
                                "      + device      = (known after apply)",
                                "      + id          = (known after apply)",
                                "      + instance_id = (known after apply)",
                                "      + region      = (known after apply)",
                                "      + volume_id   = (known after apply)",
                                "    }",
                                "",
                                "  # openstack_compute_volume_attach_v2.node_volume_attachment[5] will be created",
                                "  + resource \"openstack_compute_volume_attach_v2\" \"node_volume_attachment\" {",
                                "      + device      = (known after apply)",
                                "      + id          = (known after apply)",
                                "      + instance_id = (known after apply)",
                                "      + region      = (known after apply)",
                                "      + volume_id   = (known after apply)",
                                "    }",
                                "",
                                "  # openstack_compute_volume_attach_v2.node_volume_attachment[6] will be created",
                                "  + resource \"openstack_compute_volume_attach_v2\" \"node_volume_attachment\" {",
                                "      + device      = (known after apply)",
                                "      + id          = (known after apply)",
                                "      + instance_id = (known after apply)",
                                "      + region      = (known after apply)",
                                "      + volume_id   = (known after apply)",
                                "    }",
                                "",
                                "  # openstack_compute_volume_attach_v2.node_volume_attachment[7] will be created",
                                "  + resource \"openstack_compute_volume_attach_v2\" \"node_volume_attachment\" {",
                                "      + device      = (known after apply)",
                                "      + id          = (known after apply)",
                                "      + instance_id = (known after apply)",
                                "      + region      = (known after apply)",
                                "      + volume_id   = (known after apply)",
                                "    }",
                                "",
                                "  # openstack_compute_volume_attach_v2.node_volume_attachment[8] will be created",
                                "  + resource \"openstack_compute_volume_attach_v2\" \"node_volume_attachment\" {",
                                "      + device      = (known after apply)",
                                "      + id          = (known after apply)",
                                "      + instance_id = (known after apply)",
                                "      + region      = (known after apply)",
                                "      + volume_id   = (known after apply)",
                                "    }",
                                "",
                                "  # openstack_networking_floatingip_associate_v2.manager_floating_ip_association will be created",
                                "  + resource \"openstack_networking_floatingip_associate_v2\" \"manager_floating_ip_association\" {",
                                "      + fixed_ip    = (known after apply)",
                                "      + floating_ip = (known after apply)",
                                "      + id          = (known after apply)",
                                "      + port_id     = (known after apply)",
                                "      + region      = (known after apply)",
                                "    }",
                                "",
                                "  # openstack_networking_floatingip_v2.manager_floating_ip will be created",
                                "  + resource \"openstack_networking_floatingip_v2\" \"manager_floating_ip\" {",
                                "      + address    = (known after apply)",
                                "      + all_tags   = (known after apply)",
                                "      + dns_domain = (known after apply)",
                                "      + dns_name   = (known after apply)",
                                "      + fixed_ip   = (known after apply)",
                                "      + id         = (known after apply)",
                                "      + pool       = \"public\"",
                                "      + port_id    = (known after apply)",
                                "      + region     = (known after apply)",
                                "      + subnet_id  = (known after apply)",
                                "      + tenant_id  = (known after apply)",
                                "    }",
                                "",
                                "  # openstack_networking_network_v2.net_management will be created",
                                "  + resource \"openstack_networking_network_v2\" \"net_management\" {",
                                "      + admin_state_up          = (known after apply)",
                                "      + all_tags                = (known after apply)",
                                "      + availability_zone_hints = [",
                                "          + \"nova\",",
                                "        ]",
                                "      + dns_domain              = (known after apply)",
                                "      + external                = (known after apply)",
                                "      + id                      = (known after apply)",
                                "      + mtu                     = (known after apply)",
                                "      + name                    = \"net-testbed-management\"",
                                "      + port_security_enabled   = (known after apply)",
                                "      + qos_policy_id           = (known after apply)",
                                "      + region                  = (known after apply)",
                                "      + shared                  = (known after apply)",
                                "      + tenant_id               = (known after apply)",
                                "      + transparent_vlan        = (known after apply)",
                                "",
                                "      + segments (known after apply)",
                                "    }",
                                "",
                                "  # openstack_networking_port_v2.manager_port_management will be created",
                                "  + resource \"openstack_networking_port_v2\" \"manager_port_management\" {",
                                "      + admin_state_up         = (known after apply)",
                                "      + all_fixed_ips          = (known after apply)",
                                "      + all_security_group_ids = (known after apply)",
                                "      + all_tags               = (known after apply)",
                                "      + device_id              = (known after apply)",
                                "      + device_owner           = (known after apply)",
                                "      + dns_assignment         = (known after apply)",
                                "      + dns_name               = (known after apply)",
                                "      + id                     = (known after apply)",
                                "      + mac_address            = (known after apply)",
                                "      + network_id             = (known after apply)",
                                "      + port_security_enabled  = (known after apply)",
                                "      + qos_policy_id          = (known after apply)",
                                "      + region                 = (known after apply)",
                                "      + security_group_ids     = (known after apply)",
                                "      + tenant_id              = (known after apply)",
                                "",
                                "      + allowed_address_pairs {",
                                "          + ip_address = \"192.168.16.8/32\"",
                                "        }",
                                "",
                                "      + binding (known after apply)",
                                "",
                                "      + fixed_ip {",
                                "          + ip_address = \"192.168.16.5\"",
                                "          + subnet_id  = (known after apply)",
                                "        }",
                                "    }",
                                "",
                                "  # openstack_networking_port_v2.node_port_management[0] will be created",
                                "  + resource \"openstack_networking_port_v2\" \"node_port_management\" {",
                                "      + admin_state_up         = (known after apply)",
                                "      + all_fixed_ips          = (known after apply)",
                                "      + all_security_group_ids = (known after apply)",
                                "      + all_tags               = (known after apply)",
                                "      + device_id              = (known after apply)",
                                "      + device_owner           = (known after apply)",
                                "      + dns_assignment         = (known after apply)",
                                "      + dns_name               = (known after apply)",
                                "      + id                     = (known after apply)",
                                "      + mac_address            = (known after apply)",
                                "      + network_id             = (known after apply)",
                                "      + port_security_enabled  = (known after apply)",
                                "      + qos_policy_id          = (known after apply)",
                                "      + region                 = (known after apply)",
                                "      + security_group_ids     = (known after apply)",
                                "      + tenant_id              = (known after apply)",
                                "",
                                "      + allowed_address_pairs {",
                                "          + ip_address = \"192.168.16.254/32\"",
                                "        }",
                                "      + allowed_address_pairs {",
                                "          + ip_address = \"192.168.16.8/32\"",
                                "        }",
                                "      + allowed_address_pairs {",
                                "          + ip_address = \"192.168.16.9/32\"",
                                "        }",
                                "",
                                "      + binding (known after apply)",
                                "",
                                "      + fixed_ip {",
                                "          + ip_address = \"192.168.16.10\"",
                                "          + subnet_id  = (known after apply)",
                                "        }",
                                "    }",
                                "",
                                "  # openstack_networking_port_v2.node_port_management[1] will be created",
                                "  + resource \"openstack_networking_port_v2\" \"node_port_management\" {",
                                "      + admin_state_up         = (known after apply)",
                                "      + all_fixed_ips          = (known after apply)",
                                "      + all_security_group_ids = (known after apply)",
                                "      + all_tags               = (known after apply)",
                                "      + device_id              = (known after apply)",
                                "      + device_owner           = (known after apply)",
                                "      + dns_assignment         = (known after apply)",
                                "      + dns_name               = (known after apply)",
                                "      + id                     = (known after apply)",
                                "      + mac_address            = (known after apply)",
                                "      + network_id             = (known after apply)",
                                "      + port_security_enabled  = (known after apply)",
                                "      + qos_policy_id          = (known after apply)",
                                "      + region                 = (known after apply)",
                                "      + security_group_ids     = (known after apply)",
                                "      + tenant_id              = (known after apply)",
                                "",
                                "      + allowed_address_pairs {",
                                "          + ip_address = \"192.168.16.254/32\"",
                                "        }",
                                "      + allowed_address_pairs {",
                                "          + ip_address = \"192.168.16.8/32\"",
                                "        }",
                                "      + allowed_address_pairs {",
                                "          + ip_address = \"192.168.16.9/32\"",
                                "        }",
                                "",
                                "      + binding (known after apply)",
                                "",
                                "      + fixed_ip {",
                                "          + ip_address = \"192.168.16.11\"",
                                "          + subnet_id  = (known after apply)",
                                "        }",
                                "    }",
                                "",
                                "  # openstack_networking_port_v2.node_port_management[2] will be created",
                                "  + resource \"openstack_networking_port_v2\" \"node_port_management\" {",
                                "      + admin_state_up         = (known after apply)",
                                "      + all_fixed_ips          = (known after apply)",
                                "      + all_security_group_ids = (known after apply)",
                                "      + all_tags               = (known after apply)",
                                "      + device_id              = (known after apply)",
                                "      + device_owner           = (known after apply)",
                                "      + dns_assignment         = (known after apply)",
                                "      + dns_name               = (known after apply)",
                                "      + id                     = (known after apply)",
                                "      + mac_address            = (known after apply)",
                                "      + network_id             = (known after apply)",
                                "      + port_security_enabled  = (known after apply)",
                                "      + qos_policy_id          = (known after apply)",
                                "      + region                 = (known after apply)",
                                "      + security_group_ids     = (known after apply)",
                                "      + tenant_id              = (known after apply)",
                                "",
                                "      + allowed_address_pairs {",
                                "          + ip_address = \"192.168.16.254/32\"",
                                "        }",
                                "      + allowed_address_pairs {",
                                "          + ip_address = \"192.168.16.8/32\"",
                                "        }",
                                "      + allowed_address_pairs {",
                                "          + ip_address = \"192.168.16.9/32\"",
                                "        }",
                                "",
                                "      + binding (known after apply)",
                                "",
                                "      + fixed_ip {",
                                "          + ip_address = \"192.168.16.12\"",
                                "          + subnet_id  = (known after apply)",
                                "        }",
                                "    }",
                                "",
                                "  # openstack_networking_port_v2.node_port_management[3] will be created",
                                "  + resource \"openstack_networking_port_v2\" \"node_port_management\" {",
                                "      + admin_state_up         = (known after apply)",
                                "      + all_fixed_ips          = (known after apply)",
                                "      + all_security_group_ids = (known after apply)",
                                "      + all_tags               = (known after apply)",
                                "      + device_id              = (known after apply)",
                                "      + device_owner           = (known after apply)",
                                "      + dns_assignment         = (known after apply)",
                                "      + dns_name               = (known after apply)",
                                "      + id                     = (known after apply)",
                                "      + mac_address            = (known after apply)",
                                "      + network_id             = (known after apply)",
                                "      + port_security_enabled  = (known after apply)",
                                "      + qos_policy_id          = (known after apply)",
                                "      + region                 = (known after apply)",
                                "      + security_group_ids     = (known after apply)",
                                "      + tenant_id              = (known after apply)",
                                "",
                                "      + allowed_address_pairs {",
                                "          + ip_address = \"192.168.16.254/32\"",
                                "        }",
                                "      + allowed_address_pairs {",
                                "          + ip_address = \"192.168.16.8/32\"",
                                "        }",
                                "      + allowed_address_pairs {",
                                "          + ip_address = \"192.168.16.9/32\"",
                                "        }",
                                "",
                                "      + binding (known after apply)",
                                "",
                                "      + fixed_ip {",
                                "          + ip_address = \"192.168.16.13\"",
                                "          + subnet_id  = (known after apply)",
                                "        }",
                                "    }",
                                "",
                                "  # openstack_networking_port_v2.node_port_management[4] will be created",
                                "  + resource \"openstack_networking_port_v2\" \"node_port_management\" {",
                                "      + admin_state_up         = (known after apply)",
                                "      + all_fixed_ips          = (known after apply)",
                                "      + all_security_group_ids = (known after apply)",
                                "      + all_tags               = (known after apply)",
                                "      + device_id              = (known after apply)",
                                "      + device_owner           = (known after apply)",
                                "      + dns_assignment         = (known after apply)",
                                "      + dns_name               = (known after apply)",
                                "      + id                     = (known after apply)",
                                "      + mac_address            = (known after apply)",
                                "      + network_id             = (known after apply)",
                                "      + port_security_enabled  = (known after apply)",
                                "      + qos_policy_id          = (known after apply)",
                                "      + region                 = (known after apply)",
                                "      + security_group_ids     = (known after apply)",
                                "      + tenant_id              = (known after apply)",
                                "",
                                "      + allowed_address_pairs {",
                                "          + ip_address = \"192.168.16.254/32\"",
                                "        }",
                                "      + allowed_address_pairs {",
                                "          + ip_address = \"192.168.16.8/32\"",
                                "        }",
                                "      + allowed_address_pairs {",
                                "          + ip_address = \"192.168.16.9/32\"",
                                "        }",
                                "",
                                "      + binding (known after apply)",
                                "",
                                "      + fixed_ip {",
                                "          + ip_address = \"192.168.16.14\"",
                                "          + subnet_id  = (known after apply)",
                                "        }",
                                "    }",
                                "",
                                "  # openstack_networking_port_v2.node_port_management[5] will be created",
                                "  + resource \"openstack_networking_port_v2\" \"node_port_management\" {",
                                "      + admin_state_up         = (known after apply)",
                                "      + all_fixed_ips          = (known after apply)",
                                "      + all_security_group_ids = (known after apply)",
                                "      + all_tags               = (known after apply)",
                                "      + device_id              = (known after apply)",
                                "      + device_owner           = (known after apply)",
                                "      + dns_assignment         = (known after apply)",
                                "      + dns_name               = (known after apply)",
                                "      + id                     = (known after apply)",
                                "      + mac_address            = (known after apply)",
                                "      + network_id             = (known after apply)",
                                "      + port_security_enabled  = (known after apply)",
                                "      + qos_policy_id          = (known after apply)",
                                "      + region                 = (known after apply)",
                                "      + security_group_ids     = (known after apply)",
                                "      + tenant_id              = (known after apply)",
                                "",
                                "      + allowed_address_pairs {",
                                "          + ip_address = \"192.168.16.254/32\"",
                                "        }",
                                "      + allowed_address_pairs {",
                                "          + ip_address = \"192.168.16.8/32\"",
                                "        }",
                                "      + allowed_address_pairs {",
                                "          + ip_address = \"192.168.16.9/32\"",
                                "        }",
                                "",
                                "      + binding (known after apply)",
                                "",
                                "      + fixed_ip {",
                                "          + ip_address = \"192.168.16.15\"",
                                "          + subnet_id  = (known after apply)",
                                "        }",
                                "    }",
                                "",
                                "  # openstack_networking_router_interface_v2.router_interface will be created",
                                "  + resource \"openstack_networking_router_interface_v2\" \"router_interface\" {",
                                "      + force_destroy = false",
                                "      + id            = (known after apply)",
                                "      + port_id       = (known after apply)",
                                "      + region        = (known after apply)",
                                "      + router_id     = (known after apply)",
                                "      + subnet_id     = (known after apply)",
                                "    }",
                                "",
                                "  # openstack_networking_router_v2.router will be created",
                                "  + resource \"openstack_networking_router_v2\" \"router\" {",
                                "      + admin_state_up          = (known after apply)",
                                "      + all_tags                = (known after apply)",
                                "      + availability_zone_hints = [",
                                "          + \"nova\",",
                                "        ]",
                                "      + distributed             = (known after apply)",
                                "      + enable_snat             = (known after apply)",
                                "      + external_network_id     = \"e6be7364-bfd8-4de7-8120-8f41c69a139a\"",
                                "      + external_qos_policy_id  = (known after apply)",
                                "      + id                      = (known after apply)",
                                "      + name                    = \"testbed\"",
                                "      + region                  = (known after apply)",
                                "      + tenant_id               = (known after apply)",
                                "",
                                "      + external_fixed_ip (known after apply)",
                                "    }",
                                "",
                                "  # openstack_networking_secgroup_rule_v2.security_group_management_rule1 will be created",
                                "  + resource \"openstack_networking_secgroup_rule_v2\" \"security_group_management_rule1\" {",
                                "      + description             = \"ssh\"",
                                "      + direction               = \"ingress\"",
                                "      + ethertype               = \"IPv4\"",
                                "      + id                      = (known after apply)",
                                "      + port_range_max          = 22",
                                "      + port_range_min          = 22",
                                "      + protocol                = \"tcp\"",
                                "      + region                  = (known after apply)",
                                "      + remote_address_group_id = (known after apply)",
                                "      + remote_group_id         = (known after apply)",
                                "      + remote_ip_prefix        = \"0.0.0.0/0\"",
                                "      + security_group_id       = (known after apply)",
                                "      + tenant_id               = (known after apply)",
                                "    }",
                                "",
                                "  # openstack_networking_secgroup_rule_v2.security_group_management_rule2 will be created",
                                "  + resource \"openstack_networking_secgroup_rule_v2\" \"security_group_management_rule2\" {",
                                "      + description             = \"wireguard\"",
                                "      + direction               = \"ingress\"",
                                "      + ethertype               = \"IPv4\"",
                                "      + id                      = (known after apply)",
                                "      + port_range_max          = 51820",
                                "      + port_range_min          = 51820",
                                "      + protocol                = \"udp\"",
                                "      + region                  = (known after apply)",
                                "      + remote_address_group_id = (known after apply)",
                                "      + remote_group_id         = (known after apply)",
                                "      + remote_ip_prefix        = \"0.0.0.0/0\"",
                                "      + security_group_id       = (known after apply)",
                                "      + tenant_id               = (known after apply)",
                                "    }",
                                "",
                                "  # openstack_networking_secgroup_rule_v2.security_group_management_rule3 will be created",
                                "  + resource \"openstack_networking_secgroup_rule_v2\" \"security_group_management_rule3\" {",
                                "      + direction               = \"ingress\"",
                                "      + ethertype               = \"IPv4\"",
                                "      + id                      = (known after apply)",
                                "      + protocol                = \"tcp\"",
                                "      + region                  = (known after apply)",
                                "      + remote_address_group_id = (known after apply)",
                                "      + remote_group_id         = (known after apply)",
                                "      + remote_ip_prefix        = \"192.168.16.0/20\"",
                                "      + security_group_id       = (known after apply)",
                                "      + tenant_id               = (known after apply)",
                                "    }",
                                "",
                                "  # openstack_networking_secgroup_rule_v2.security_group_management_rule4 will be created",
                                "  + resource \"openstack_networking_secgroup_rule_v2\" \"security_group_management_rule4\" {",
                                "      + direction               = \"ingress\"",
                                "      + ethertype               = \"IPv4\"",
                                "      + id                      = (known after apply)",
                                "      + protocol                = \"udp\"",
                                "      + region                  = (known after apply)",
                                "      + remote_address_group_id = (known after apply)",
                                "      + remote_group_id         = (known after apply)",
                                "      + remote_ip_prefix        = \"192.168.16.0/20\"",
                                "      + security_group_id       = (known after apply)",
                                "      + tenant_id               = (known after apply)",
                                "    }",
                                "",
                                "  # openstack_networking_secgroup_rule_v2.security_group_management_rule5 will be created",
                                "  + resource \"openstack_networking_secgroup_rule_v2\" \"security_group_management_rule5\" {",
                                "      + direction               = \"ingress\"",
                                "      + ethertype               = \"IPv4\"",
                                "      + id                      = (known after apply)",
                                "      + protocol                = \"icmp\"",
                                "      + region                  = (known after apply)",
                                "      + remote_address_group_id = (known after apply)",
                                "      + remote_group_id         = (known after apply)",
                                "      + remote_ip_prefix        = \"0.0.0.0/0\"",
                                "      + security_group_id       = (known after apply)",
                                "      + tenant_id               = (known after apply)",
                                "    }",
                                "",
                                "  # openstack_networking_secgroup_rule_v2.security_group_node_rule1 will be created",
                                "  + resource \"openstack_networking_secgroup_rule_v2\" \"security_group_node_rule1\" {",
                                "      + direction               = \"ingress\"",
                                "      + ethertype               = \"IPv4\"",
                                "      + id                      = (known after apply)",
                                "      + protocol                = \"tcp\"",
                                "      + region                  = (known after apply)",
                                "      + remote_address_group_id = (known after apply)",
                                "      + remote_group_id         = (known after apply)",
                                "      + remote_ip_prefix        = \"0.0.0.0/0\"",
                                "      + security_group_id       = (known after apply)",
                                "      + tenant_id               = (known after apply)",
                                "    }",
                                "",
                                "  # openstack_networking_secgroup_rule_v2.security_group_node_rule2 will be created",
                                "  + resource \"openstack_networking_secgroup_rule_v2\" \"security_group_node_rule2\" {",
                                "      + direction               = \"ingress\"",
                                "      + ethertype               = \"IPv4\"",
                                "      + id                      = (known after apply)",
                                "      + protocol                = \"udp\"",
                                "      + region                  = (known after apply)",
                                "      + remote_address_group_id = (known after apply)",
                                "      + remote_group_id         = (known after apply)",
                                "      + remote_ip_prefix        = \"0.0.0.0/0\"",
                                "      + security_group_id       = (known after apply)",
                                "      + tenant_id               = (known after apply)",
                                "    }",
                                "",
                                "  # openstack_networking_secgroup_rule_v2.security_group_node_rule3 will be created",
                                "  + resource \"openstack_networking_secgroup_rule_v2\" \"security_group_node_rule3\" {",
                                "      + direction               = \"ingress\"",
                                "      + ethertype               = \"IPv4\"",
                                "      + id                      = (known after apply)",
                                "      + protocol                = \"icmp\"",
                                "      + region                  = (known after apply)",
                                "      + remote_address_group_id = (known after apply)",
                                "      + remote_group_id         = (known after apply)",
                                "      + remote_ip_prefix        = \"0.0.0.0/0\"",
                                "      + security_group_id       = (known after apply)",
                                "      + tenant_id               = (known after apply)",
                                "    }",
                                "",
                                "  # openstack_networking_secgroup_rule_v2.security_group_rule_vrrp will be created",
                                "  + resource \"openstack_networking_secgroup_rule_v2\" \"security_group_rule_vrrp\" {",
                                "      + description             = \"vrrp\"",
                                "      + direction               = \"ingress\"",
                                "      + ethertype               = \"IPv4\"",
                                "      + id                      = (known after apply)",
                                "      + protocol                = \"112\"",
                                "      + region                  = (known after apply)",
                                "      + remote_address_group_id = (known after apply)",
                                "      + remote_group_id         = (known after apply)",
                                "      + remote_ip_prefix        = \"0.0.0.0/0\"",
                                "      + security_group_id       = (known after apply)",
                                "      + tenant_id               = (known after apply)",
                                "    }",
                                "",
                                "  # openstack_networking_secgroup_v2.security_group_management will be created",
                                "  + resource \"openstack_networking_secgroup_v2\" \"security_group_management\" {",
                                "      + all_tags    = (known after apply)",
                                "      + description = \"management security group\"",
                                "      + id          = (known after apply)",
                                "      + name        = \"testbed-management\"",
                                "      + region      = (known after apply)",
                                "      + stateful    = (known after apply)",
                                "      + tenant_id   = (known after apply)",
                                "    }",
                                "",
                                "  # openstack_networking_secgroup_v2.security_group_node will be created",
                                "  + resource \"openstack_networking_secgroup_v2\" \"security_group_node\" {",
                                "      + all_tags    = (known after apply)",
                                "      + description = \"node security group\"",
                                "      + id          = (known after apply)",
                                "      + name        = \"testbed-node\"",
                                "      + region      = (known after apply)",
                                "      + stateful    = (known after apply)",
                                "      + tenant_id   = (known after apply)",
                                "    }",
                                "",
                                "  # openstack_networking_subnet_v2.subnet_management will be created",
                                "  + resource \"openstack_networking_subnet_v2\" \"subnet_management\" {",
                                "      + all_tags          = (known after apply)",
                                "      + cidr              = \"192.168.16.0/20\"",
                                "      + dns_nameservers   = [",
                                "          + \"8.8.8.8\",",
                                "          + \"9.9.9.9\",",
                                "        ]",
                                "      + enable_dhcp       = true",
                                "      + gateway_ip        = (known after apply)",
                                "      + id                = (known after apply)",
                                "      + ip_version        = 4",
                                "      + ipv6_address_mode = (known after apply)",
                                "      + ipv6_ra_mode      = (known after apply)",
                                "      + name              = \"subnet-testbed-management\"",
                                "      + network_id        = (known after apply)",
                                "      + no_gateway        = false",
                                "      + region            = (known after apply)",
                                "      + service_types     = (known after apply)",
                                "      + tenant_id         = (known after apply)",
                                "",
                                "      + allocation_pool {",
                                "          + end   = \"192.168.31.250\"",
                                "          + start = \"192.168.31.200\"",
                                "        }",
                                "    }",
                                "",
                                "  # terraform_data.image will be created",
                                "  + resource \"terraform_data\" \"image\" {",
                                "      + id     = (known after apply)",
                                "      + input  = \"Ubuntu 24.04\"",
                                "      + output = (known after apply)",
                                "    }",
                                "",
                                "  # terraform_data.image_node will be created",
                                "  + resource \"terraform_data\" \"image_node\" {",
                                "      + id     = (known after apply)",
                                "      + input  = \"Ubuntu 24.04\"",
                                "      + output = (known after apply)",
                                "    }",
                                "",
                                "Plan: 64 to add, 0 to change, 0 to destroy.",
                                "",
                                "Changes to Outputs:",
                                "  + manager_address = (sensitive value)",
                                "  + private_key     = (sensitive value)",
                                "terraform_data.image_node: Creating...",
                                "terraform_data.image: Creating...",
                                "terraform_data.image_node: Creation complete after 0s [id=15076263-b98f-8a1e-6728-793c2a492a4f]",
                                "terraform_data.image: Creation complete after 0s [id=529e8ade-efc7-96ff-a4ad-79bfc4eed7eb]",
                                "data.openstack_images_image_v2.image: Reading...",
                                "data.openstack_images_image_v2.image_node: Reading...",
                                "openstack_blockstorage_volume_v3.node_volume[8]: Creating...",
                                "openstack_blockstorage_volume_v3.node_volume[3]: Creating...",
                                "openstack_blockstorage_volume_v3.node_volume[1]: Creating...",
                                "openstack_blockstorage_volume_v3.node_volume[4]: Creating...",
                                "openstack_blockstorage_volume_v3.node_volume[2]: Creating...",
                                "openstack_blockstorage_volume_v3.node_volume[0]: Creating...",
                                "openstack_compute_keypair_v2.key: Creating...",
                                "openstack_networking_network_v2.net_management: Creating...",
                                "data.openstack_images_image_v2.image: Read complete after 1s [id=846820b2-039e-4b42-adad-daf72e0f8ea4]",
                                "openstack_blockstorage_volume_v3.node_volume[6]: Creating...",
                                "openstack_compute_keypair_v2.key: Creation complete after 1s [id=testbed]",
                                "data.openstack_images_image_v2.image_node: Read complete after 1s [id=846820b2-039e-4b42-adad-daf72e0f8ea4]",
                                "openstack_blockstorage_volume_v3.node_volume[7]: Creating...",
                                "openstack_blockstorage_volume_v3.node_volume[5]: Creating...",
                                "openstack_networking_network_v2.net_management: Creation complete after 1s [id=efb372d3-0e12-4094-ab09-ea954e656478]",
                                "openstack_blockstorage_volume_v3.manager_base_volume[0]: Creating...",
                                "openstack_blockstorage_volume_v3.node_volume[2]: Creation complete after 4s [id=6f4643f0-329b-4ae2-97c7-d348648fd970]",
                                "local_sensitive_file.id_rsa: Creating...",
                                "local_sensitive_file.id_rsa: Creation complete after 0s [id=d4355e1f232bca0bfbf7b62060f5199acdefa892]",
                                "openstack_blockstorage_volume_v3.node_volume[0]: Creation complete after 4s [id=9c6b6763-9704-41ec-9b6d-97eb80a5145c]",
                                "local_file.id_rsa_pub: Creating...",
                                "openstack_blockstorage_volume_v3.node_volume[8]: Creation complete after 4s [id=5f47789f-2a1a-442a-8a3c-644a4f92e5a8]",
                                "openstack_blockstorage_volume_v3.node_volume[4]: Creation complete after 4s [id=fb58bf16-a1c6-4abb-925b-29c6a6a6f100]",
                                "openstack_blockstorage_volume_v3.node_base_volume[1]: Creating...",
                                "openstack_networking_subnet_v2.subnet_management: Creating...",
                                "openstack_blockstorage_volume_v3.node_volume[1]: Creation complete after 4s [id=5660bb62-e2c9-45b2-ae11-8ab0f829b194]",
                                "local_file.id_rsa_pub: Creation complete after 0s [id=d54a83a615742cbaaac15d6ca6f05a14c6d0c5ab]",
                                "openstack_blockstorage_volume_v3.node_base_volume[0]: Creating...",
                                "openstack_blockstorage_volume_v3.node_base_volume[4]: Creating...",
                                "openstack_blockstorage_volume_v3.node_base_volume[2]: Creating...",
                                "openstack_blockstorage_volume_v3.node_volume[3]: Creation complete after 4s [id=6644f191-69e9-4f3a-baff-cffe4152595b]",
                                "openstack_blockstorage_volume_v3.node_base_volume[3]: Creating...",
                                "openstack_blockstorage_volume_v3.node_volume[6]: Creation complete after 3s [id=4534650b-23a0-46c8-8bf1-44520caee637]",
                                "openstack_blockstorage_volume_v3.node_volume[7]: Creation complete after 3s [id=f659b61f-c39f-426f-a640-6a753951661d]",
                                "openstack_blockstorage_volume_v3.node_base_volume[5]: Creating...",
                                "openstack_blockstorage_volume_v3.node_volume[5]: Creation complete after 3s [id=1558935b-4c66-4614-97e5-cda14abd2641]",
                                "openstack_blockstorage_volume_v3.manager_base_volume[0]: Creation complete after 4s [id=65266dda-9f19-479d-a42c-1592584b970f]",
                                "openstack_networking_subnet_v2.subnet_management: Creation complete after 1s [id=ee6d269f-6556-408a-a662-603f9f99eb43]",
                                "openstack_networking_router_v2.router: Creating...",
                                "openstack_blockstorage_volume_v3.node_base_volume[1]: Creation complete after 3s [id=6fef90f9-4ce5-456f-a170-152290f19313]",
                                "openstack_blockstorage_volume_v3.node_base_volume[0]: Creation complete after 3s [id=2d259cae-7b79-479e-b978-3fffee75aa51]",
                                "openstack_blockstorage_volume_v3.node_base_volume[4]: Creation complete after 3s [id=47e600ec-05ca-44a6-93d0-5ca3104e3639]",
                                "openstack_blockstorage_volume_v3.node_base_volume[2]: Creation complete after 4s [id=35b8d51e-716e-455d-9a95-995756ae5900]",
                                "openstack_blockstorage_volume_v3.node_base_volume[3]: Creation complete after 4s [id=f90a4e80-1f36-4d14-b7da-de7b20e9fbca]",
                                "openstack_blockstorage_volume_v3.node_base_volume[5]: Creation complete after 4s [id=7a9487cb-ec9e-4508-9c5a-f692700fc5a4]",
                                "openstack_networking_router_v2.router: Creation complete after 3s [id=78bdffa7-7274-45bc-a222-3c8067a8b67a]",
                                "openstack_networking_secgroup_v2.security_group_management: Creating...",
                                "openstack_networking_secgroup_v2.security_group_node: Creating...",
                                "openstack_networking_router_interface_v2.router_interface: Creating...",
                                "openstack_networking_secgroup_v2.security_group_node: Creation complete after 0s [id=0e28c216-6c91-4be2-bd8a-e93830fffbeb]",
                                "openstack_networking_secgroup_rule_v2.security_group_node_rule3: Creating...",
                                "openstack_networking_secgroup_rule_v2.security_group_node_rule2: Creating...",
                                "openstack_networking_port_v2.node_port_management[5]: Creating...",
                                "openstack_networking_secgroup_rule_v2.security_group_node_rule1: Creating...",
                                "openstack_networking_secgroup_rule_v2.security_group_rule_vrrp: Creating...",
                                "openstack_networking_port_v2.node_port_management[3]: Creating...",
                                "openstack_networking_port_v2.node_port_management[1]: Creating...",
                                "openstack_networking_port_v2.node_port_management[0]: Creating...",
                                "openstack_networking_secgroup_rule_v2.security_group_node_rule3: Creation complete after 0s [id=387089e4-f760-4928-88f5-a915100e55d6]",
                                "openstack_networking_port_v2.node_port_management[2]: Creating...",
                                "openstack_networking_secgroup_rule_v2.security_group_node_rule2: Creation complete after 0s [id=b73046cc-2cee-4e90-9574-a7fc0413601b]",
                                "openstack_networking_port_v2.node_port_management[4]: Creating...",
                                "openstack_networking_secgroup_rule_v2.security_group_rule_vrrp: Creation complete after 0s [id=1d63a497-952f-4017-a681-67dbf229b9f0]",
                                "openstack_networking_port_v2.node_port_management[1]: Creation complete after 1s [id=4e0612ba-f5fc-449f-bdf0-52cf3f9e0ff4]",
                                "openstack_networking_port_v2.node_port_management[5]: Creation complete after 1s [id=b963b7df-f3b4-4efc-a2f3-fc1a7ac620e0]",
                                "openstack_networking_port_v2.node_port_management[0]: Creation complete after 1s [id=52a531cd-5463-4412-a687-49e3322e1d3a]",
                                "openstack_networking_secgroup_rule_v2.security_group_node_rule1: Creation complete after 1s [id=ab75e10b-5a7d-4465-a77b-a2d27ac85a87]",
                                "openstack_networking_secgroup_v2.security_group_management: Creation complete after 1s [id=4576d5df-b4e8-453a-9f85-e9033140afde]",
                                "openstack_networking_secgroup_rule_v2.security_group_management_rule5: Creating...",
                                "openstack_networking_secgroup_rule_v2.security_group_management_rule1: Creating...",
                                "openstack_networking_port_v2.manager_port_management: Creating...",
                                "openstack_networking_secgroup_rule_v2.security_group_management_rule3: Creating...",
                                "openstack_networking_secgroup_rule_v2.security_group_management_rule2: Creating...",
                                "openstack_networking_secgroup_rule_v2.security_group_management_rule4: Creating...",
                                "openstack_networking_port_v2.node_port_management[2]: Creation complete after 1s [id=3efcac55-5642-4046-a9a6-94f0970dacfb]",
                                "openstack_networking_secgroup_rule_v2.security_group_management_rule5: Creation complete after 0s [id=791e61d3-0ebb-4c2f-b01f-4a99d7a46541]",
                                "openstack_networking_port_v2.node_port_management[4]: Creation complete after 1s [id=fd0bab54-e5d7-4115-8ac6-6d5e24c663da]",
                                "openstack_networking_port_v2.node_port_management[3]: Creation complete after 1s [id=b86f1b60-f237-4c8b-a846-4b44402e937f]",
                                "openstack_networking_secgroup_rule_v2.security_group_management_rule1: Creation complete after 0s [id=f8964514-2650-45f3-9372-ee021595be3f]",
                                "openstack_networking_secgroup_rule_v2.security_group_management_rule3: Creation complete after 0s [id=4dd1cb5f-5d60-43d2-bcd1-f648f83f847e]",
                                "openstack_networking_secgroup_rule_v2.security_group_management_rule2: Creation complete after 0s [id=c8b4c78c-fd27-4b8c-98e4-38b549c80b55]",
                                "openstack_networking_secgroup_rule_v2.security_group_management_rule4: Creation complete after 1s [id=e6ef18d5-506a-44c4-be35-c43969468703]",
                                "openstack_networking_port_v2.manager_port_management: Creation complete after 1s [id=6419139a-cf37-4920-a5ce-6fbaf77a4267]",
                                "openstack_networking_router_interface_v2.router_interface: Creation complete after 4s [id=e6f7b8e7-2ad3-4d1c-a6b4-399b8439e436]",
                                "openstack_networking_floatingip_v2.manager_floating_ip: Creating...",
                                "openstack_compute_instance_v2.node_server[1]: Creating...",
                                "openstack_compute_instance_v2.node_server[0]: Creating...",
                                "openstack_compute_instance_v2.node_server[5]: Creating...",
                                "openstack_compute_instance_v2.node_server[4]: Creating...",
                                "openstack_compute_instance_v2.node_server[2]: Creating...",
                                "openstack_compute_instance_v2.node_server[3]: Creating...",
                                "openstack_networking_floatingip_v2.manager_floating_ip: Creation complete after 3s [id=4b75e324-3592-41b1-8c39-d8b21170c67a]",
                                "local_file.MANAGER_ADDRESS: Creating...",
                                "local_file.inventory: Creating...",
                                "openstack_networking_floatingip_associate_v2.manager_floating_ip_association: Creating...",
                                "local_file.inventory: Creation complete after 0s [id=5cef27c57f3cca0ffaf63b133cf6c70a2ab823cb]",
                                "local_file.MANAGER_ADDRESS: Creation complete after 0s [id=2a52d35ad08ef77d309e2f503983de36521d64f8]",
                                "openstack_networking_floatingip_associate_v2.manager_floating_ip_association: Creation complete after 1s [id=4b75e324-3592-41b1-8c39-d8b21170c67a]",
                                "openstack_compute_instance_v2.node_server[1]: Still creating... [10s elapsed]",
                                "openstack_compute_instance_v2.node_server[0]: Still creating... [10s elapsed]",
                                "openstack_compute_instance_v2.node_server[5]: Still creating... [10s elapsed]",
                                "openstack_compute_instance_v2.node_server[4]: Still creating... [10s elapsed]",
                                "openstack_compute_instance_v2.node_server[2]: Still creating... [10s elapsed]",
                                "openstack_compute_instance_v2.node_server[3]: Still creating... [10s elapsed]",
                                "openstack_compute_instance_v2.node_server[0]: Still creating... [20s elapsed]",
                                "openstack_compute_instance_v2.node_server[1]: Still creating... [20s elapsed]",
                                "openstack_compute_instance_v2.node_server[5]: Still creating... [20s elapsed]",
                                "openstack_compute_instance_v2.node_server[4]: Still creating... [20s elapsed]",
                                "openstack_compute_instance_v2.node_server[2]: Still creating... [20s elapsed]",
                                "openstack_compute_instance_v2.node_server[3]: Still creating... [20s elapsed]",
                                "openstack_compute_instance_v2.node_server[3]: Creation complete after 21s [id=05d0fb61-2d0a-4090-b8f0-bd104996720e]",
                                "openstack_compute_instance_v2.node_server[0]: Creation complete after 21s [id=aaa02988-9362-4770-8564-8f17ae7318b2]",
                                "openstack_compute_instance_v2.node_server[1]: Still creating... [30s elapsed]",
                                "openstack_compute_instance_v2.node_server[5]: Still creating... [30s elapsed]",
                                "openstack_compute_instance_v2.node_server[4]: Still creating... [30s elapsed]",
                                "openstack_compute_instance_v2.node_server[2]: Still creating... [30s elapsed]",
                                "openstack_compute_instance_v2.node_server[5]: Creation complete after 31s [id=f7128b88-877f-46fc-9d1d-3f4c1596413d]",
                                "openstack_compute_instance_v2.node_server[2]: Creation complete after 31s [id=84c66398-f55f-4354-b3cf-775d88768d57]",
                                "openstack_compute_instance_v2.node_server[1]: Creation complete after 31s [id=45422329-13ac-46a3-a475-2d2bacc66157]",
                                "openstack_compute_instance_v2.node_server[4]: Creation complete after 31s [id=e22977cc-084b-4b2a-bbb9-fad0560981cd]",
                                "null_resource.node_semaphore: Creating...",
                                "null_resource.node_semaphore: Creation complete after 0s [id=45525888549353150]",
                                "openstack_compute_volume_attach_v2.node_volume_attachment[4]: Creating...",
                                "openstack_compute_volume_attach_v2.node_volume_attachment[1]: Creating...",
                                "openstack_compute_volume_attach_v2.node_volume_attachment[7]: Creating...",
                                "openstack_compute_volume_attach_v2.node_volume_attachment[6]: Creating...",
                                "openstack_compute_volume_attach_v2.node_volume_attachment[3]: Creating...",
                                "openstack_compute_volume_attach_v2.node_volume_attachment[8]: Creating...",
                                "openstack_compute_volume_attach_v2.node_volume_attachment[2]: Creating...",
                                "openstack_compute_volume_attach_v2.node_volume_attachment[5]: Creating...",
                                "openstack_compute_instance_v2.manager_server: Creating...",
                                "openstack_compute_volume_attach_v2.node_volume_attachment[0]: Creating...",
                                "openstack_compute_volume_attach_v2.node_volume_attachment[6]: Creation complete after 4s [id=05d0fb61-2d0a-4090-b8f0-bd104996720e/4534650b-23a0-46c8-8bf1-44520caee637]",
                                "openstack_compute_volume_attach_v2.node_volume_attachment[1]: Creation complete after 4s [id=e22977cc-084b-4b2a-bbb9-fad0560981cd/5660bb62-e2c9-45b2-ae11-8ab0f829b194]",
                                "openstack_compute_volume_attach_v2.node_volume_attachment[8]: Creation complete after 4s [id=f7128b88-877f-46fc-9d1d-3f4c1596413d/5f47789f-2a1a-442a-8a3c-644a4f92e5a8]",
                                "openstack_compute_volume_attach_v2.node_volume_attachment[4]: Creation complete after 10s [id=e22977cc-084b-4b2a-bbb9-fad0560981cd/fb58bf16-a1c6-4abb-925b-29c6a6a6f100]",
                                "openstack_compute_volume_attach_v2.node_volume_attachment[3]: Creation complete after 10s [id=05d0fb61-2d0a-4090-b8f0-bd104996720e/6644f191-69e9-4f3a-baff-cffe4152595b]",
                                "openstack_compute_volume_attach_v2.node_volume_attachment[2]: Creation complete after 10s [id=f7128b88-877f-46fc-9d1d-3f4c1596413d/6f4643f0-329b-4ae2-97c7-d348648fd970]",
                                "openstack_compute_volume_attach_v2.node_volume_attachment[0]: Creation complete after 10s [id=05d0fb61-2d0a-4090-b8f0-bd104996720e/9c6b6763-9704-41ec-9b6d-97eb80a5145c]",
                                "openstack_compute_volume_attach_v2.node_volume_attachment[7]: Creation complete after 10s [id=e22977cc-084b-4b2a-bbb9-fad0560981cd/f659b61f-c39f-426f-a640-6a753951661d]",
                                "openstack_compute_volume_attach_v2.node_volume_attachment[5]: Creation complete after 10s [id=f7128b88-877f-46fc-9d1d-3f4c1596413d/1558935b-4c66-4614-97e5-cda14abd2641]",
                                "openstack_compute_instance_v2.manager_server: Still creating... [10s elapsed]",
                                "openstack_compute_instance_v2.manager_server: Still creating... [20s elapsed]",
                                "openstack_compute_instance_v2.manager_server: Creation complete after 21s [id=17506fa8-3f0b-4351-b562-9644c43fa35e]",
                                "",
                                "Apply complete! Resources: 64 added, 0 changed, 0 destroyed.",
                                "",
                                "Outputs:",
                                "",
                                "manager_address = <sensitive>",
                                "private_key = <sensitive>"
                            ],
                            "zuul_log_id": "0242ac1f-6606-1b2b-a63a-000000000009-1-orchestrator"
                        }
                    },
                    "task": {
                        "duration": {
                            "end": "2026-02-24T00:02:32.028881Z",
                            "start": "2026-02-24T00:01:21.929460Z"
                        },
                        "id": "0242ac1f-6606-1b2b-a63a-000000000009",
                        "name": "Create infrastructure (latest)"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "ansible.builtin.command",
                            "changed": false,
                            "false_condition": "_manager_version != \"latest\"",
                            "skip_reason": "Conditional result was False",
                            "skipped": true
                        }
                    },
                    "task": {
                        "duration": {
                            "end": "2026-02-24T00:02:32.582693Z",
                            "start": "2026-02-24T00:02:32.071670Z"
                        },
                        "id": "0242ac1f-6606-1b2b-a63a-00000000000a",
                        "name": "Create infrastructure (stable)"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "ansible.builtin.slurp",
                            "changed": false,
                            "content": "TUFOQUdFUl9BRERSRVNTPTgxLjE2My4xOTIuMTk1Cg==",
                            "encoding": "base64",
                            "invocation": {
                                "module_args": {
                                    "src": "/home/zuul-testbed02/src/github.com/osism/testbed/terraform/.MANAGER_ADDRESS.ci"
                                }
                            },
                            "source": "/home/zuul-testbed02/src/github.com/osism/testbed/terraform/.MANAGER_ADDRESS.ci"
                        }
                    },
                    "task": {
                        "duration": {
                            "end": "2026-02-24T00:02:33.094056Z",
                            "start": "2026-02-24T00:02:32.601410Z"
                        },
                        "id": "0242ac1f-6606-1b2b-a63a-00000000000b",
                        "name": "Fetch manager address"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "ansible.builtin.set_fact",
                            "ansible_facts": {
                                "manager_host": "81.163.192.195"
                            },
                            "changed": false
                        }
                    },
                    "task": {
                        "duration": {
                            "end": "2026-02-24T00:02:33.176977Z",
                            "start": "2026-02-24T00:02:33.101404Z"
                        },
                        "id": "0242ac1f-6606-1b2b-a63a-00000000000c",
                        "name": "Set manager_host address"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "ansible.builtin.command",
                            "changed": true,
                            "msg": "All items completed",
                            "results": [
                                {
                                    "ansible_loop_var": "item",
                                    "changed": true,
                                    "cmd": [
                                        "~/venv/bin/ansible-galaxy",
                                        "collection",
                                        "install",
                                        "--force",
                                        "/home/zuul-testbed02/src/github.com/osism/ansible-collection-commons"
                                    ],
                                    "delta": "0:00:00.692699",
                                    "end": "2026-02-24 00:02:34.174158",
                                    "failed": false,
                                    "invocation": {
                                        "module_args": {
                                            "_raw_params": "~/venv/bin/ansible-galaxy collection install --force \"/home/zuul-testbed02/src/github.com/osism/ansible-collection-commons\"\n",
                                            "_uses_shell": false,
                                            "argv": null,
                                            "chdir": "/home/zuul-testbed02",
                                            "creates": null,
                                            "executable": null,
                                            "expand_argument_vars": true,
                                            "removes": null,
                                            "stdin": null,
                                            "stdin_add_newline": true,
                                            "strip_empty_ends": true,
                                            "zuul_ansible_split_streams": false,
                                            "zuul_log_id": "in-loop-ignore",
                                            "zuul_output_max_bytes": 1073741824
                                        }
                                    },
                                    "item": "commons",
                                    "msg": "",
                                    "rc": 0,
                                    "start": "2026-02-24 00:02:33.481459",
                                    "stderr": "",
                                    "stderr_lines": [],
                                    "stdout": "[WARNING]: Collection osism.services does not support Ansible version 2.15.2\n[WARNING]: Collection osism.commons does not support Ansible version 2.15.2\nStarting galaxy collection install process\nProcess install dependency map\nStarting collection install process\nInstalling 'osism.commons:999.0.0' to '/home/zuul-testbed02/.ansible/collections/ansible_collections/osism/commons'\nCreated collection for osism.commons:999.0.0 at /home/zuul-testbed02/.ansible/collections/ansible_collections/osism/commons\nosism.commons:999.0.0 was installed successfully",
                                    "stdout_lines": [
                                        "[WARNING]: Collection osism.services does not support Ansible version 2.15.2",
                                        "[WARNING]: Collection osism.commons does not support Ansible version 2.15.2",
                                        "Starting galaxy collection install process",
                                        "Process install dependency map",
                                        "Starting collection install process",
                                        "Installing 'osism.commons:999.0.0' to '/home/zuul-testbed02/.ansible/collections/ansible_collections/osism/commons'",
                                        "Created collection for osism.commons:999.0.0 at /home/zuul-testbed02/.ansible/collections/ansible_collections/osism/commons",
                                        "osism.commons:999.0.0 was installed successfully"
                                    ],
                                    "zuul_log_id": "in-loop-ignore"
                                },
                                {
                                    "ansible_loop_var": "item",
                                    "changed": true,
                                    "cmd": [
                                        "~/venv/bin/ansible-galaxy",
                                        "collection",
                                        "install",
                                        "--force",
                                        "/home/zuul-testbed02/src/github.com/osism/ansible-collection-services"
                                    ],
                                    "delta": "0:00:00.740117",
                                    "end": "2026-02-24 00:02:35.231165",
                                    "failed": false,
                                    "invocation": {
                                        "module_args": {
                                            "_raw_params": "~/venv/bin/ansible-galaxy collection install --force \"/home/zuul-testbed02/src/github.com/osism/ansible-collection-services\"\n",
                                            "_uses_shell": false,
                                            "argv": null,
                                            "chdir": "/home/zuul-testbed02",
                                            "creates": null,
                                            "executable": null,
                                            "expand_argument_vars": true,
                                            "removes": null,
                                            "stdin": null,
                                            "stdin_add_newline": true,
                                            "strip_empty_ends": true,
                                            "zuul_ansible_split_streams": false,
                                            "zuul_log_id": "in-loop-ignore",
                                            "zuul_output_max_bytes": 1073741824
                                        }
                                    },
                                    "item": "services",
                                    "msg": "",
                                    "rc": 0,
                                    "start": "2026-02-24 00:02:34.491048",
                                    "stderr": "",
                                    "stderr_lines": [],
                                    "stdout": "[WARNING]: Collection osism.commons does not support Ansible version 2.15.2\n[WARNING]: Collection osism.services does not support Ansible version 2.15.2\nStarting galaxy collection install process\nProcess install dependency map\nStarting collection install process\nInstalling 'osism.services:999.0.0' to '/home/zuul-testbed02/.ansible/collections/ansible_collections/osism/services'\nCreated collection for osism.services:999.0.0 at /home/zuul-testbed02/.ansible/collections/ansible_collections/osism/services\nosism.services:999.0.0 was installed successfully",
                                    "stdout_lines": [
                                        "[WARNING]: Collection osism.commons does not support Ansible version 2.15.2",
                                        "[WARNING]: Collection osism.services does not support Ansible version 2.15.2",
                                        "Starting galaxy collection install process",
                                        "Process install dependency map",
                                        "Starting collection install process",
                                        "Installing 'osism.services:999.0.0' to '/home/zuul-testbed02/.ansible/collections/ansible_collections/osism/services'",
                                        "Created collection for osism.services:999.0.0 at /home/zuul-testbed02/.ansible/collections/ansible_collections/osism/services",
                                        "osism.services:999.0.0 was installed successfully"
                                    ],
                                    "zuul_log_id": "in-loop-ignore"
                                }
                            ]
                        }
                    },
                    "task": {
                        "duration": {
                            "end": "2026-02-24T00:02:35.287816Z",
                            "start": "2026-02-24T00:02:33.186509Z"
                        },
                        "id": "0242ac1f-6606-1b2b-a63a-00000000000d",
                        "name": "Update ansible collections"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "ansible.builtin.wait_for",
                            "changed": false,
                            "elapsed": 10,
                            "invocation": {
                                "module_args": {
                                    "active_connection_states": [
                                        "ESTABLISHED",
                                        "FIN_WAIT1",
                                        "FIN_WAIT2",
                                        "SYN_RECV",
                                        "SYN_SENT",
                                        "TIME_WAIT"
                                    ],
                                    "connect_timeout": 5,
                                    "delay": 10,
                                    "exclude_hosts": null,
                                    "host": "81.163.192.195",
                                    "msg": null,
                                    "path": null,
                                    "port": 22,
                                    "search_regex": "OpenSSH",
                                    "sleep": 1,
                                    "state": "started",
                                    "timeout": 300
                                }
                            },
                            "match_groupdict": {},
                            "match_groups": [],
                            "path": null,
                            "port": 22,
                            "search_regex": "OpenSSH",
                            "state": "started"
                        }
                    },
                    "task": {
                        "duration": {
                            "end": "2026-02-24T00:02:45.979522Z",
                            "start": "2026-02-24T00:02:35.300780Z"
                        },
                        "id": "0242ac1f-6606-1b2b-a63a-00000000000e",
                        "name": "Wait up to 300 seconds for port 22 to become open and contain \"OpenSSH\""
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "ansible.builtin.pause",
                            "changed": false,
                            "delta": 60,
                            "echo": true,
                            "rc": 0,
                            "start": "2026-02-24 00:02:46.019157",
                            "stderr": "",
                            "stdout": "Paused for 60.0 seconds",
                            "stop": "2026-02-24 00:03:46.021011",
                            "user_input": ""
                        }
                    },
                    "task": {
                        "duration": {
                            "end": "2026-02-24T00:03:46.050713Z",
                            "start": "2026-02-24T00:02:45.995813Z"
                        },
                        "id": "0242ac1f-6606-1b2b-a63a-00000000000f",
                        "name": "Wait a little longer for the manager so that everything is ready"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "ansible.builtin.shell",
                            "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result",
                            "changed": true
                        }
                    },
                    "task": {
                        "duration": {
                            "end": "2026-02-24T00:03:47.658227Z",
                            "start": "2026-02-24T00:03:46.138781Z"
                        },
                        "id": "0242ac1f-6606-1b2b-a63a-000000000010",
                        "name": "Fetch manager ssh hostkey"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "ansible.builtin.shell",
                            "changed": true,
                            "cmd": "cp /home/zuul-testbed02/src/github.com/osism/testbed/terraform/.id* .",
                            "delta": "0:00:00.007781",
                            "end": "2026-02-24 00:03:48.032355",
                            "invocation": {
                                "module_args": {
                                    "_raw_params": "cp /home/zuul-testbed02/src/github.com/osism/testbed/terraform/.id* .",
                                    "_uses_shell": true,
                                    "argv": null,
                                    "chdir": "/home/zuul-testbed02/src/github.com/osism/testbed/ansible",
                                    "creates": null,
                                    "executable": null,
                                    "expand_argument_vars": true,
                                    "removes": null,
                                    "stdin": null,
                                    "stdin_add_newline": true,
                                    "strip_empty_ends": true,
                                    "zuul_ansible_split_streams": false,
                                    "zuul_log_id": "0242ac1f-6606-1b2b-a63a-000000000011-1-orchestrator",
                                    "zuul_output_max_bytes": 1073741824
                                }
                            },
                            "msg": "",
                            "rc": 0,
                            "start": "2026-02-24 00:03:48.024574",
                            "stderr": "",
                            "stderr_lines": [],
                            "stdout": "",
                            "stdout_lines": [],
                            "zuul_log_id": "0242ac1f-6606-1b2b-a63a-000000000011-1-orchestrator"
                        }
                    },
                    "task": {
                        "duration": {
                            "end": "2026-02-24T00:03:48.226482Z",
                            "start": "2026-02-24T00:03:47.716199Z"
                        },
                        "id": "0242ac1f-6606-1b2b-a63a-000000000011",
                        "name": "Get ssh keypair from terraform environment"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "ansible.builtin.debug",
                            "changed": false,
                            "msg": "The task 'Run manager part 0' runs an Ansible playbook on the manager. There is no further output of this here. It takes a few minutes for this task to complete."
                        }
                    },
                    "task": {
                        "duration": {
                            "end": "2026-02-24T00:03:48.293740Z",
                            "start": "2026-02-24T00:03:48.242136Z"
                        },
                        "id": "0242ac1f-6606-1b2b-a63a-000000000012",
                        "name": "Point out that the following task takes some time and does not give any output"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "ansible.builtin.shell",
                            "changed": true,
                            "cmd": "set -o pipefail\n~/venv/bin/ansible-playbook \\\n  -i /home/zuul-testbed02/src/github.com/osism/testbed/terraform/inventory.ci \\\n  --key-file /home/zuul-testbed02/src/github.com/osism/testbed/terraform/.id_rsa.ci \\\n  -e cloud=ci \\\n  -e terraform_environment=ci-ubuntu-24.04 \\\n  -e repo_path=/home/zuul-testbed02/src/github.com \\\n  manager-part-0.yml | tee -a ansible-manager-part-0.log\n",
                            "delta": "0:09:36.695834",
                            "end": "2026-02-24 00:13:25.354599",
                            "invocation": {
                                "module_args": {
                                    "_raw_params": "set -o pipefail\n~/venv/bin/ansible-playbook \\\n  -i /home/zuul-testbed02/src/github.com/osism/testbed/terraform/inventory.ci \\\n  --key-file /home/zuul-testbed02/src/github.com/osism/testbed/terraform/.id_rsa.ci \\\n  -e cloud=ci \\\n  -e terraform_environment=ci-ubuntu-24.04 \\\n  -e repo_path=/home/zuul-testbed02/src/github.com \\\n  manager-part-0.yml | tee -a ansible-manager-part-0.log\n",
                                    "_uses_shell": true,
                                    "argv": null,
                                    "chdir": "/home/zuul-testbed02/src/github.com/osism/testbed/ansible",
                                    "creates": null,
                                    "executable": "/bin/bash",
                                    "expand_argument_vars": true,
                                    "removes": null,
                                    "stdin": null,
                                    "stdin_add_newline": true,
                                    "strip_empty_ends": true,
                                    "zuul_ansible_split_streams": false,
                                    "zuul_log_id": "0242ac1f-6606-1b2b-a63a-000000000013-1-orchestrator",
                                    "zuul_output_max_bytes": 1073741824
                                }
                            },
                            "msg": "",
                            "rc": 0,
                            "start": "2026-02-24 00:03:48.658765",
                            "stderr": "",
                            "stderr_lines": [],
                            "stdout": "[WARNING]: Collection osism.commons does not support Ansible version 2.15.2\n\nPLAY [Wait for cloud-init to finish] *******************************************\n\nTASK [Check /var/lib/cloud/instance/boot-finished] *****************************\nok: [testbed-manager]\n\nPLAY [Run manager part 0] ******************************************************\n\nTASK [Gathering Facts] *********************************************************\nok: [testbed-manager]\n\nTASK [Get home directory of ansible user] **************************************\nok: [testbed-manager]\n\nTASK [Set repo_path fact] ******************************************************\nskipping: [testbed-manager]\n\nTASK [Update package cache] ****************************************************\nskipping: [testbed-manager]\n\nTASK [Install required packages] ***********************************************\nskipping: [testbed-manager]\n\nTASK [Remove some python packages] *********************************************\nskipping: [testbed-manager]\n\nTASK [Set venv_command fact (RedHat)] ******************************************\nskipping: [testbed-manager]\n\nTASK [Fail if Ubuntu version is lower than 24.04] ******************************\nskipping: [testbed-manager]\n\nTASK [Fail if Debian version is lower than 12] *********************************\nskipping: [testbed-manager]\n\nTASK [Set APT options on manager] **********************************************\nchanged: [testbed-manager]\n\nTASK [Update APT cache and run dist-upgrade] ***********************************\nchanged: [testbed-manager]\n\nTASK [Install HWE kernel package on Ubuntu] ************************************\nchanged: [testbed-manager]\n\nTASK [Install required packages] ***********************************************\nchanged: [testbed-manager]\n\nTASK [Remove some python packages] *********************************************\nchanged: [testbed-manager]\n\nTASK [Set venv_command fact (Debian)] ******************************************\nok: [testbed-manager]\n\nTASK [Get current user] ********************************************************\nok: [testbed-manager]\n\nTASK [Create venv directory] ***************************************************\nchanged: [testbed-manager]\n\nTASK [Install netaddr in venv] *************************************************\nchanged: [testbed-manager]\n\nTASK [Install ansible-core in venv] ********************************************\nchanged: [testbed-manager]\n\nTASK [Install requests >= 2.32.2] **********************************************\nchanged: [testbed-manager]\n\nTASK [Install docker >= 7.1.0] *************************************************\nchanged: [testbed-manager]\n\nTASK [Create directories in /opt/src] ******************************************\nchanged: [testbed-manager] => (item=osism/ansible-collection-commons)\nchanged: [testbed-manager] => (item=osism/ansible-collection-services)\n\nTASK [Sync sources in /opt/src] ************************************************\n[DEPRECATION WARNING]: The connection's stdin object is deprecated. Call \ndisplay.prompt_until(msg) instead. This feature will be removed in version \n2.19. Deprecation warnings can be disabled by setting \ndeprecation_warnings=False in ansible.cfg.\nchanged: [testbed-manager] => (item=osism/ansible-collection-commons)\nchanged: [testbed-manager] => (item=osism/ansible-collection-services)\n\nTASK [Create /usr/share/ansible directory] *************************************\nchanged: [testbed-manager]\n\nTASK [Install collections from Ansible galaxy] *********************************\nchanged: [testbed-manager] => (item=ansible.netcommon)\nchanged: [testbed-manager] => (item=ansible.posix)\nchanged: [testbed-manager] => (item=community.docker>=3.10.2)\n\nTASK [Install local collections] ***********************************************\nchanged: [testbed-manager] => (item=ansible-collection-commons)\nchanged: [testbed-manager] => (item=ansible-collection-services)\n\nPLAY [Create operator user] ****************************************************\n\nTASK [Gathering Facts] *********************************************************\nok: [testbed-manager]\n\nTASK [osism.commons.operator : Gather variables for each operating system] *****\nok: [testbed-manager]\n\nTASK [osism.commons.operator : Set operator_groups variable to default value] ***\nok: [testbed-manager]\n\nTASK [osism.commons.operator : Create operator group] **************************\nchanged: [testbed-manager]\n\nTASK [osism.commons.operator : Create user] ************************************\nchanged: [testbed-manager]\n\nTASK [osism.commons.operator : Add user to additional groups] ******************\nchanged: [testbed-manager] => (item=adm)\nchanged: [testbed-manager] => (item=sudo)\n\nTASK [osism.commons.operator : Copy user sudoers file] *************************\nchanged: [testbed-manager]\n\nTASK [osism.commons.operator : Set language variables in .bashrc configuration file] ***\nchanged: [testbed-manager] => (item=export LANGUAGE=C.UTF-8)\nchanged: [testbed-manager] => (item=export LANG=C.UTF-8)\nchanged: [testbed-manager] => (item=export LC_ALL=C.UTF-8)\n\nTASK [osism.commons.operator : Set custom environment variables in .bashrc configuration file] ***\nskipping: [testbed-manager]\n\nTASK [osism.commons.operator : Set custom PS1 prompt in .bashrc configuration file] ***\nskipping: [testbed-manager]\n\nTASK [osism.commons.operator : Create .ssh directory] **************************\nchanged: [testbed-manager]\n\nTASK [osism.commons.operator : Check number of SSH authorized keys] ************\nskipping: [testbed-manager]\n\nTASK [osism.commons.operator : Set ssh authorized keys] ************************\nchanged: [testbed-manager] => (item=None)\nchanged: [testbed-manager]\n\nTASK [osism.commons.operator : Delete ssh authorized keys] *********************\nskipping: [testbed-manager]\n\nTASK [osism.commons.operator : Set authorized GitHub accounts] *****************\nskipping: [testbed-manager]\n\nTASK [osism.commons.operator : Delete authorized GitHub accounts] **************\nskipping: [testbed-manager]\n\nTASK [osism.commons.operator : Set password] ***********************************\nskipping: [testbed-manager]\n\nTASK [osism.commons.operator : Unset & lock password] **************************\nok: [testbed-manager]\n\nPLAY [Run manager part 0] ******************************************************\n\nTASK [Gathering Facts] *********************************************************\nok: [testbed-manager]\n\nTASK [Recursively change ownership of /opt/venv] *******************************\nchanged: [testbed-manager]\n\nPLAY RECAP *********************************************************************\ntestbed-manager            : ok=33   changed=23   unreachable=0    failed=0    skipped=14   rescued=0    ignored=0   ",
                            "stdout_lines": [
                                "[WARNING]: Collection osism.commons does not support Ansible version 2.15.2",
                                "",
                                "PLAY [Wait for cloud-init to finish] *******************************************",
                                "",
                                "TASK [Check /var/lib/cloud/instance/boot-finished] *****************************",
                                "ok: [testbed-manager]",
                                "",
                                "PLAY [Run manager part 0] ******************************************************",
                                "",
                                "TASK [Gathering Facts] *********************************************************",
                                "ok: [testbed-manager]",
                                "",
                                "TASK [Get home directory of ansible user] **************************************",
                                "ok: [testbed-manager]",
                                "",
                                "TASK [Set repo_path fact] ******************************************************",
                                "skipping: [testbed-manager]",
                                "",
                                "TASK [Update package cache] ****************************************************",
                                "skipping: [testbed-manager]",
                                "",
                                "TASK [Install required packages] ***********************************************",
                                "skipping: [testbed-manager]",
                                "",
                                "TASK [Remove some python packages] *********************************************",
                                "skipping: [testbed-manager]",
                                "",
                                "TASK [Set venv_command fact (RedHat)] ******************************************",
                                "skipping: [testbed-manager]",
                                "",
                                "TASK [Fail if Ubuntu version is lower than 24.04] ******************************",
                                "skipping: [testbed-manager]",
                                "",
                                "TASK [Fail if Debian version is lower than 12] *********************************",
                                "skipping: [testbed-manager]",
                                "",
                                "TASK [Set APT options on manager] **********************************************",
                                "changed: [testbed-manager]",
                                "",
                                "TASK [Update APT cache and run dist-upgrade] ***********************************",
                                "changed: [testbed-manager]",
                                "",
                                "TASK [Install HWE kernel package on Ubuntu] ************************************",
                                "changed: [testbed-manager]",
                                "",
                                "TASK [Install required packages] ***********************************************",
                                "changed: [testbed-manager]",
                                "",
                                "TASK [Remove some python packages] *********************************************",
                                "changed: [testbed-manager]",
                                "",
                                "TASK [Set venv_command fact (Debian)] ******************************************",
                                "ok: [testbed-manager]",
                                "",
                                "TASK [Get current user] ********************************************************",
                                "ok: [testbed-manager]",
                                "",
                                "TASK [Create venv directory] ***************************************************",
                                "changed: [testbed-manager]",
                                "",
                                "TASK [Install netaddr in venv] *************************************************",
                                "changed: [testbed-manager]",
                                "",
                                "TASK [Install ansible-core in venv] ********************************************",
                                "changed: [testbed-manager]",
                                "",
                                "TASK [Install requests >= 2.32.2] **********************************************",
                                "changed: [testbed-manager]",
                                "",
                                "TASK [Install docker >= 7.1.0] *************************************************",
                                "changed: [testbed-manager]",
                                "",
                                "TASK [Create directories in /opt/src] ******************************************",
                                "changed: [testbed-manager] => (item=osism/ansible-collection-commons)",
                                "changed: [testbed-manager] => (item=osism/ansible-collection-services)",
                                "",
                                "TASK [Sync sources in /opt/src] ************************************************",
                                "[DEPRECATION WARNING]: The connection's stdin object is deprecated. Call ",
                                "display.prompt_until(msg) instead. This feature will be removed in version ",
                                "2.19. Deprecation warnings can be disabled by setting ",
                                "deprecation_warnings=False in ansible.cfg.",
                                "changed: [testbed-manager] => (item=osism/ansible-collection-commons)",
                                "changed: [testbed-manager] => (item=osism/ansible-collection-services)",
                                "",
                                "TASK [Create /usr/share/ansible directory] *************************************",
                                "changed: [testbed-manager]",
                                "",
                                "TASK [Install collections from Ansible galaxy] *********************************",
                                "changed: [testbed-manager] => (item=ansible.netcommon)",
                                "changed: [testbed-manager] => (item=ansible.posix)",
                                "changed: [testbed-manager] => (item=community.docker>=3.10.2)",
                                "",
                                "TASK [Install local collections] ***********************************************",
                                "changed: [testbed-manager] => (item=ansible-collection-commons)",
                                "changed: [testbed-manager] => (item=ansible-collection-services)",
                                "",
                                "PLAY [Create operator user] ****************************************************",
                                "",
                                "TASK [Gathering Facts] *********************************************************",
                                "ok: [testbed-manager]",
                                "",
                                "TASK [osism.commons.operator : Gather variables for each operating system] *****",
                                "ok: [testbed-manager]",
                                "",
                                "TASK [osism.commons.operator : Set operator_groups variable to default value] ***",
                                "ok: [testbed-manager]",
                                "",
                                "TASK [osism.commons.operator : Create operator group] **************************",
                                "changed: [testbed-manager]",
                                "",
                                "TASK [osism.commons.operator : Create user] ************************************",
                                "changed: [testbed-manager]",
                                "",
                                "TASK [osism.commons.operator : Add user to additional groups] ******************",
                                "changed: [testbed-manager] => (item=adm)",
                                "changed: [testbed-manager] => (item=sudo)",
                                "",
                                "TASK [osism.commons.operator : Copy user sudoers file] *************************",
                                "changed: [testbed-manager]",
                                "",
                                "TASK [osism.commons.operator : Set language variables in .bashrc configuration file] ***",
                                "changed: [testbed-manager] => (item=export LANGUAGE=C.UTF-8)",
                                "changed: [testbed-manager] => (item=export LANG=C.UTF-8)",
                                "changed: [testbed-manager] => (item=export LC_ALL=C.UTF-8)",
                                "",
                                "TASK [osism.commons.operator : Set custom environment variables in .bashrc configuration file] ***",
                                "skipping: [testbed-manager]",
                                "",
                                "TASK [osism.commons.operator : Set custom PS1 prompt in .bashrc configuration file] ***",
                                "skipping: [testbed-manager]",
                                "",
                                "TASK [osism.commons.operator : Create .ssh directory] **************************",
                                "changed: [testbed-manager]",
                                "",
                                "TASK [osism.commons.operator : Check number of SSH authorized keys] ************",
                                "skipping: [testbed-manager]",
                                "",
                                "TASK [osism.commons.operator : Set ssh authorized keys] ************************",
                                "changed: [testbed-manager] => (item=None)",
                                "changed: [testbed-manager]",
                                "",
                                "TASK [osism.commons.operator : Delete ssh authorized keys] *********************",
                                "skipping: [testbed-manager]",
                                "",
                                "TASK [osism.commons.operator : Set authorized GitHub accounts] *****************",
                                "skipping: [testbed-manager]",
                                "",
                                "TASK [osism.commons.operator : Delete authorized GitHub accounts] **************",
                                "skipping: [testbed-manager]",
                                "",
                                "TASK [osism.commons.operator : Set password] ***********************************",
                                "skipping: [testbed-manager]",
                                "",
                                "TASK [osism.commons.operator : Unset & lock password] **************************",
                                "ok: [testbed-manager]",
                                "",
                                "PLAY [Run manager part 0] ******************************************************",
                                "",
                                "TASK [Gathering Facts] *********************************************************",
                                "ok: [testbed-manager]",
                                "",
                                "TASK [Recursively change ownership of /opt/venv] *******************************",
                                "changed: [testbed-manager]",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "testbed-manager            : ok=33   changed=23   unreachable=0    failed=0    skipped=14   rescued=0    ignored=0   "
                            ],
                            "zuul_log_id": "0242ac1f-6606-1b2b-a63a-000000000013-1-orchestrator"
                        }
                    },
                    "task": {
                        "duration": {
                            "end": "2026-02-24T00:13:25.743570Z",
                            "start": "2026-02-24T00:03:48.332445Z"
                        },
                        "id": "0242ac1f-6606-1b2b-a63a-000000000013",
                        "name": "Run manager part 0"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "ansible.builtin.debug",
                            "changed": false,
                            "msg": "It is now already possible to log in to the manager with 'make login'."
                        }
                    },
                    "task": {
                        "duration": {
                            "end": "2026-02-24T00:13:25.800897Z",
                            "start": "2026-02-24T00:13:25.752048Z"
                        },
                        "id": "0242ac1f-6606-1b2b-a63a-000000000014",
                        "name": "Point out that the log in on the manager is now possible"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "ansible.builtin.debug",
                            "changed": false,
                            "msg": "The task 'Run manager part 1 + 2' runs an Ansible playbook on the manager. There is no further output of this here. It takes a few minuts for this task to complete."
                        }
                    },
                    "task": {
                        "duration": {
                            "end": "2026-02-24T00:13:25.853033Z",
                            "start": "2026-02-24T00:13:25.811480Z"
                        },
                        "id": "0242ac1f-6606-1b2b-a63a-000000000015",
                        "name": "Point out that the following task takes some time and does not give any output"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "ansible.builtin.shell",
                            "changed": true,
                            "cmd": "set -o pipefail\n~/venv/bin/ansible-playbook \\\n  -i /home/zuul-testbed02/src/github.com/osism/testbed/terraform/inventory.ci \\\n  --key-file /home/zuul-testbed02/src/github.com/osism/testbed/terraform/.id_rsa.ci \\\n  -e cloud=ci \\\n  -e terraform_environment=ci-ubuntu-24.04 \\\n  -e repo_path=/home/zuul-testbed02/src/github.com \\\n  -e manager_version=latest \\\n  manager-part-1.yml | tee -a ansible-manager-part-1.log\n",
                            "delta": "0:02:32.636728",
                            "end": "2026-02-24 00:15:58.850998",
                            "invocation": {
                                "module_args": {
                                    "_raw_params": "set -o pipefail\n~/venv/bin/ansible-playbook \\\n  -i /home/zuul-testbed02/src/github.com/osism/testbed/terraform/inventory.ci \\\n  --key-file /home/zuul-testbed02/src/github.com/osism/testbed/terraform/.id_rsa.ci \\\n  -e cloud=ci \\\n  -e terraform_environment=ci-ubuntu-24.04 \\\n  -e repo_path=/home/zuul-testbed02/src/github.com \\\n  -e manager_version=latest \\\n  manager-part-1.yml | tee -a ansible-manager-part-1.log\n",
                                    "_uses_shell": true,
                                    "argv": null,
                                    "chdir": "/home/zuul-testbed02/src/github.com/osism/testbed/ansible",
                                    "creates": null,
                                    "executable": "/bin/bash",
                                    "expand_argument_vars": true,
                                    "removes": null,
                                    "stdin": null,
                                    "stdin_add_newline": true,
                                    "strip_empty_ends": true,
                                    "zuul_ansible_split_streams": false,
                                    "zuul_log_id": "0242ac1f-6606-1b2b-a63a-000000000016-1-orchestrator",
                                    "zuul_output_max_bytes": 1073741824
                                }
                            },
                            "msg": "",
                            "rc": 0,
                            "start": "2026-02-24 00:13:26.214270",
                            "stderr": "",
                            "stderr_lines": [],
                            "stdout": "[WARNING]: Collection osism.commons does not support Ansible version 2.15.2\n\nPLAY [Run manager part 1] ******************************************************\n\nTASK [Gathering Facts] *********************************************************\nok: [testbed-manager]\n\nTASK [Set venv_command fact (RedHat)] ******************************************\nskipping: [testbed-manager]\n\nTASK [Set venv_command fact (Debian)] ******************************************\nok: [testbed-manager]\n\nTASK [osism.commons.repository : Gather variables for each operating system] ***\nok: [testbed-manager]\n\nTASK [osism.commons.repository : Set repository_default fact to default value] ***\nok: [testbed-manager]\n\nTASK [osism.commons.repository : Set repositories to default] ******************\nok: [testbed-manager]\n\nTASK [osism.commons.repository : Include distribution specific repository tasks] ***\nincluded: /home/zuul-testbed02/.ansible/collections/ansible_collections/osism/commons/roles/repository/tasks/Ubuntu.yml for testbed-manager\n\nTASK [osism.commons.repository : Create /etc/apt/sources.list.d directory] *****\nok: [testbed-manager]\n\nTASK [osism.commons.repository : Include tasks for Ubuntu < 24.04] *************\nskipping: [testbed-manager]\n\nTASK [osism.commons.repository : Copy 99osism apt configuration] ***************\nchanged: [testbed-manager]\n\nTASK [osism.commons.repository : Remove sources.list file] *********************\nok: [testbed-manager]\n\nTASK [osism.commons.repository : Copy ubuntu.sources file] *********************\nchanged: [testbed-manager]\n\nTASK [osism.commons.repository : Update package cache] *************************\nchanged: [testbed-manager]\n\nTASK [Get home directory of ansible user] **************************************\nok: [testbed-manager]\n\nTASK [Set repo_path fact] ******************************************************\nskipping: [testbed-manager]\n\nTASK [Copy SSH public key] *****************************************************\nchanged: [testbed-manager]\n\nTASK [Copy SSH private key] ****************************************************\nchanged: [testbed-manager]\n\nTASK [Create configuration directory] ******************************************\nchanged: [testbed-manager]\n\nTASK [Copy testbed repo] *******************************************************\n[DEPRECATION WARNING]: The connection's stdin object is deprecated. Call \ndisplay.prompt_until(msg) instead. This feature will be removed in version \n2.19. Deprecation warnings can be disabled by setting \ndeprecation_warnings=False in ansible.cfg.\nchanged: [testbed-manager]\n\nTASK [Install python requirements in venv] *************************************\nok: [testbed-manager] => (item=Jinja2)\nok: [testbed-manager] => (item=PyYAML)\nok: [testbed-manager] => (item=packaging)\nchanged: [testbed-manager] => (item=python-gilt==1.2.3)\nok: [testbed-manager] => (item=requests>=2.32.2)\nok: [testbed-manager] => (item=docker>=7.1.0)\n\nTASK [Copy testbed custom CA certificate on Debian/Ubuntu] *********************\nchanged: [testbed-manager]\n\nTASK [Copy testbed custom CA certificate on CentOS] ****************************\nskipping: [testbed-manager]\n\nTASK [Run update-ca-certificates on Debian/Ubuntu] *****************************\nchanged: [testbed-manager]\n\nTASK [Run update-ca-trust on RedHat] *******************************************\nskipping: [testbed-manager]\n\nTASK [Run manager part 2] ******************************************************\nchanged: [testbed-manager]\n\nRUNNING HANDLER [osism.commons.repository : Force update of package cache] *****\nok: [testbed-manager]\n\nPLAY RECAP *********************************************************************\ntestbed-manager            : ok=21   changed=11   unreachable=0    failed=0    skipped=5    rescued=0    ignored=0   ",
                            "stdout_lines": [
                                "[WARNING]: Collection osism.commons does not support Ansible version 2.15.2",
                                "",
                                "PLAY [Run manager part 1] ******************************************************",
                                "",
                                "TASK [Gathering Facts] *********************************************************",
                                "ok: [testbed-manager]",
                                "",
                                "TASK [Set venv_command fact (RedHat)] ******************************************",
                                "skipping: [testbed-manager]",
                                "",
                                "TASK [Set venv_command fact (Debian)] ******************************************",
                                "ok: [testbed-manager]",
                                "",
                                "TASK [osism.commons.repository : Gather variables for each operating system] ***",
                                "ok: [testbed-manager]",
                                "",
                                "TASK [osism.commons.repository : Set repository_default fact to default value] ***",
                                "ok: [testbed-manager]",
                                "",
                                "TASK [osism.commons.repository : Set repositories to default] ******************",
                                "ok: [testbed-manager]",
                                "",
                                "TASK [osism.commons.repository : Include distribution specific repository tasks] ***",
                                "included: /home/zuul-testbed02/.ansible/collections/ansible_collections/osism/commons/roles/repository/tasks/Ubuntu.yml for testbed-manager",
                                "",
                                "TASK [osism.commons.repository : Create /etc/apt/sources.list.d directory] *****",
                                "ok: [testbed-manager]",
                                "",
                                "TASK [osism.commons.repository : Include tasks for Ubuntu < 24.04] *************",
                                "skipping: [testbed-manager]",
                                "",
                                "TASK [osism.commons.repository : Copy 99osism apt configuration] ***************",
                                "changed: [testbed-manager]",
                                "",
                                "TASK [osism.commons.repository : Remove sources.list file] *********************",
                                "ok: [testbed-manager]",
                                "",
                                "TASK [osism.commons.repository : Copy ubuntu.sources file] *********************",
                                "changed: [testbed-manager]",
                                "",
                                "TASK [osism.commons.repository : Update package cache] *************************",
                                "changed: [testbed-manager]",
                                "",
                                "TASK [Get home directory of ansible user] **************************************",
                                "ok: [testbed-manager]",
                                "",
                                "TASK [Set repo_path fact] ******************************************************",
                                "skipping: [testbed-manager]",
                                "",
                                "TASK [Copy SSH public key] *****************************************************",
                                "changed: [testbed-manager]",
                                "",
                                "TASK [Copy SSH private key] ****************************************************",
                                "changed: [testbed-manager]",
                                "",
                                "TASK [Create configuration directory] ******************************************",
                                "changed: [testbed-manager]",
                                "",
                                "TASK [Copy testbed repo] *******************************************************",
                                "[DEPRECATION WARNING]: The connection's stdin object is deprecated. Call ",
                                "display.prompt_until(msg) instead. This feature will be removed in version ",
                                "2.19. Deprecation warnings can be disabled by setting ",
                                "deprecation_warnings=False in ansible.cfg.",
                                "changed: [testbed-manager]",
                                "",
                                "TASK [Install python requirements in venv] *************************************",
                                "ok: [testbed-manager] => (item=Jinja2)",
                                "ok: [testbed-manager] => (item=PyYAML)",
                                "ok: [testbed-manager] => (item=packaging)",
                                "changed: [testbed-manager] => (item=python-gilt==1.2.3)",
                                "ok: [testbed-manager] => (item=requests>=2.32.2)",
                                "ok: [testbed-manager] => (item=docker>=7.1.0)",
                                "",
                                "TASK [Copy testbed custom CA certificate on Debian/Ubuntu] *********************",
                                "changed: [testbed-manager]",
                                "",
                                "TASK [Copy testbed custom CA certificate on CentOS] ****************************",
                                "skipping: [testbed-manager]",
                                "",
                                "TASK [Run update-ca-certificates on Debian/Ubuntu] *****************************",
                                "changed: [testbed-manager]",
                                "",
                                "TASK [Run update-ca-trust on RedHat] *******************************************",
                                "skipping: [testbed-manager]",
                                "",
                                "TASK [Run manager part 2] ******************************************************",
                                "changed: [testbed-manager]",
                                "",
                                "RUNNING HANDLER [osism.commons.repository : Force update of package cache] *****",
                                "ok: [testbed-manager]",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "testbed-manager            : ok=21   changed=11   unreachable=0    failed=0    skipped=5    rescued=0    ignored=0   "
                            ],
                            "zuul_log_id": "0242ac1f-6606-1b2b-a63a-000000000016-1-orchestrator"
                        }
                    },
                    "task": {
                        "duration": {
                            "end": "2026-02-24T00:15:59.020334Z",
                            "start": "2026-02-24T00:13:25.889357Z"
                        },
                        "id": "0242ac1f-6606-1b2b-a63a-000000000016",
                        "name": "Run manager part 1 + 2"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "ansible.builtin.command",
                            "changed": true,
                            "cmd": [
                                "ssh",
                                "-i",
                                "/home/zuul-testbed02/src/github.com/osism/testbed/terraform/.id_rsa.ci",
                                "dragon@81.163.192.195",
                                "sudo",
                                "shutdown",
                                "-r",
                                "now"
                            ],
                            "delta": "0:00:01.036884",
                            "end": "2026-02-24 00:16:00.363564",
                            "failed_when_result": false,
                            "invocation": {
                                "module_args": {
                                    "_raw_params": "ssh -i /home/zuul-testbed02/src/github.com/osism/testbed/terraform/.id_rsa.ci dragon@81.163.192.195 sudo shutdown -r now",
                                    "_uses_shell": false,
                                    "argv": null,
                                    "chdir": null,
                                    "creates": null,
                                    "executable": null,
                                    "expand_argument_vars": true,
                                    "removes": null,
                                    "stdin": null,
                                    "stdin_add_newline": true,
                                    "strip_empty_ends": true,
                                    "zuul_ansible_split_streams": false,
                                    "zuul_log_id": "0242ac1f-6606-1b2b-a63a-000000000017-1-orchestrator",
                                    "zuul_output_max_bytes": 1073741824
                                }
                            },
                            "msg": "",
                            "rc": 0,
                            "start": "2026-02-24 00:15:59.326680",
                            "stderr": "",
                            "stderr_lines": [],
                            "stdout": "",
                            "stdout_lines": [],
                            "zuul_log_id": "0242ac1f-6606-1b2b-a63a-000000000017-1-orchestrator"
                        }
                    },
                    "task": {
                        "duration": {
                            "end": "2026-02-24T00:16:00.565221Z",
                            "start": "2026-02-24T00:15:59.053818Z"
                        },
                        "id": "0242ac1f-6606-1b2b-a63a-000000000017",
                        "name": "Reboot manager"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "ansible.builtin.wait_for",
                            "changed": false,
                            "elapsed": 16,
                            "invocation": {
                                "module_args": {
                                    "active_connection_states": [
                                        "ESTABLISHED",
                                        "FIN_WAIT1",
                                        "FIN_WAIT2",
                                        "SYN_RECV",
                                        "SYN_SENT",
                                        "TIME_WAIT"
                                    ],
                                    "connect_timeout": 5,
                                    "delay": 10,
                                    "exclude_hosts": null,
                                    "host": "81.163.192.195",
                                    "msg": null,
                                    "path": null,
                                    "port": 22,
                                    "search_regex": "OpenSSH",
                                    "sleep": 1,
                                    "state": "started",
                                    "timeout": 300
                                }
                            },
                            "match_groupdict": {},
                            "match_groups": [],
                            "path": null,
                            "port": 22,
                            "search_regex": "OpenSSH",
                            "state": "started"
                        }
                    },
                    "task": {
                        "duration": {
                            "end": "2026-02-24T00:16:17.007473Z",
                            "start": "2026-02-24T00:16:00.581258Z"
                        },
                        "id": "0242ac1f-6606-1b2b-a63a-000000000018",
                        "name": "Wait up to 300 seconds for port 22 to become open and contain \"OpenSSH\""
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "ansible.builtin.pause",
                            "changed": false,
                            "delta": 60,
                            "echo": true,
                            "rc": 0,
                            "start": "2026-02-24 00:16:17.039969",
                            "stderr": "",
                            "stdout": "Paused for 60.0 seconds",
                            "stop": "2026-02-24 00:17:17.042945",
                            "user_input": ""
                        }
                    },
                    "task": {
                        "duration": {
                            "end": "2026-02-24T00:17:17.069521Z",
                            "start": "2026-02-24T00:16:17.017870Z"
                        },
                        "id": "0242ac1f-6606-1b2b-a63a-000000000019",
                        "name": "Wait a little longer for the manager so that everything is ready"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "ansible.builtin.command",
                            "changed": true,
                            "cmd": [
                                "ssh",
                                "-i",
                                "/home/zuul-testbed02/src/github.com/osism/testbed/terraform/.id_rsa.ci",
                                "dragon@81.163.192.195",
                                "/opt/configuration/scripts/deploy-manager.sh"
                            ],
                            "delta": "0:24:47.921878",
                            "end": "2026-02-24 00:42:05.728005",
                            "invocation": {
                                "module_args": {
                                    "_raw_params": "ssh -i /home/zuul-testbed02/src/github.com/osism/testbed/terraform/.id_rsa.ci dragon@81.163.192.195 /opt/configuration/scripts/deploy-manager.sh",
                                    "_uses_shell": false,
                                    "argv": null,
                                    "chdir": null,
                                    "creates": null,
                                    "executable": null,
                                    "expand_argument_vars": true,
                                    "removes": null,
                                    "stdin": null,
                                    "stdin_add_newline": true,
                                    "strip_empty_ends": true,
                                    "zuul_ansible_split_streams": false,
                                    "zuul_log_id": "0242ac1f-6606-1b2b-a63a-00000000001a-1-orchestrator",
                                    "zuul_output_max_bytes": 1073741824
                                }
                            },
                            "msg": "",
                            "rc": 0,
                            "start": "2026-02-24 00:17:17.806127",
                            "stderr": "",
                            "stderr_lines": [],
                            "stdout": "\n# DEPLOY MANAGER\n\n+ set -e\n+ echo\n+ echo '# DEPLOY MANAGER'\n+ echo\n+ cat /opt/manager-vars.sh\nexport NUMBER_OF_NODES=6\n\nexport CEPH_VERSION=reef\nexport CONFIGURATION_VERSION=main\nexport MANAGER_VERSION=latest\nexport OPENSTACK_VERSION=2024.2\n\nexport ARA=false\nexport DEPLOY_MODE=manager\nexport TEMPEST=true\nexport IS_ZUUL=true\n\nexport MANAGER_PUBLIC_IP_ADDRESS=81.163.192.195\nexport EXTERNAL_API=false\n\nexport IMAGE_USER=ubuntu\nexport IMAGE_NODE_USER=ubuntu\n\nexport CEPH_STACK=ceph-ansible\n\n+ echo\n+ source /opt/configuration/scripts/include.sh\n++ export INTERACTIVE=false\n++ INTERACTIVE=false\n++ export OSISM_APPLY_RETRY=1\n++ OSISM_APPLY_RETRY=1\n+ source /opt/manager-vars.sh\n++ export NUMBER_OF_NODES=6\n++ NUMBER_OF_NODES=6\n++ export CEPH_VERSION=reef\n++ CEPH_VERSION=reef\n++ export CONFIGURATION_VERSION=main\n++ CONFIGURATION_VERSION=main\n++ export MANAGER_VERSION=latest\n++ MANAGER_VERSION=latest\n++ export OPENSTACK_VERSION=2024.2\n++ OPENSTACK_VERSION=2024.2\n++ export ARA=false\n++ ARA=false\n++ export DEPLOY_MODE=manager\n++ DEPLOY_MODE=manager\n++ export TEMPEST=true\n++ TEMPEST=true\n++ export IS_ZUUL=true\n++ IS_ZUUL=true\n++ export MANAGER_PUBLIC_IP_ADDRESS=81.163.192.195\n++ MANAGER_PUBLIC_IP_ADDRESS=81.163.192.195\n++ export EXTERNAL_API=false\n++ EXTERNAL_API=false\n++ export IMAGE_USER=ubuntu\n++ IMAGE_USER=ubuntu\n++ export IMAGE_NODE_USER=ubuntu\n++ IMAGE_NODE_USER=ubuntu\n++ export CEPH_STACK=ceph-ansible\n++ CEPH_STACK=ceph-ansible\n+ sudo ln -sf /opt/configuration/contrib/semver2.sh /usr/local/bin/semver\n+ docker version\nClient: Docker Engine - Community\n Version:           27.5.1\n API version:       1.47\n Go version:        go1.22.11\n Git commit:        9f9e405\n Built:             Wed Jan 22 13:41:48 2025\n OS/Arch:           linux/amd64\n Context:           default\n\nServer: Docker Engine - Community\n Engine:\n  Version:          27.5.1\n  API version:      1.47 (minimum version 1.24)\n  Go version:       go1.22.11\n  Git commit:       4c9b3b0\n  Built:            Wed Jan 22 13:41:48 2025\n  OS/Arch:          linux/amd64\n  Experimental:     false\n containerd:\n  Version:          v2.2.1\n  GitCommit:        dea7da592f5d1d2b7755e3a161be07f43fad8f75\n runc:\n  Version:          1.3.4\n  GitCommit:        v1.3.4-0-gd6d73eb8\n docker-init:\n  Version:          0.19.0\n  GitCommit:        de40ad0\n+ sh -c /opt/configuration/scripts/deploy/000-manager.sh\n+ set -e\n+ source /opt/manager-vars.sh\n++ export NUMBER_OF_NODES=6\n++ NUMBER_OF_NODES=6\n++ export CEPH_VERSION=reef\n++ CEPH_VERSION=reef\n++ export CONFIGURATION_VERSION=main\n++ CONFIGURATION_VERSION=main\n++ export MANAGER_VERSION=latest\n++ MANAGER_VERSION=latest\n++ export OPENSTACK_VERSION=2024.2\n++ OPENSTACK_VERSION=2024.2\n++ export ARA=false\n++ ARA=false\n++ export DEPLOY_MODE=manager\n++ DEPLOY_MODE=manager\n++ export TEMPEST=true\n++ TEMPEST=true\n++ export IS_ZUUL=true\n++ IS_ZUUL=true\n++ export MANAGER_PUBLIC_IP_ADDRESS=81.163.192.195\n++ MANAGER_PUBLIC_IP_ADDRESS=81.163.192.195\n++ export EXTERNAL_API=false\n++ EXTERNAL_API=false\n++ export IMAGE_USER=ubuntu\n++ IMAGE_USER=ubuntu\n++ export IMAGE_NODE_USER=ubuntu\n++ IMAGE_NODE_USER=ubuntu\n++ export CEPH_STACK=ceph-ansible\n++ CEPH_STACK=ceph-ansible\n+ source /opt/configuration/scripts/include.sh\n++ export INTERACTIVE=false\n++ INTERACTIVE=false\n++ export OSISM_APPLY_RETRY=1\n++ OSISM_APPLY_RETRY=1\n+ [[ latest != \\l\\a\\t\\e\\s\\t ]]\n+ [[ latest == \\l\\a\\t\\e\\s\\t ]]\n+ /opt/configuration/scripts/set-ceph-version.sh reef\n+ set -e\n+ VERSION=reef\n++ grep '^ceph_version:' /opt/configuration/environments/manager/configuration.yml\n+ [[ -n ceph_version: reef ]]\n+ sed -i 's/ceph_version: .*/ceph_version: reef/g' /opt/configuration/environments/manager/configuration.yml\n+ /opt/configuration/scripts/set-openstack-version.sh 2024.2\n+ set -e\n+ VERSION=2024.2\n++ grep '^openstack_version:' /opt/configuration/environments/manager/configuration.yml\n+ [[ -n openstack_version: 2024.2 ]]\n+ sed -i 's/openstack_version: .*/openstack_version: 2024.2/g' /opt/configuration/environments/manager/configuration.yml\n+ [[ ceph-ansible == \\r\\o\\o\\k ]]\n++ semver latest 7.0.0\n+ [[ -1 -ge 0 ]]\n+ [[ latest == \\l\\a\\t\\e\\s\\t ]]\n+ echo 'enable_osism_kubernetes: true'\n++ semver latest 10.0.0-0\n+ [[ -1 -ge 0 ]]\n++ semver 2024.2 2025.1\n+ [[ -1 -ge 0 ]]\n+ /opt/configuration/scripts/enable-resource-nodes.sh\n+ [[ -e /opt/venv/bin/activate ]]\n+ source /opt/venv/bin/activate\n++ deactivate nondestructive\n++ '[' -n '' ']'\n++ '[' -n '' ']'\n++ hash -r\n++ '[' -n '' ']'\n++ unset VIRTUAL_ENV\n++ unset VIRTUAL_ENV_PROMPT\n++ '[' '!' nondestructive = nondestructive ']'\n++ '[' linux-gnu = cygwin ']'\n++ '[' linux-gnu = msys ']'\n++ export VIRTUAL_ENV=/opt/venv\n++ VIRTUAL_ENV=/opt/venv\n++ _OLD_VIRTUAL_PATH=/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/usr/games:/usr/local/games:/snap/bin\n++ PATH=/opt/venv/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/usr/games:/usr/local/games:/snap/bin\n++ export PATH\n++ '[' -n '' ']'\n++ '[' -z '' ']'\n++ _OLD_VIRTUAL_PS1=\n++ PS1='(venv) '\n++ export PS1\n++ VIRTUAL_ENV_PROMPT='(venv) '\n++ export VIRTUAL_ENV_PROMPT\n++ hash -r\n+ ansible-playbook -i testbed-manager, --vault-password-file /opt/configuration/environments/.vault_pass /opt/configuration/ansible/manager-part-3.yml\n\nPLAY [Copy custom facts] *******************************************************\n\nTASK [Create custom facts directory] *******************************************\nok: [testbed-manager]\n\nTASK [Copy fact files] *********************************************************\nchanged: [testbed-manager]\n\nPLAY [Before the deployment of the manager] ************************************\n\nTASK [Gathering Facts] *********************************************************\nok: [testbed-manager]\n\nTASK [Get /opt/manager-vars.sh] ************************************************\nok: [testbed-manager]\n\nTASK [Add ara_server_mariadb_volume_type parameter] ****************************\nchanged: [testbed-manager]\n\nTASK [Add netbox_enable parameter] *********************************************\nskipping: [testbed-manager]\n\nTASK [Install HWE kernel package on Ubuntu] ************************************\nchanged: [testbed-manager]\n\nTASK [Check if /etc/OTC_region exist] ******************************************\nok: [testbed-manager]\n\nTASK [Add nova_compute_virt_type parameter] ************************************\nskipping: [testbed-manager]\n\nPLAY [Apply role traefik] ******************************************************\n\nTASK [Gathering Facts] *********************************************************\nok: [testbed-manager]\n\nTASK [Apply traefik role] ******************************************************\nincluded: osism.services.traefik for testbed-manager\n\nTASK [osism.services.traefik : Include config tasks] ***************************\nincluded: /usr/share/ansible/collections/ansible_collections/osism/services/roles/traefik/tasks/config.yml for testbed-manager\n\nTASK [osism.services.traefik : Create required directories] ********************\nchanged: [testbed-manager] => (item=/opt/traefik)\nchanged: [testbed-manager] => (item=/opt/traefik/certificates)\nchanged: [testbed-manager] => (item=/opt/traefik/configuration)\n\nTASK [osism.services.traefik : Copy configuration files] ***********************\nchanged: [testbed-manager] => (item=traefik.yml)\nchanged: [testbed-manager] => (item=traefik.env)\nchanged: [testbed-manager] => (item=certificates.yml)\n\nTASK [osism.services.traefik : Copy certificate cert files] ********************\nchanged: [testbed-manager] => (item=None)\nchanged: [testbed-manager]\n\nTASK [osism.services.traefik : Copy certificate key files] *********************\nchanged: [testbed-manager] => (item=None)\nchanged: [testbed-manager]\n\nTASK [osism.services.traefik : Copy dynamic configuration] *********************\nskipping: [testbed-manager]\n\nTASK [osism.services.traefik : Remove dynamic configuration] *******************\nok: [testbed-manager]\n\nTASK [osism.services.traefik : Include service tasks] **************************\nincluded: /usr/share/ansible/collections/ansible_collections/osism/services/roles/traefik/tasks/service.yml for testbed-manager\n\nTASK [osism.services.traefik : Create traefik external network] ****************\nchanged: [testbed-manager]\n\nTASK [osism.services.traefik : Copy docker-compose.yml file] *******************\nchanged: [testbed-manager]\n\nTASK [osism.services.traefik : Manage traefik service] *************************\nchanged: [testbed-manager]\n\nRUNNING HANDLER [osism.services.traefik : Restart traefik service] *************\nskipping: [testbed-manager]\n\nPLAY [Deploy manager service] **************************************************\n\nTASK [Gathering Facts] *********************************************************\nok: [testbed-manager]\n\nTASK [Apply manager role] ******************************************************\nincluded: osism.services.manager for testbed-manager\n\nTASK [osism.services.manager : Include install tasks] **************************\nincluded: /usr/share/ansible/collections/ansible_collections/osism/services/roles/manager/tasks/install-Debian-family.yml for testbed-manager\n\nTASK [osism.services.manager : Install required packages] **********************\nok: [testbed-manager]\n\nTASK [osism.services.manager : Gather variables for each operating system] *****\nok: [testbed-manager]\n\nTASK [osism.services.manager : Include config tasks] ***************************\nincluded: /usr/share/ansible/collections/ansible_collections/osism/services/roles/manager/tasks/config.yml for testbed-manager\n\nTASK [osism.services.manager : Create required directories] ********************\nchanged: [testbed-manager] => (item=/opt/ansible)\nchanged: [testbed-manager] => (item=/opt/archive)\nchanged: [testbed-manager] => (item=/opt/manager/configuration)\nchanged: [testbed-manager] => (item=/opt/manager/data)\nok: [testbed-manager] => (item=/opt/manager)\nchanged: [testbed-manager] => (item=/opt/manager/secrets)\nchanged: [testbed-manager] => (item=/opt/ansible/secrets)\nchanged: [testbed-manager] => (item=/opt/state)\n\nTASK [osism.services.manager : Copy all environment file] **********************\nchanged: [testbed-manager]\n\nTASK [osism.services.manager : Copy client environment file] *******************\nchanged: [testbed-manager]\n\nTASK [osism.services.manager : Include ara config tasks] ***********************\nincluded: /usr/share/ansible/collections/ansible_collections/osism/services/roles/manager/tasks/config-ara.yml for testbed-manager\n\nTASK [osism.services.manager : Copy ARA environment files] *********************\nchanged: [testbed-manager] => (item=ara)\nchanged: [testbed-manager] => (item=ara-server)\n\nTASK [osism.services.manager : Copy MariaDB environment file] ******************\nchanged: [testbed-manager]\n\nTASK [osism.services.manager : Include vault config tasks] *********************\nskipping: [testbed-manager]\n\nTASK [osism.services.manager : Include frontend config tasks] ******************\nincluded: /usr/share/ansible/collections/ansible_collections/osism/services/roles/manager/tasks/config-frontend.yml for testbed-manager\n\nTASK [osism.services.manager : Copy frontend environment file] *****************\nchanged: [testbed-manager]\n\nTASK [osism.services.manager : Include ansible config tasks] *******************\nincluded: /usr/share/ansible/collections/ansible_collections/osism/services/roles/manager/tasks/config-ansible.yml for testbed-manager\n\nTASK [osism.services.manager : Copy private ssh keys] **************************\nchanged: [testbed-manager] => (item=None)\nchanged: [testbed-manager] => (item=None)\nchanged: [testbed-manager]\n\nTASK [osism.services.manager : Copy ansible environment file] ******************\nchanged: [testbed-manager]\n\nTASK [osism.services.manager : Include netbox config tasks] ********************\nskipping: [testbed-manager]\n\nTASK [osism.services.manager : Include celery config tasks] ********************\nincluded: /usr/share/ansible/collections/ansible_collections/osism/services/roles/manager/tasks/config-celery.yml for testbed-manager\n\nTASK [osism.services.manager : Set fs.inotify.max_user_watches] ****************\nchanged: [testbed-manager]\n\nTASK [osism.services.manager : Set fs.inotify.max_user_instances] **************\nchanged: [testbed-manager]\n\nTASK [osism.services.manager : Copy celery environment files] ******************\nchanged: [testbed-manager] => (item=conductor)\nchanged: [testbed-manager] => (item=openstack)\n\nTASK [osism.services.manager : Copy listener environment file] *****************\nchanged: [testbed-manager]\n\nTASK [osism.services.manager : Check for conductor.yml] ************************\nok: [testbed-manager]\n\nTASK [osism.services.manager : Copy conductor configuration file] **************\nchanged: [testbed-manager]\n\nTASK [osism.services.manager : Copy empty conductor configuration file] ********\nskipping: [testbed-manager]\n\nTASK [osism.services.manager : Include wrapper config tasks] *******************\nincluded: /usr/share/ansible/collections/ansible_collections/osism/services/roles/manager/tasks/config-wrapper.yml for testbed-manager\n\nTASK [osism.services.manager : Include wrapper vars file] **********************\nok: [testbed-manager]\n\nTASK [osism.services.manager : Copy wrapper scripts] ***************************\nchanged: [testbed-manager] => (item=osism)\nchanged: [testbed-manager] => (item=osism-update-docker)\nchanged: [testbed-manager] => (item=osism-update-manager)\n\nTASK [osism.services.manager : Copy cilium wrapper script] *********************\nchanged: [testbed-manager]\n\nTASK [osism.services.manager : Copy hubble wrapper script] *********************\nchanged: [testbed-manager]\n\nTASK [osism.services.manager : Copy flux wrapper script] ***********************\nchanged: [testbed-manager]\n\nTASK [osism.services.manager : Include scripts config tasks] *******************\nincluded: /usr/share/ansible/collections/ansible_collections/osism/services/roles/manager/tasks/config-scripts.yml for testbed-manager\n\nTASK [osism.services.manager : Include scripts vars file] **********************\nok: [testbed-manager]\n\nTASK [osism.services.manager : Copy scripts] ***********************************\nchanged: [testbed-manager] => (item=osism-include)\n\nTASK [osism.services.manager : Include service tasks] **************************\nincluded: /usr/share/ansible/collections/ansible_collections/osism/services/roles/manager/tasks/service.yml for testbed-manager\n\nTASK [osism.services.manager : Copy manager systemd unit file] *****************\nchanged: [testbed-manager]\n\nTASK [osism.services.manager : Create traefik external network] ****************\nok: [testbed-manager]\n\nTASK [osism.services.manager : Set mariadb healthcheck for mariadb < 11.0.0] ***\nskipping: [testbed-manager]\n\nTASK [osism.services.manager : Set mariadb healthcheck for mariadb >= 11.0.0] ***\nok: [testbed-manager]\n\nTASK [osism.services.manager : Copy docker-compose.yml file] *******************\nchanged: [testbed-manager]\n\nTASK [osism.services.manager : Pull container images] **************************\nchanged: [testbed-manager]\n\nTASK [osism.services.manager : Stop and disable old service docker-compose@manager] ***\nok: [testbed-manager]\n\nTASK [osism.services.manager : Do a manual start of the manager service] *******\nskipping: [testbed-manager]\n\nTASK [osism.services.manager : Manage manager service] *************************\nchanged: [testbed-manager]\n\nTASK [osism.services.manager : Register that manager service was started] ******\nok: [testbed-manager]\n\nTASK [osism.services.manager : Flush handlers] *********************************\n\nRUNNING HANDLER [osism.services.manager : Restart manager service] *************\nskipping: [testbed-manager]\n\nRUNNING HANDLER [osism.services.manager : Wait for manager service to start] ***\nPausing for 60 seconds\nchanged: [testbed-manager]\n\nRUNNING HANDLER [osism.services.manager : Ensure that all containers are up] ***\nchanged: [testbed-manager]\n\nRUNNING HANDLER [osism.services.manager : Wait for an healthy manager service] ***\nFAILED - RETRYING: [testbed-manager]: Wait for an healthy manager service (50 retries left).\nFAILED - RETRYING: [testbed-manager]: Wait for an healthy manager service (49 retries left).\nchanged: [testbed-manager]\n\nRUNNING HANDLER [osism.services.manager : Copy osismclient bash completion script] ***\nchanged: [testbed-manager]\n\nTASK [osism.services.manager : Include initialize tasks] ***********************\nincluded: /usr/share/ansible/collections/ansible_collections/osism/services/roles/manager/tasks/initialize.yml for testbed-manager\n\nTASK [osism.services.manager : Flush handlers] *********************************\n\nTASK [osism.services.manager : Include vault initialize tasks] *****************\nskipping: [testbed-manager]\n\nTASK [osism.services.manager : Include version verification tasks] *************\nincluded: /usr/share/ansible/collections/ansible_collections/osism/services/roles/manager/tasks/verify-versions.yml for testbed-manager\n\nTASK [osism.services.manager : Deploy service manager version check script] ****\nchanged: [testbed-manager]\n\nTASK [osism.services.manager : Execute service manager version check] **********\nok: [testbed-manager]\n\nTASK [osism.services.manager : Display version check results] ******************\nok: [testbed-manager] => {\n    \"version_check_result.stdout_lines\": [\n        \"=== OSISM Container Version Check ===\",\n        \"Checking running containers against expected versions...\",\n        \"\",\n        \"Checking service: inventory_reconciler (Inventory Reconciler Service)\",\n        \"  Expected: registry.osism.tech/osism/inventory-reconciler:latest\",\n        \"  Enabled:  true\",\n        \"  Running:  registry.osism.tech/osism/inventory-reconciler:latest\",\n        \"  Status:   \u2705 MATCH\",\n        \"\",\n        \"Checking service: osism-ansible (OSISM Ansible Service)\",\n        \"  Expected: registry.osism.tech/osism/osism-ansible:latest\",\n        \"  Enabled:  true\",\n        \"  Running:  registry.osism.tech/osism/osism-ansible:latest\",\n        \"  Status:   \u2705 MATCH\",\n        \"\",\n        \"Checking service: osism-kubernetes (Osism-Kubernetes Service)\",\n        \"  Expected: registry.osism.tech/osism/osism-kubernetes:latest\",\n        \"  Enabled:  true\",\n        \"  Running:  registry.osism.tech/osism/osism-kubernetes:latest\",\n        \"  Status:   \u2705 MATCH\",\n        \"\",\n        \"Checking service: ceph-ansible (Ceph-Ansible Service)\",\n        \"  Expected: registry.osism.tech/osism/ceph-ansible:reef\",\n        \"  Enabled:  true\",\n        \"  Running:  registry.osism.tech/osism/ceph-ansible:reef\",\n        \"  Status:   \u2705 MATCH\",\n        \"\",\n        \"Checking service: kolla-ansible (Kolla-Ansible Service)\",\n        \"  Expected: registry.osism.tech/osism/kolla-ansible:2024.2\",\n        \"  Enabled:  true\",\n        \"  Running:  registry.osism.tech/osism/kolla-ansible:2024.2\",\n        \"  Status:   \u2705 MATCH\",\n        \"\",\n        \"Checking service: osismclient (OSISM Client)\",\n        \"  Expected: registry.osism.tech/osism/osism:latest\",\n        \"  Enabled:  true\",\n        \"  Running:  registry.osism.tech/osism/osism:latest\",\n        \"  Status:   \u2705 MATCH\",\n        \"\",\n        \"Checking service: ara-server (ARA Server)\",\n        \"  Expected: registry.osism.tech/osism/ara-server:1.7.3\",\n        \"  Enabled:  true\",\n        \"  Running:  registry.osism.tech/osism/ara-server:1.7.3\",\n        \"  Status:   \u2705 MATCH\",\n        \"\",\n        \"Checking service: mariadb (MariaDB for ARA)\",\n        \"  Expected: registry.osism.tech/dockerhub/library/mariadb:11.8.4\",\n        \"  Enabled:  true\",\n        \"  Running:  registry.osism.tech/dockerhub/library/mariadb:11.8.4\",\n        \"  Status:   \u2705 MATCH\",\n        \"\",\n        \"Checking service: frontend (OSISM Frontend)\",\n        \"  Expected: registry.osism.tech/osism/osism-frontend:latest\",\n        \"  Enabled:  true\",\n        \"  Running:  registry.osism.tech/osism/osism-frontend:latest\",\n        \"  Status:   \u2705 MATCH\",\n        \"\",\n        \"Checking service: redis (Redis Cache)\",\n        \"  Expected: registry.osism.tech/dockerhub/library/redis:7.4.7-alpine\",\n        \"  Enabled:  true\",\n        \"  Running:  registry.osism.tech/dockerhub/library/redis:7.4.7-alpine\",\n        \"  Status:   \u2705 MATCH\",\n        \"\",\n        \"Checking service: api (OSISM API Service)\",\n        \"  Expected: registry.osism.tech/osism/osism:latest\",\n        \"  Enabled:  true\",\n        \"  Running:  registry.osism.tech/osism/osism:latest\",\n        \"  Status:   \u2705 MATCH\",\n        \"\",\n        \"Checking service: listener (OpenStack Event Listener)\",\n        \"  Expected: registry.osism.tech/osism/osism:latest\",\n        \"  Enabled:  true\",\n        \"  Running:  registry.osism.tech/osism/osism:latest\",\n        \"  Status:   \u2705 MATCH\",\n        \"\",\n        \"Checking service: openstack (OpenStack Integration)\",\n        \"  Expected: registry.osism.tech/osism/osism:latest\",\n        \"  Enabled:  true\",\n        \"  Running:  registry.osism.tech/osism/osism:latest\",\n        \"  Status:   \u2705 MATCH\",\n        \"\",\n        \"Checking service: beat (Celery Beat Scheduler)\",\n        \"  Expected: registry.osism.tech/osism/osism:latest\",\n        \"  Enabled:  true\",\n        \"  Running:  registry.osism.tech/osism/osism:latest\",\n        \"  Status:   \u2705 MATCH\",\n        \"\",\n        \"Checking service: flower (Celery Flower Monitor)\",\n        \"  Expected: registry.osism.tech/osism/osism:latest\",\n        \"  Enabled:  true\",\n        \"  Running:  registry.osism.tech/osism/osism:latest\",\n        \"  Status:   \u2705 MATCH\",\n        \"\",\n        \"=== Summary ===\",\n        \"Errors (version mismatches): 0\",\n        \"Warnings (expected containers not running): 0\",\n        \"\",\n        \"\u2705 All running containers match expected versions!\"\n    ]\n}\n\nTASK [osism.services.manager : Skip version check due to service configuration] ***\nskipping: [testbed-manager]\n\nPLAY RECAP *********************************************************************\ntestbed-manager            : ok=70   changed=37   unreachable=0    failed=0    skipped=12   rescued=0    ignored=0   \n\n+ [[ -e /opt/venv/bin/activate ]]\n+ deactivate\n+ '[' -n /usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/usr/games:/usr/local/games:/snap/bin ']'\n+ PATH=/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/usr/games:/usr/local/games:/snap/bin\n+ export PATH\n+ unset _OLD_VIRTUAL_PATH\n+ '[' -n '' ']'\n+ hash -r\n+ '[' -n '' ']'\n+ unset VIRTUAL_ENV\n+ unset VIRTUAL_ENV_PROMPT\n+ '[' '!' '' = nondestructive ']'\n+ unset -f deactivate\n+ cp /home/dragon/.ssh/id_rsa.pub /opt/ansible/secrets/id_rsa.operator.pub\n+ [[ ceph-ansible == \\c\\e\\p\\h\\-\\a\\n\\s\\i\\b\\l\\e ]]\n+ wait_for_container_healthy 60 ceph-ansible\n+ local max_attempts=60\n+ local name=ceph-ansible\n+ local attempt_num=1\n++ /usr/bin/docker inspect -f '{{.State.Health.Status}}' ceph-ansible\n+ [[ healthy == \\h\\e\\a\\l\\t\\h\\y ]]\n+ wait_for_container_healthy 60 kolla-ansible\n+ local max_attempts=60\n+ local name=kolla-ansible\n+ local attempt_num=1\n++ /usr/bin/docker inspect -f '{{.State.Health.Status}}' kolla-ansible\n+ [[ healthy == \\h\\e\\a\\l\\t\\h\\y ]]\n+ wait_for_container_healthy 60 osism-ansible\n+ local max_attempts=60\n+ local name=osism-ansible\n+ local attempt_num=1\n++ /usr/bin/docker inspect -f '{{.State.Health.Status}}' osism-ansible\n+ [[ healthy == \\h\\e\\a\\l\\t\\h\\y ]]\n+ [[ true == \\t\\r\\u\\e ]]\n+ sh -c /opt/configuration/scripts/disable-ara.sh\n+ docker compose --project-directory /opt/manager ps\nNAME                             IMAGE                                                      COMMAND                  SERVICE                CREATED              STATUS                        PORTS\nceph-ansible                     registry.osism.tech/osism/ceph-ansible:reef                \"/entrypoint.sh osis\u2026\"   ceph-ansible           About a minute ago   Up About a minute (healthy)   \nkolla-ansible                    registry.osism.tech/osism/kolla-ansible:2024.2             \"/entrypoint.sh osis\u2026\"   kolla-ansible          About a minute ago   Up About a minute (healthy)   \nmanager-api-1                    registry.osism.tech/osism/osism:latest                     \"/sbin/tini -- osism\u2026\"   api                    About a minute ago   Up About a minute (healthy)   192.168.16.5:8000->8000/tcp\nmanager-ara-server-1             registry.osism.tech/osism/ara-server:1.7.3                 \"sh -c '/wait && /ru\u2026\"   ara-server             About a minute ago   Up About a minute (healthy)   8000/tcp\nmanager-beat-1                   registry.osism.tech/osism/osism:latest                     \"/sbin/tini -- osism\u2026\"   beat                   About a minute ago   Up About a minute (healthy)   \nmanager-flower-1                 registry.osism.tech/osism/osism:latest                     \"/sbin/tini -- osism\u2026\"   flower                 About a minute ago   Up About a minute (healthy)   \nmanager-inventory_reconciler-1   registry.osism.tech/osism/inventory-reconciler:latest      \"/sbin/tini -- /entr\u2026\"   inventory_reconciler   About a minute ago   Up 56 seconds (healthy)       \nmanager-listener-1               registry.osism.tech/osism/osism:latest                     \"/sbin/tini -- osism\u2026\"   listener               About a minute ago   Up About a minute (healthy)   \nmanager-mariadb-1                registry.osism.tech/dockerhub/library/mariadb:11.8.4       \"docker-entrypoint.s\u2026\"   mariadb                About a minute ago   Up About a minute (healthy)   3306/tcp\nmanager-openstack-1              registry.osism.tech/osism/osism:latest                     \"/sbin/tini -- osism\u2026\"   openstack              About a minute ago   Up About a minute (healthy)   \nmanager-redis-1                  registry.osism.tech/dockerhub/library/redis:7.4.7-alpine   \"docker-entrypoint.s\u2026\"   redis                  About a minute ago   Up About a minute (healthy)   6379/tcp\nosism-ansible                    registry.osism.tech/osism/osism-ansible:latest             \"/entrypoint.sh osis\u2026\"   osism-ansible          About a minute ago   Up About a minute (healthy)   \nosism-frontend                   registry.osism.tech/osism/osism-frontend:latest            \"docker-entrypoint.s\u2026\"   frontend               About a minute ago   Up About a minute             192.168.16.5:3000->3000/tcp\nosism-kubernetes                 registry.osism.tech/osism/osism-kubernetes:latest          \"/entrypoint.sh osis\u2026\"   osism-kubernetes       About a minute ago   Up About a minute (healthy)   \nosismclient                      registry.osism.tech/osism/osism:latest                     \"/sbin/tini -- sleep\u2026\"   osismclient            About a minute ago   Up About a minute (healthy)   \n++ semver latest 7.0.0\n+ [[ -1 -ge 0 ]]\n+ [[ latest == \\l\\a\\t\\e\\s\\t ]]\n+ sed -i s/community.general.yaml/osism.commons.still_alive/ /opt/configuration/environments/ansible.cfg\n+ osism apply resolvconf -l testbed-manager\n\u001b[32m2026-02-24 00:22:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mPrepare task for execution of resolvconf.\u001b[0m\n\u001b[32m2026-02-24 00:22:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5a9194ec-1896-415a-bfdb-560ac86dbab4 (resolvconf) was prepared for execution.\u001b[0m\n\u001b[32m2026-02-24 00:22:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task 5a9194ec-1896-415a-bfdb-560ac86dbab4 (resolvconf) has been started and output is visible here.\u001b[0m\n\nPLAY [Apply role resolvconf] ***************************************************\n\nTASK [Gathering Facts] *********************************************************\nTuesday 24 February 2026  00:22:04 +0000 (0:00:00.101)       0:00:00.101 ****** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\nTASK [osism.commons.resolvconf : Check minimum and maximum number of name servers] ***\nTuesday 24 February 2026  00:22:07 +0000 (0:00:03.314)       0:00:03.416 ****** \n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\nTASK [osism.commons.resolvconf : Include resolvconf tasks] *********************\nTuesday 24 February 2026  00:22:07 +0000 (0:00:00.063)       0:00:03.480 ****** \n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/resolvconf/tasks/configure-resolv.yml for testbed-manager\u001b[0m\n\nTASK [osism.commons.resolvconf : Include distribution specific installation tasks] ***\nTuesday 24 February 2026  00:22:07 +0000 (0:00:00.089)       0:00:03.569 ****** \n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/resolvconf/tasks/install-Debian-family.yml for testbed-manager\u001b[0m\n\nTASK [osism.commons.resolvconf : Remove packages configuring /etc/resolv.conf] ***\nTuesday 24 February 2026  00:22:07 +0000 (0:00:00.066)       0:00:03.635 ****** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\nTASK [osism.commons.resolvconf : Install package systemd-resolved] *************\nTuesday 24 February 2026  00:22:08 +0000 (0:00:00.866)       0:00:04.502 ****** \n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\nTASK [osism.commons.resolvconf : Retrieve file status of /etc/resolv.conf] *****\nTuesday 24 February 2026  00:22:08 +0000 (0:00:00.061)       0:00:04.564 ****** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\nTASK [osism.commons.resolvconf : Archive existing file /etc/resolv.conf] *******\nTuesday 24 February 2026  00:22:09 +0000 (0:00:01.440)       0:00:06.004 ****** \n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\nTASK [osism.commons.resolvconf : Link /run/systemd/resolve/stub-resolv.conf to /etc/resolv.conf] ***\nTuesday 24 February 2026  00:22:10 +0000 (0:00:00.072)       0:00:06.077 ****** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nTASK [osism.commons.resolvconf : Copy configuration files] *********************\nTuesday 24 February 2026  00:22:10 +0000 (0:00:00.485)       0:00:06.562 ****** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nTASK [osism.commons.resolvconf : Start/enable systemd-resolved service] ********\nTuesday 24 February 2026  00:22:11 +0000 (0:00:00.956)       0:00:07.519 ****** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\nTASK [osism.commons.resolvconf : Include distribution specific configuration tasks] ***\nTuesday 24 February 2026  00:22:12 +0000 (0:00:00.874)       0:00:08.394 ****** \n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/resolvconf/tasks/configure-Debian-family.yml for testbed-manager\u001b[0m\n\nTASK [osism.commons.resolvconf : Restart systemd-resolved service] *************\nTuesday 24 February 2026  00:22:12 +0000 (0:00:00.077)       0:00:08.471 ****** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;33mtestbed-manager\u001b[0m            : \u001b[0;32mok=10  \u001b[0m \u001b[0;33mchanged=3   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=3   \u001b[0m rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nTuesday 24 February 2026  00:22:13 +0000 (0:00:01.099)       0:00:09.571 ****** \n=============================================================================== \nGathering Facts --------------------------------------------------------- 3.31s\nosism.commons.resolvconf : Retrieve file status of /etc/resolv.conf ----- 1.44s\nosism.commons.resolvconf : Restart systemd-resolved service ------------- 1.10s\nosism.commons.resolvconf : Copy configuration files --------------------- 0.96s\nosism.commons.resolvconf : Start/enable systemd-resolved service -------- 0.87s\nosism.commons.resolvconf : Remove packages configuring /etc/resolv.conf --- 0.87s\nosism.commons.resolvconf : Link /run/systemd/resolve/stub-resolv.conf to /etc/resolv.conf --- 0.49s\nosism.commons.resolvconf : Include resolvconf tasks --------------------- 0.09s\nosism.commons.resolvconf : Include distribution specific configuration tasks --- 0.08s\nosism.commons.resolvconf : Archive existing file /etc/resolv.conf ------- 0.07s\nosism.commons.resolvconf : Include distribution specific installation tasks --- 0.07s\nosism.commons.resolvconf : Check minimum and maximum number of name servers --- 0.06s\nosism.commons.resolvconf : Install package systemd-resolved ------------- 0.06s\n+ osism apply sshconfig\n\u001b[32m2026-02-24 00:22:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mPrepare task for execution of sshconfig.\u001b[0m\n\u001b[32m2026-02-24 00:22:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 0ea583c6-54c2-46df-adef-59f07b9f4681 (sshconfig) was prepared for execution.\u001b[0m\n\u001b[32m2026-02-24 00:22:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task 0ea583c6-54c2-46df-adef-59f07b9f4681 (sshconfig) has been started and output is visible here.\u001b[0m\n\nPLAY [Apply role sshconfig] ****************************************************\n\nTASK [osism.commons.sshconfig : Get home directory of operator user] ***********\nTuesday 24 February 2026  00:22:29 +0000 (0:00:00.117)       0:00:00.117 ****** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\nTASK [osism.commons.sshconfig : Ensure .ssh/config.d exist] ********************\nTuesday 24 February 2026  00:22:30 +0000 (0:00:00.467)       0:00:00.584 ****** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nTASK [osism.commons.sshconfig : Ensure config for each host exist] *************\nTuesday 24 February 2026  00:22:30 +0000 (0:00:00.365)       0:00:00.950 ****** \n\u001b[0;33mchanged: [testbed-manager] => (item=testbed-node-0)\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item=testbed-node-1)\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item=testbed-node-2)\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item=testbed-node-3)\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item=testbed-node-4)\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item=testbed-node-5)\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item=testbed-manager)\u001b[0m\n\nTASK [osism.commons.sshconfig : Add extra config] ******************************\nTuesday 24 February 2026  00:22:35 +0000 (0:00:05.158)       0:00:06.109 ****** \n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\nTASK [osism.commons.sshconfig : Assemble ssh config] ***************************\nTuesday 24 February 2026  00:22:35 +0000 (0:00:00.077)       0:00:06.186 ****** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;33mtestbed-manager\u001b[0m            : \u001b[0;32mok=4   \u001b[0m \u001b[0;33mchanged=3   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=1   \u001b[0m rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nTuesday 24 February 2026  00:22:36 +0000 (0:00:00.568)       0:00:06.755 ****** \n=============================================================================== \nosism.commons.sshconfig : Ensure config for each host exist ------------- 5.16s\nosism.commons.sshconfig : Assemble ssh config --------------------------- 0.57s\nosism.commons.sshconfig : Get home directory of operator user ----------- 0.47s\nosism.commons.sshconfig : Ensure .ssh/config.d exist -------------------- 0.37s\nosism.commons.sshconfig : Add extra config ------------------------------ 0.08s\n+ osism apply known-hosts\n\u001b[32m2026-02-24 00:22:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mPrepare task for execution of known-hosts.\u001b[0m\n\u001b[32m2026-02-24 00:22:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 688403ab-c18b-4137-889d-96137831af39 (known-hosts) was prepared for execution.\u001b[0m\n\u001b[32m2026-02-24 00:22:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task 688403ab-c18b-4137-889d-96137831af39 (known-hosts) has been started and output is visible here.\u001b[0m\n\nPLAY [Apply role known_hosts] **************************************************\n\nTASK [osism.commons.known_hosts : Run ssh-keyscan for all hosts with hostname] ***\nTuesday 24 February 2026  00:22:52 +0000 (0:00:00.166)       0:00:00.166 ****** \n\u001b[0;32mok: [testbed-manager] => (item=testbed-node-3)\u001b[0m\n\u001b[0;32mok: [testbed-manager] => (item=testbed-node-4)\u001b[0m\n\u001b[0;32mok: [testbed-manager] => (item=testbed-node-5)\u001b[0m\n\u001b[0;32mok: [testbed-manager] => (item=testbed-manager)\u001b[0m\n\u001b[0;32mok: [testbed-manager] => (item=testbed-node-0)\u001b[0m\n\u001b[0;32mok: [testbed-manager] => (item=testbed-node-1)\u001b[0m\n\u001b[0;32mok: [testbed-manager] => (item=testbed-node-2)\u001b[0m\n\nTASK [osism.commons.known_hosts : Write scanned known_hosts entries for all hosts with hostname] ***\nTuesday 24 February 2026  00:22:58 +0000 (0:00:05.860)       0:00:06.026 ****** \n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/known_hosts/tasks/write-scanned.yml for testbed-manager => (item=Scanned entries of testbed-node-3)\u001b[0m\n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/known_hosts/tasks/write-scanned.yml for testbed-manager => (item=Scanned entries of testbed-node-4)\u001b[0m\n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/known_hosts/tasks/write-scanned.yml for testbed-manager => (item=Scanned entries of testbed-node-5)\u001b[0m\n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/known_hosts/tasks/write-scanned.yml for testbed-manager => (item=Scanned entries of testbed-manager)\u001b[0m\n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/known_hosts/tasks/write-scanned.yml for testbed-manager => (item=Scanned entries of testbed-node-0)\u001b[0m\n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/known_hosts/tasks/write-scanned.yml for testbed-manager => (item=Scanned entries of testbed-node-1)\u001b[0m\n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/known_hosts/tasks/write-scanned.yml for testbed-manager => (item=Scanned entries of testbed-node-2)\u001b[0m\n\nTASK [osism.commons.known_hosts : Write scanned known_hosts entries] ***********\nTuesday 24 February 2026  00:22:58 +0000 (0:00:00.161)       0:00:06.188 ****** \n\u001b[0;33mchanged: [testbed-manager] => (item=testbed-node-3 ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABgQDOWRdk1N8OIeC4uCbXt2ifzDinxCcJxpFg+cy8/O9qdNiU6uF6MYSNDWdTmj5LsQfBOv9B2tGl+oFtMoV6HFtFDsUgRunw1ZTuofhokH0eYBBrukiCLi7dxc7SnmFZ2ITcQjZB2BhCHKwtvOxTBksVGswVmjoPGjlciYKDTaeLZD3zYiLjKGP1UKtIfjzrxuwVuHbzGMwnExq8l2JeYF84/mrQP2FRBfH/UkCVyuvZnHhQAHRoURNmhrCwaepBbx2ZaZNUEdmxJnkJMWJtvpk3C7BPgiGEPwLlhM79cqKzgtFZkZ+7/qZCZACjPF9aaTds5utFpV74TDHtAQrUKM5KAI7dFIf9jeNVWw6r0vgL44maMvPIYQD6vinwzlujJKG0cBCq678Up0uTRF07q5NpCBf0oGWOC8LoxdycSfI2Vs3ljaUt+b3T4R0BO5sNT2JCHRdcUNZwFAGwZ4RW4RCGD0j5w26RMy9pp+ofRgsAUEYDHRkZ/4WHfMxIUMjwmU0=)\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item=testbed-node-3 ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBL2uknstYHYYZxd3ycHLBWlZpyvOW+d9YIVVFgEYz3Zxjh4zS4EdPoYhodFG1Qr3nQQ5Rvp4lzvZiJjxhFxTZ28=)\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item=testbed-node-3 ssh-ed25519 AAAAC3NzaC1lZDI1NTE5AAAAIN2iQH254D6gxizx/kDK2ek50uPnRRnQlJj02mWpt5cH)\u001b[0m\n\nTASK [osism.commons.known_hosts : Write scanned known_hosts entries] ***********\nTuesday 24 February 2026  00:22:59 +0000 (0:00:01.122)       0:00:07.311 ****** \n\u001b[0;33mchanged: [testbed-manager] => (item=testbed-node-4 ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABgQCQjWzk/nNVedMF8uWXwcthkgBn54kNzWDVJ5eAPCCqD++s091Bnx29/cjP7Q+lEdxp8BnEg6BnF1cH9i1RmNbAgQCZEjLF7WwCK0M9ptx6Ebp7M9j0UtRuFhtNBOIefbKOwTKJw21yOvzNX2xzvH8/WS1ODjJeH0ZbhWhXHgdKisLQjwJmaxGD+nYxDFAhWFjcfBE5kb+STYmdxujCZ6b8pARZS4XJUEPFZZgx7U0jylvG8nLUN7CgCxy9breJpHMKtjk3Mm7zSrH0Nhop1eiz4c/v8/Zag21EPO9jo1R3fvIOeAYIvQ3eQdBfRULWc89Ijgdvq8z6+OeFRx0wCP6Zzcw9w3nE1if7IkikF4MDaCXIIlf/KW77/6r1iWWol+aOBNR64BA9d9VjstePAQAHlQ/a/qONP5kD1NRXAEhAwuJzdxjvX7RjyuntHNE6n/kcDEfacZKEMThOs/WyIVRVkAgvjgj9wSkSYMqp1kkhKY5QsfOocegMHtkKtqb4FB8=)\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item=testbed-node-4 ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBOly8FEun0/C8rfMAzTzUQzKUsbnR8IbBcxxSLcEEY/FPTQCBr4O94lHghhDYXhW3M0OoKHX8XqwcnbnJumkl4g=)\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item=testbed-node-4 ssh-ed25519 AAAAC3NzaC1lZDI1NTE5AAAAINkXgg1KUois1zDu7aBhB83ATBEugpnY7XNMswDNHl3w)\u001b[0m\n\nTASK [osism.commons.known_hosts : Write scanned known_hosts entries] ***********\nTuesday 24 February 2026  00:23:00 +0000 (0:00:01.017)       0:00:08.328 ****** \n\u001b[0;33mchanged: [testbed-manager] => (item=testbed-node-5 ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBAeSrLLHw6GzSjU89slqhT47VMqLM9hsem1yhDpwlj3R+Q4qo3kzcxiGcXhAd6bD1fjqCK3Wfu1+cKO9fhEy1g4=)\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item=testbed-node-5 ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABgQCsN9mDKfSY623Rz4GeNVRKPyAQ2P/kyJ1M+ciS9jyXpN6jHd1oAzAqVipldMZnQkPqLtbfhsAgFdIVodbBCIlTURpF1N64E61lwTsipSvaqRW+gWwNykSp0Kd2tgjA8QDzk0CdLtTIR0Y3G2xepnzsMFF7ybQyNx80kovfbvbA/M5u8pekTSQ9hO27dVqafm/Rkg6Se7e7C3Va9Bunhj+O5UUJofrtdFJRpOMYUilHjqXgehDZpamdm9zIFDyxlXLqB/oJmcr3Zn9JFLgwMuZfSc+z8G/c5H1lJLYJr0Pd/0rsb826cqxKrFy//JDih1XP8Q/66TDP9wmvBJts2MrsmXe1d6zukaxINB436jntilYM+bzRL+8dH9eXexj/sUZZiBCyfGAlTJr+I4RARGU49FvPvKQtz1SQhQoUB1OR/6FooWx6yv1CA/Dw5PiZuSvECNDR++LE6ZvuLqAMvwVHm7zxgSx5q0VNtIqN+PTHN6yq9r2rnhErhLoOuQThUm8=)\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item=testbed-node-5 ssh-ed25519 AAAAC3NzaC1lZDI1NTE5AAAAIP4XvFEG+/G6UU398M8h8QkghZwwa+RxCDcBl6tu+MbK)\u001b[0m\n\nTASK [osism.commons.known_hosts : Write scanned known_hosts entries] ***********\nTuesday 24 February 2026  00:23:01 +0000 (0:00:01.000)       0:00:09.329 ****** \n\u001b[0;33mchanged: [testbed-manager] => (item=testbed-manager ssh-ed25519 AAAAC3NzaC1lZDI1NTE5AAAAIDfU67w4g0t5ruLoohFtsOVj0FEJKCFA6yTUXHoTeiU+)\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item=testbed-manager ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABgQCUeJXF5lVxZV05KJ56am57/G4Pj6M8Ir07kqn5lXoOZXaKtAM034AYvC6FfkuxmYUsc69IbF5cYYNzWSTi7uigANnV3Gtkq8OlDelwVdV3vQDp+71dO6t30HlzvuEvk4qjeYkciisK/RE5RqVZCkqkMQIQR7TunEGspNpQyMP0Jb6CHkRMFAhPzIDv+Hcun0aWylP6WK+44TYcReX6RIOILPhPUXYV9H0ZFoC2L/hup+kkY1S85c8WaqH/OcR/XnOLwODXyRpMm2DX9kwhU0QeIeO+VqU4Jp6aXAfjSLbFbgAcm7LMAA7qdTd3liWDXJO22pgua2aAolsOe8gcFWZtMlD0AwtOUBd5c2MFrrn85aoaINKiXu/fOJxsV67SX729tb+1aXygYdUsjDxJBNHKdkmaLGpWBVgkz++0yy6DS6viWLi2nDL/3p80JUPbLv57s8zpev1x9ZMHJxbWkCjGgqjZ7HZ2WAsNklDWNgvZ03i/KiiUoaCCDjIofbR2R0E=)\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item=testbed-manager ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBAAEM4MNA1yZvjsAJIz6pihEIhOsw75Cu+tJPTJEbZDNCdZ3BprKWAIEwVr1UFlQTAVPSVDKMXHBd3XzGkopW84=)\u001b[0m\n\nTASK [osism.commons.known_hosts : Write scanned known_hosts entries] ***********\nTuesday 24 February 2026  00:23:02 +0000 (0:00:00.982)       0:00:10.312 ****** \n\u001b[0;33mchanged: [testbed-manager] => (item=testbed-node-0 ssh-ed25519 AAAAC3NzaC1lZDI1NTE5AAAAIC9+a2YIH9Deu7Kp5MuasTBGpKoKvAwa8BiKo1b3H76+)\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item=testbed-node-0 ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABgQC1zWCyl2+3CBj3nqxQhXBIhbNhR0LfsVZ8HtUXLylfIQOHN5bMATCm8HbT8KZ3HbGbluD5kffIIFrqxlSvMyuKnSFBlIwg1dSEaJztZQm1QPmZEkPMD+C+8HudVc5itevm9fwKDVWPt0lZ8zKpFxzF6Rum/bfFKJQcmPxl+zIL46mr4PbBLibBlHbAUs1ZnkZXwGXtkghKmrl8lvFskCXPllSKZcLPeM6QImpZa99TwLNhC1FnkJhLWtW25iDh53zIkoCFizujMM38wHNGIiu7rKB/CMgnX8LsKYZoSYatqHivgegBJd79m3myxBkGOsQ0YmCGGeUPRpOzabncaSPmDRsWpTafchsA7JJX6xXc4Q8+v/Zx+CNOg2oPpPnMJegXxgBAMkXeiWCaB5sbHLFmCD8eBFrOxtBeS0RqaBSHylkQBJ4q6aNJ2jIK0i2XnDVTbTrfhwkeIY9Uz5z8DuKm9HJ3x6rPJofnDSdPLdXXT3XDnj6B1riH5sVGIbKRglk=)\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item=testbed-node-0 ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBA+CLgYVB+RtOt+86BTUxoGVVBxkcCRXmZpt5JEy6gYHueijrlVosdVEchOoU/pc/10Ze4k+BS7Mg3o6u6GUV8s=)\u001b[0m\n\nTASK [osism.commons.known_hosts : Write scanned known_hosts entries] ***********\nTuesday 24 February 2026  00:23:03 +0000 (0:00:01.043)       0:00:11.355 ****** \n\u001b[0;33mchanged: [testbed-manager] => (item=testbed-node-1 ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBPjiHNE7pTy7o8Vd28ZoO+w4U30fFOydWvvQBQp9qayGswAO7W20tyqB2rWFbwwhAjAYAGZX2sFLF0X/ulaZkVA=)\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item=testbed-node-1 ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABgQCbVubwVzGRukWm151/SYPu5+ZUCffbxzUGqgLxjp6PNUEvt1yzTcQ1bVHQ4y09uUCZjwDql6LL3gOIfV6fpdzXAd/FEDQt2fSoHvRXxH8BafXLEVwctwrdcoLPcJkpgl+qMgysgFgL9bHa0Rd6jP2Npur0vk3Y5ju723LOiRroRByOYiHC4XNS7DxHu6u+L/D6hVqQz5jwzP0tDXCF/+LNJ6uiaTDOd+sywTVq7ITDItNLaXIZGpsNjubQy1lMqyVZeaI5Le1/wwJDWhsrBrIyFm3Kemhecn/PVJZmPPk2OFvWp48Uyo0+NubcquE50ezJVbAmhbFG5c6HUrpVHcWLNk4LzZ6F5bRpR6V42QvBvcgwnMtzBw8QRXf96EFtzv4K5EG4Y+1Ew1sWzjWXYD/ZlA6Cc9qu5D23ebFPZIiY8o8dhBlj4iXHxO3kABxCDEJS49apyfvvKwFQ5+FjMCy88w1eO5LOwT8BXpUXcXgNcNWHidtPZ4ufzsqLYoP9Fhk=)\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item=testbed-node-1 ssh-ed25519 AAAAC3NzaC1lZDI1NTE5AAAAIPlYo1pFim6htda1AoAJkI0dbcELCmM5QFZYRxrp2T8p)\u001b[0m\n\nTASK [osism.commons.known_hosts : Write scanned known_hosts entries] ***********\nTuesday 24 February 2026  00:23:04 +0000 (0:00:00.978)       0:00:12.334 ****** \n\u001b[0;33mchanged: [testbed-manager] => (item=testbed-node-2 ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBNMB0mFLquOvLDLHWMx+FbOs6ia8nAHzsLjwVkHj6z+8y/yvQBEmH3hQRj/jml5GScCC1Ncbk/HPjJCUU52aLP4=)\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item=testbed-node-2 ssh-ed25519 AAAAC3NzaC1lZDI1NTE5AAAAID9TiTT29JB5SGhfDhxbtI4boFbWaY41+GMclNzBxg2p)\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item=testbed-node-2 ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABgQDCMRk074A67PfDSfiSZySlWWdBg82NzopzXCqwhSXtyjr/LVN4q3Z3VWst7PL1hjux2gaZvX6+h/RWI8ub/hOF8VMqPQ90NdJOy53UuMtsrIlovBWpDHMUa3v5FrrtAFY1Fve1POj+5kcPG69xUj/N6wLkiLab8+VqDnJLia7IsrYFnU4xShfL963gkMDvZvamLExqWws9ZZHfZc37vaYL9ntX3bcmjW76hTp+q/jMjXIxy5MRNZP+0YLq6Ig7BQb8VrrHSMx0FQgWX0UGso0BQQcaRjUl5T9yWYCJBMt3ciRZvs+Rordy878/xbYLADmgU4Vyy6RFkGxkoOQPiNMtZfMx0kLSqd8FYlysLgjvUjKFAv5uB3MAWD2Sp4CSIKI/Hbn7lzTEGIcgUopve3iif2R+oTGLje8I2g0Tqbxl53+oOWjZrPnYuCFarbJ2siPpQljjEVmsI1Y72JHFcQyRpJUp0rReRH1BpnxiPIFzr76k4hIgPNqbBi0o8zJri2k=)\u001b[0m\n\nTASK [osism.commons.known_hosts : Run ssh-keyscan for all hosts with ansible_host] ***\nTuesday 24 February 2026  00:23:05 +0000 (0:00:00.981)       0:00:13.315 ****** \n\u001b[0;32mok: [testbed-manager] => (item=testbed-node-3)\u001b[0m\n\u001b[0;32mok: [testbed-manager] => (item=testbed-node-4)\u001b[0m\n\u001b[0;32mok: [testbed-manager] => (item=testbed-node-5)\u001b[0m\n\u001b[0;32mok: [testbed-manager] => (item=testbed-manager)\u001b[0m\n\u001b[0;32mok: [testbed-manager] => (item=testbed-node-0)\u001b[0m\n\u001b[0;32mok: [testbed-manager] => (item=testbed-node-1)\u001b[0m\n\u001b[0;32mok: [testbed-manager] => (item=testbed-node-2)\u001b[0m\n\nTASK [osism.commons.known_hosts : Write scanned known_hosts entries for all hosts with ansible_host] ***\nTuesday 24 February 2026  00:23:10 +0000 (0:00:05.048)       0:00:18.363 ****** \n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/known_hosts/tasks/write-scanned.yml for testbed-manager => (item=Scanned entries of testbed-node-3)\u001b[0m\n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/known_hosts/tasks/write-scanned.yml for testbed-manager => (item=Scanned entries of testbed-node-4)\u001b[0m\n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/known_hosts/tasks/write-scanned.yml for testbed-manager => (item=Scanned entries of testbed-node-5)\u001b[0m\n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/known_hosts/tasks/write-scanned.yml for testbed-manager => (item=Scanned entries of testbed-manager)\u001b[0m\n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/known_hosts/tasks/write-scanned.yml for testbed-manager => (item=Scanned entries of testbed-node-0)\u001b[0m\n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/known_hosts/tasks/write-scanned.yml for testbed-manager => (item=Scanned entries of testbed-node-1)\u001b[0m\n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/known_hosts/tasks/write-scanned.yml for testbed-manager => (item=Scanned entries of testbed-node-2)\u001b[0m\n\nTASK [osism.commons.known_hosts : Write scanned known_hosts entries] ***********\nTuesday 24 February 2026  00:23:11 +0000 (0:00:00.171)       0:00:18.535 ****** \n\u001b[0;33mchanged: [testbed-manager] => (item=192.168.16.13 ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBL2uknstYHYYZxd3ycHLBWlZpyvOW+d9YIVVFgEYz3Zxjh4zS4EdPoYhodFG1Qr3nQQ5Rvp4lzvZiJjxhFxTZ28=)\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item=192.168.16.13 ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABgQDOWRdk1N8OIeC4uCbXt2ifzDinxCcJxpFg+cy8/O9qdNiU6uF6MYSNDWdTmj5LsQfBOv9B2tGl+oFtMoV6HFtFDsUgRunw1ZTuofhokH0eYBBrukiCLi7dxc7SnmFZ2ITcQjZB2BhCHKwtvOxTBksVGswVmjoPGjlciYKDTaeLZD3zYiLjKGP1UKtIfjzrxuwVuHbzGMwnExq8l2JeYF84/mrQP2FRBfH/UkCVyuvZnHhQAHRoURNmhrCwaepBbx2ZaZNUEdmxJnkJMWJtvpk3C7BPgiGEPwLlhM79cqKzgtFZkZ+7/qZCZACjPF9aaTds5utFpV74TDHtAQrUKM5KAI7dFIf9jeNVWw6r0vgL44maMvPIYQD6vinwzlujJKG0cBCq678Up0uTRF07q5NpCBf0oGWOC8LoxdycSfI2Vs3ljaUt+b3T4R0BO5sNT2JCHRdcUNZwFAGwZ4RW4RCGD0j5w26RMy9pp+ofRgsAUEYDHRkZ/4WHfMxIUMjwmU0=)\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item=192.168.16.13 ssh-ed25519 AAAAC3NzaC1lZDI1NTE5AAAAIN2iQH254D6gxizx/kDK2ek50uPnRRnQlJj02mWpt5cH)\u001b[0m\n\nTASK [osism.commons.known_hosts : Write scanned known_hosts entries] ***********\nTuesday 24 February 2026  00:23:12 +0000 (0:00:01.013)       0:00:19.549 ****** \n\u001b[0;33mchanged: [testbed-manager] => (item=192.168.16.14 ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBOly8FEun0/C8rfMAzTzUQzKUsbnR8IbBcxxSLcEEY/FPTQCBr4O94lHghhDYXhW3M0OoKHX8XqwcnbnJumkl4g=)\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item=192.168.16.14 ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABgQCQjWzk/nNVedMF8uWXwcthkgBn54kNzWDVJ5eAPCCqD++s091Bnx29/cjP7Q+lEdxp8BnEg6BnF1cH9i1RmNbAgQCZEjLF7WwCK0M9ptx6Ebp7M9j0UtRuFhtNBOIefbKOwTKJw21yOvzNX2xzvH8/WS1ODjJeH0ZbhWhXHgdKisLQjwJmaxGD+nYxDFAhWFjcfBE5kb+STYmdxujCZ6b8pARZS4XJUEPFZZgx7U0jylvG8nLUN7CgCxy9breJpHMKtjk3Mm7zSrH0Nhop1eiz4c/v8/Zag21EPO9jo1R3fvIOeAYIvQ3eQdBfRULWc89Ijgdvq8z6+OeFRx0wCP6Zzcw9w3nE1if7IkikF4MDaCXIIlf/KW77/6r1iWWol+aOBNR64BA9d9VjstePAQAHlQ/a/qONP5kD1NRXAEhAwuJzdxjvX7RjyuntHNE6n/kcDEfacZKEMThOs/WyIVRVkAgvjgj9wSkSYMqp1kkhKY5QsfOocegMHtkKtqb4FB8=)\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item=192.168.16.14 ssh-ed25519 AAAAC3NzaC1lZDI1NTE5AAAAINkXgg1KUois1zDu7aBhB83ATBEugpnY7XNMswDNHl3w)\u001b[0m\n\nTASK [osism.commons.known_hosts : Write scanned known_hosts entries] ***********\nTuesday 24 February 2026  00:23:13 +0000 (0:00:01.018)       0:00:20.568 ****** \n\u001b[0;33mchanged: [testbed-manager] => (item=192.168.16.15 ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABgQCsN9mDKfSY623Rz4GeNVRKPyAQ2P/kyJ1M+ciS9jyXpN6jHd1oAzAqVipldMZnQkPqLtbfhsAgFdIVodbBCIlTURpF1N64E61lwTsipSvaqRW+gWwNykSp0Kd2tgjA8QDzk0CdLtTIR0Y3G2xepnzsMFF7ybQyNx80kovfbvbA/M5u8pekTSQ9hO27dVqafm/Rkg6Se7e7C3Va9Bunhj+O5UUJofrtdFJRpOMYUilHjqXgehDZpamdm9zIFDyxlXLqB/oJmcr3Zn9JFLgwMuZfSc+z8G/c5H1lJLYJr0Pd/0rsb826cqxKrFy//JDih1XP8Q/66TDP9wmvBJts2MrsmXe1d6zukaxINB436jntilYM+bzRL+8dH9eXexj/sUZZiBCyfGAlTJr+I4RARGU49FvPvKQtz1SQhQoUB1OR/6FooWx6yv1CA/Dw5PiZuSvECNDR++LE6ZvuLqAMvwVHm7zxgSx5q0VNtIqN+PTHN6yq9r2rnhErhLoOuQThUm8=)\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item=192.168.16.15 ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBAeSrLLHw6GzSjU89slqhT47VMqLM9hsem1yhDpwlj3R+Q4qo3kzcxiGcXhAd6bD1fjqCK3Wfu1+cKO9fhEy1g4=)\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item=192.168.16.15 ssh-ed25519 AAAAC3NzaC1lZDI1NTE5AAAAIP4XvFEG+/G6UU398M8h8QkghZwwa+RxCDcBl6tu+MbK)\u001b[0m\n\nTASK [osism.commons.known_hosts : Write scanned known_hosts entries] ***********\nTuesday 24 February 2026  00:23:14 +0000 (0:00:01.005)       0:00:21.573 ****** \n\u001b[0;33mchanged: [testbed-manager] => (item=192.168.16.5 ssh-ed25519 AAAAC3NzaC1lZDI1NTE5AAAAIDfU67w4g0t5ruLoohFtsOVj0FEJKCFA6yTUXHoTeiU+)\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item=192.168.16.5 ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABgQCUeJXF5lVxZV05KJ56am57/G4Pj6M8Ir07kqn5lXoOZXaKtAM034AYvC6FfkuxmYUsc69IbF5cYYNzWSTi7uigANnV3Gtkq8OlDelwVdV3vQDp+71dO6t30HlzvuEvk4qjeYkciisK/RE5RqVZCkqkMQIQR7TunEGspNpQyMP0Jb6CHkRMFAhPzIDv+Hcun0aWylP6WK+44TYcReX6RIOILPhPUXYV9H0ZFoC2L/hup+kkY1S85c8WaqH/OcR/XnOLwODXyRpMm2DX9kwhU0QeIeO+VqU4Jp6aXAfjSLbFbgAcm7LMAA7qdTd3liWDXJO22pgua2aAolsOe8gcFWZtMlD0AwtOUBd5c2MFrrn85aoaINKiXu/fOJxsV67SX729tb+1aXygYdUsjDxJBNHKdkmaLGpWBVgkz++0yy6DS6viWLi2nDL/3p80JUPbLv57s8zpev1x9ZMHJxbWkCjGgqjZ7HZ2WAsNklDWNgvZ03i/KiiUoaCCDjIofbR2R0E=)\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item=192.168.16.5 ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBAAEM4MNA1yZvjsAJIz6pihEIhOsw75Cu+tJPTJEbZDNCdZ3BprKWAIEwVr1UFlQTAVPSVDKMXHBd3XzGkopW84=)\u001b[0m\n\nTASK [osism.commons.known_hosts : Write scanned known_hosts entries] ***********\nTuesday 24 February 2026  00:23:15 +0000 (0:00:01.013)       0:00:22.587 ****** \n\u001b[0;33mchanged: [testbed-manager] => (item=192.168.16.10 ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBA+CLgYVB+RtOt+86BTUxoGVVBxkcCRXmZpt5JEy6gYHueijrlVosdVEchOoU/pc/10Ze4k+BS7Mg3o6u6GUV8s=)\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item=192.168.16.10 ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABgQC1zWCyl2+3CBj3nqxQhXBIhbNhR0LfsVZ8HtUXLylfIQOHN5bMATCm8HbT8KZ3HbGbluD5kffIIFrqxlSvMyuKnSFBlIwg1dSEaJztZQm1QPmZEkPMD+C+8HudVc5itevm9fwKDVWPt0lZ8zKpFxzF6Rum/bfFKJQcmPxl+zIL46mr4PbBLibBlHbAUs1ZnkZXwGXtkghKmrl8lvFskCXPllSKZcLPeM6QImpZa99TwLNhC1FnkJhLWtW25iDh53zIkoCFizujMM38wHNGIiu7rKB/CMgnX8LsKYZoSYatqHivgegBJd79m3myxBkGOsQ0YmCGGeUPRpOzabncaSPmDRsWpTafchsA7JJX6xXc4Q8+v/Zx+CNOg2oPpPnMJegXxgBAMkXeiWCaB5sbHLFmCD8eBFrOxtBeS0RqaBSHylkQBJ4q6aNJ2jIK0i2XnDVTbTrfhwkeIY9Uz5z8DuKm9HJ3x6rPJofnDSdPLdXXT3XDnj6B1riH5sVGIbKRglk=)\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item=192.168.16.10 ssh-ed25519 AAAAC3NzaC1lZDI1NTE5AAAAIC9+a2YIH9Deu7Kp5MuasTBGpKoKvAwa8BiKo1b3H76+)\u001b[0m\n\nTASK [osism.commons.known_hosts : Write scanned known_hosts entries] ***********\nTuesday 24 February 2026  00:23:16 +0000 (0:00:01.020)       0:00:23.607 ****** \n\u001b[0;33mchanged: [testbed-manager] => (item=192.168.16.11 ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABgQCbVubwVzGRukWm151/SYPu5+ZUCffbxzUGqgLxjp6PNUEvt1yzTcQ1bVHQ4y09uUCZjwDql6LL3gOIfV6fpdzXAd/FEDQt2fSoHvRXxH8BafXLEVwctwrdcoLPcJkpgl+qMgysgFgL9bHa0Rd6jP2Npur0vk3Y5ju723LOiRroRByOYiHC4XNS7DxHu6u+L/D6hVqQz5jwzP0tDXCF/+LNJ6uiaTDOd+sywTVq7ITDItNLaXIZGpsNjubQy1lMqyVZeaI5Le1/wwJDWhsrBrIyFm3Kemhecn/PVJZmPPk2OFvWp48Uyo0+NubcquE50ezJVbAmhbFG5c6HUrpVHcWLNk4LzZ6F5bRpR6V42QvBvcgwnMtzBw8QRXf96EFtzv4K5EG4Y+1Ew1sWzjWXYD/ZlA6Cc9qu5D23ebFPZIiY8o8dhBlj4iXHxO3kABxCDEJS49apyfvvKwFQ5+FjMCy88w1eO5LOwT8BXpUXcXgNcNWHidtPZ4ufzsqLYoP9Fhk=)\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item=192.168.16.11 ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBPjiHNE7pTy7o8Vd28ZoO+w4U30fFOydWvvQBQp9qayGswAO7W20tyqB2rWFbwwhAjAYAGZX2sFLF0X/ulaZkVA=)\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item=192.168.16.11 ssh-ed25519 AAAAC3NzaC1lZDI1NTE5AAAAIPlYo1pFim6htda1AoAJkI0dbcELCmM5QFZYRxrp2T8p)\u001b[0m\n\nTASK [osism.commons.known_hosts : Write scanned known_hosts entries] ***********\nTuesday 24 February 2026  00:23:17 +0000 (0:00:00.988)       0:00:24.595 ****** \n\u001b[0;33mchanged: [testbed-manager] => (item=192.168.16.12 ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBNMB0mFLquOvLDLHWMx+FbOs6ia8nAHzsLjwVkHj6z+8y/yvQBEmH3hQRj/jml5GScCC1Ncbk/HPjJCUU52aLP4=)\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item=192.168.16.12 ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABgQDCMRk074A67PfDSfiSZySlWWdBg82NzopzXCqwhSXtyjr/LVN4q3Z3VWst7PL1hjux2gaZvX6+h/RWI8ub/hOF8VMqPQ90NdJOy53UuMtsrIlovBWpDHMUa3v5FrrtAFY1Fve1POj+5kcPG69xUj/N6wLkiLab8+VqDnJLia7IsrYFnU4xShfL963gkMDvZvamLExqWws9ZZHfZc37vaYL9ntX3bcmjW76hTp+q/jMjXIxy5MRNZP+0YLq6Ig7BQb8VrrHSMx0FQgWX0UGso0BQQcaRjUl5T9yWYCJBMt3ciRZvs+Rordy878/xbYLADmgU4Vyy6RFkGxkoOQPiNMtZfMx0kLSqd8FYlysLgjvUjKFAv5uB3MAWD2Sp4CSIKI/Hbn7lzTEGIcgUopve3iif2R+oTGLje8I2g0Tqbxl53+oOWjZrPnYuCFarbJ2siPpQljjEVmsI1Y72JHFcQyRpJUp0rReRH1BpnxiPIFzr76k4hIgPNqbBi0o8zJri2k=)\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item=192.168.16.12 ssh-ed25519 AAAAC3NzaC1lZDI1NTE5AAAAID9TiTT29JB5SGhfDhxbtI4boFbWaY41+GMclNzBxg2p)\u001b[0m\n\nTASK [osism.commons.known_hosts : Write static known_hosts entries] ************\nTuesday 24 February 2026  00:23:18 +0000 (0:00:01.014)       0:00:25.609 ****** \n\u001b[0;36mskipping: [testbed-manager] => (item=testbed-node-3) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager] => (item=testbed-node-4) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager] => (item=testbed-node-5) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager] => (item=testbed-manager) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager] => (item=testbed-node-0) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager] => (item=testbed-node-1) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager] => (item=testbed-node-2) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\nTASK [osism.commons.known_hosts : Write extra known_hosts entries] *************\nTuesday 24 February 2026  00:23:18 +0000 (0:00:00.161)       0:00:25.771 ****** \n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\nTASK [osism.commons.known_hosts : Delete known_hosts entries] ******************\nTuesday 24 February 2026  00:23:18 +0000 (0:00:00.043)       0:00:25.814 ****** \n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\nTASK [osism.commons.known_hosts : Set file permissions] ************************\nTuesday 24 February 2026  00:23:18 +0000 (0:00:00.050)       0:00:25.865 ****** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;33mtestbed-manager\u001b[0m            : \u001b[0;32mok=31  \u001b[0m \u001b[0;33mchanged=15  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=3   \u001b[0m rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nTuesday 24 February 2026  00:23:19 +0000 (0:00:00.700)       0:00:26.565 ****** \n=============================================================================== \nosism.commons.known_hosts : Run ssh-keyscan for all hosts with hostname --- 5.86s\nosism.commons.known_hosts : Run ssh-keyscan for all hosts with ansible_host --- 5.05s\nosism.commons.known_hosts : Write scanned known_hosts entries ----------- 1.12s\nosism.commons.known_hosts : Write scanned known_hosts entries ----------- 1.04s\nosism.commons.known_hosts : Write scanned known_hosts entries ----------- 1.02s\nosism.commons.known_hosts : Write scanned known_hosts entries ----------- 1.02s\nosism.commons.known_hosts : Write scanned known_hosts entries ----------- 1.02s\nosism.commons.known_hosts : Write scanned known_hosts entries ----------- 1.01s\nosism.commons.known_hosts : Write scanned known_hosts entries ----------- 1.01s\nosism.commons.known_hosts : Write scanned known_hosts entries ----------- 1.01s\nosism.commons.known_hosts : Write scanned known_hosts entries ----------- 1.01s\nosism.commons.known_hosts : Write scanned known_hosts entries ----------- 1.00s\nosism.commons.known_hosts : Write scanned known_hosts entries ----------- 0.99s\nosism.commons.known_hosts : Write scanned known_hosts entries ----------- 0.98s\nosism.commons.known_hosts : Write scanned known_hosts entries ----------- 0.98s\nosism.commons.known_hosts : Write scanned known_hosts entries ----------- 0.98s\nosism.commons.known_hosts : Set file permissions ------------------------ 0.70s\nosism.commons.known_hosts : Write scanned known_hosts entries for all hosts with ansible_host --- 0.17s\nosism.commons.known_hosts : Write scanned known_hosts entries for all hosts with hostname --- 0.16s\nosism.commons.known_hosts : Write static known_hosts entries ------------ 0.16s\n+ osism apply squid\n\u001b[32m2026-02-24 00:23:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mPrepare task for execution of squid.\u001b[0m\n\u001b[32m2026-02-24 00:23:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5bf1eadf-39a3-4b5b-9eb1-63c073d32e17 (squid) was prepared for execution.\u001b[0m\n\u001b[32m2026-02-24 00:23:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task 5bf1eadf-39a3-4b5b-9eb1-63c073d32e17 (squid) has been started and output is visible here.\u001b[0m\n\nPLAY [Apply role squid] ********************************************************\n\nTASK [osism.services.squid : Include install tasks] ****************************\nTuesday 24 February 2026  00:23:35 +0000 (0:00:00.159)       0:00:00.159 ****** \n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/services/roles/squid/tasks/install-Debian-family.yml for testbed-manager\u001b[0m\n\nTASK [osism.services.squid : Install required packages] ************************\nTuesday 24 February 2026  00:23:35 +0000 (0:00:00.079)       0:00:00.238 ****** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\nTASK [osism.services.squid : Create required directories] **********************\nTuesday 24 February 2026  00:23:37 +0000 (0:00:01.349)       0:00:01.588 ****** \n\u001b[0;33mchanged: [testbed-manager] => (item=/opt/squid/configuration)\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item=/opt/squid/configuration/conf.d)\u001b[0m\n\u001b[0;32mok: [testbed-manager] => (item=/opt/squid)\u001b[0m\n\nTASK [osism.services.squid : Copy squid configuration files] *******************\nTuesday 24 February 2026  00:23:38 +0000 (0:00:01.014)       0:00:02.602 ****** \n\u001b[0;33mchanged: [testbed-manager] => (item=osism.conf)\u001b[0m\n\nTASK [osism.services.squid : Remove osism_allow_list.conf configuration file] ***\nTuesday 24 February 2026  00:23:39 +0000 (0:00:00.929)       0:00:03.532 ****** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\nTASK [osism.services.squid : Copy docker-compose.yml file] *********************\nTuesday 24 February 2026  00:23:39 +0000 (0:00:00.298)       0:00:03.830 ****** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nTASK [osism.services.squid : Manage squid service] *****************************\nTuesday 24 February 2026  00:23:40 +0000 (0:00:00.847)       0:00:04.678 ****** \n\u001b[1;30mFAILED - RETRYING: [testbed-manager]: Manage squid service (10 retries left).\u001b[0m\n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\nRUNNING HANDLER [osism.services.squid : Restart squid service] *****************\nTuesday 24 February 2026  00:24:10 +0000 (0:00:30.698)       0:00:35.376 ****** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nRUNNING HANDLER [osism.services.squid : Wait for squid service to start] *******\nTuesday 24 February 2026  00:24:22 +0000 (0:00:11.737)       0:00:47.113 ****** \nPausing for 60 seconds\n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nRUNNING HANDLER [osism.services.squid : Register that squid service was restarted] ***\nTuesday 24 February 2026  00:25:22 +0000 (0:01:00.075)       0:01:47.189 ****** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\nRUNNING HANDLER [osism.services.squid : Wait for an healthy squid service] *****\nTuesday 24 February 2026  00:25:22 +0000 (0:00:00.067)       0:01:47.257 ****** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;33mtestbed-manager\u001b[0m            : \u001b[0;32mok=11  \u001b[0m \u001b[0;33mchanged=6   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nTuesday 24 February 2026  00:25:23 +0000 (0:00:00.608)       0:01:47.865 ****** \n=============================================================================== \nosism.services.squid : Wait for squid service to start ----------------- 60.08s\nosism.services.squid : Manage squid service ---------------------------- 30.70s\nosism.services.squid : Restart squid service --------------------------- 11.74s\nosism.services.squid : Install required packages ------------------------ 1.35s\nosism.services.squid : Create required directories ---------------------- 1.01s\nosism.services.squid : Copy squid configuration files ------------------- 0.93s\nosism.services.squid : Copy docker-compose.yml file --------------------- 0.85s\nosism.services.squid : Wait for an healthy squid service ---------------- 0.61s\nosism.services.squid : Remove osism_allow_list.conf configuration file --- 0.30s\nosism.services.squid : Include install tasks ---------------------------- 0.08s\nosism.services.squid : Register that squid service was restarted -------- 0.07s\n+ [[ latest != \\l\\a\\t\\e\\s\\t ]]\n+ /opt/configuration/scripts/set-kolla-namespace.sh kolla\n+ set -e\n+ NAMESPACE=kolla\n+ sed -i 's#docker_namespace: .*#docker_namespace: kolla#g' /opt/configuration/inventory/group_vars/all/kolla.yml\n++ semver latest 9.0.0\n+ [[ -1 -lt 0 ]]\n+ [[ latest != \\l\\a\\t\\e\\s\\t ]]\n+ osism apply operator -u ubuntu -l testbed-nodes\n\u001b[32m2026-02-24 00:25:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mPrepare task for execution of operator.\u001b[0m\n\u001b[32m2026-02-24 00:25:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4ac53be9-301e-43c7-aa44-9dcc07469c96 (operator) was prepared for execution.\u001b[0m\n\u001b[32m2026-02-24 00:25:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task 4ac53be9-301e-43c7-aa44-9dcc07469c96 (operator) has been started and output is visible here.\u001b[0m\n\nPLAY [Make ssh pipelining working] *********************************************\n\nTASK [Gathering Facts] *********************************************************\nTuesday 24 February 2026  00:25:39 +0000 (0:00:00.104)       0:00:00.104 ****** \n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\nTASK [Do not require tty for all users] ****************************************\nTuesday 24 February 2026  00:25:42 +0000 (0:00:03.366)       0:00:03.471 ****** \n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\nPLAY [Apply role operator] *****************************************************\n\nTASK [osism.commons.operator : Gather variables for each operating system] *****\nTuesday 24 February 2026  00:25:43 +0000 (0:00:00.751)       0:00:04.222 ****** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [osism.commons.operator : Set operator_groups variable to default value] ***\nTuesday 24 February 2026  00:25:43 +0000 (0:00:00.155)       0:00:04.378 ****** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [osism.commons.operator : Create operator group] **************************\nTuesday 24 February 2026  00:25:43 +0000 (0:00:00.127)       0:00:04.505 ****** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\nTASK [osism.commons.operator : Create user] ************************************\nTuesday 24 February 2026  00:25:44 +0000 (0:00:00.609)       0:00:05.114 ****** \n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\nTASK [osism.commons.operator : Add user to additional groups] ******************\nTuesday 24 February 2026  00:25:45 +0000 (0:00:00.801)       0:00:05.916 ****** \n\u001b[0;33mchanged: [testbed-node-0] => (item=adm)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=adm)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=adm)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=adm)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item=adm)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=adm)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=sudo)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=sudo)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=sudo)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=sudo)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item=sudo)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=sudo)\u001b[0m\n\nTASK [osism.commons.operator : Copy user sudoers file] *************************\nTuesday 24 February 2026  00:25:46 +0000 (0:00:01.185)       0:00:07.101 ****** \n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\nTASK [osism.commons.operator : Set language variables in .bashrc configuration file] ***\nTuesday 24 February 2026  00:25:47 +0000 (0:00:01.219)       0:00:08.320 ****** \n\u001b[0;33mchanged: [testbed-node-5] => (item=export LANGUAGE=C.UTF-8)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=export LANGUAGE=C.UTF-8)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item=export LANGUAGE=C.UTF-8)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=export LANGUAGE=C.UTF-8)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=export LANGUAGE=C.UTF-8)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=export LANGUAGE=C.UTF-8)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=export LANG=C.UTF-8)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=export LANG=C.UTF-8)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item=export LANG=C.UTF-8)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=export LANG=C.UTF-8)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=export LANG=C.UTF-8)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=export LANG=C.UTF-8)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=export LC_ALL=C.UTF-8)\u001b[0m\n\u001b[1;35m[WARNING]: Module remote_tmp /root/.ansible/tmp did not exist and was created\u001b[0m\n\u001b[1;35mwith a mode of 0700, this may cause issues when running as another user. To\u001b[0m\n\u001b[1;35mavoid this, create the remote_tmp dir with the correct permissions manually\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=export LC_ALL=C.UTF-8)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=export LC_ALL=C.UTF-8)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item=export LC_ALL=C.UTF-8)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=export LC_ALL=C.UTF-8)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=export LC_ALL=C.UTF-8)\u001b[0m\n\nTASK [osism.commons.operator : Set custom environment variables in .bashrc configuration file] ***\nTuesday 24 February 2026  00:25:48 +0000 (0:00:01.222)       0:00:09.543 ****** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [osism.commons.operator : Set custom PS1 prompt in .bashrc configuration file] ***\nTuesday 24 February 2026  00:25:49 +0000 (0:00:00.147)       0:00:09.690 ****** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [osism.commons.operator : Create .ssh directory] **************************\nTuesday 24 February 2026  00:25:49 +0000 (0:00:00.177)       0:00:09.868 ****** \n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [osism.commons.operator : Check number of SSH authorized keys] ************\nTuesday 24 February 2026  00:25:49 +0000 (0:00:00.560)       0:00:10.428 ****** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [osism.commons.operator : Set ssh authorized keys] ************************\nTuesday 24 February 2026  00:25:50 +0000 (0:00:00.154)       0:00:10.583 ****** \n\u001b[0;33mchanged: [testbed-node-0] => (item=None)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=None)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item=None)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=None)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=None)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=None)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\nTASK [osism.commons.operator : Delete ssh authorized keys] *********************\nTuesday 24 February 2026  00:25:50 +0000 (0:00:00.733)       0:00:11.317 ****** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [osism.commons.operator : Set authorized GitHub accounts] *****************\nTuesday 24 February 2026  00:25:50 +0000 (0:00:00.142)       0:00:11.459 ****** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [osism.commons.operator : Delete authorized GitHub accounts] **************\nTuesday 24 February 2026  00:25:51 +0000 (0:00:00.134)       0:00:11.594 ****** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [osism.commons.operator : Set password] ***********************************\nTuesday 24 February 2026  00:25:51 +0000 (0:00:00.137)       0:00:11.732 ****** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nTASK [osism.commons.operator : Unset & lock password] **************************\nTuesday 24 February 2026  00:25:51 +0000 (0:00:00.650)       0:00:12.382 ****** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=12  \u001b[0m \u001b[0;33mchanged=8   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=7   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=12  \u001b[0m \u001b[0;33mchanged=8   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=7   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=12  \u001b[0m \u001b[0;33mchanged=8   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=7   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-3\u001b[0m             : \u001b[0;32mok=12  \u001b[0m \u001b[0;33mchanged=8   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=7   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-4\u001b[0m             : \u001b[0;32mok=12  \u001b[0m \u001b[0;33mchanged=8   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=7   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-5\u001b[0m             : \u001b[0;32mok=12  \u001b[0m \u001b[0;33mchanged=8   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=7   \u001b[0m rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nTuesday 24 February 2026  00:25:52 +0000 (0:00:00.205)       0:00:12.588 ****** \n=============================================================================== \nGathering Facts --------------------------------------------------------- 3.37s\nosism.commons.operator : Set language variables in .bashrc configuration file --- 1.22s\nosism.commons.operator : Copy user sudoers file ------------------------- 1.22s\nosism.commons.operator : Add user to additional groups ------------------ 1.19s\nosism.commons.operator : Create user ------------------------------------ 0.80s\nDo not require tty for all users ---------------------------------------- 0.75s\nosism.commons.operator : Set ssh authorized keys ------------------------ 0.73s\nosism.commons.operator : Set password ----------------------------------- 0.65s\nosism.commons.operator : Create operator group -------------------------- 0.61s\nosism.commons.operator : Create .ssh directory -------------------------- 0.56s\nosism.commons.operator : Unset & lock password -------------------------- 0.21s\nosism.commons.operator : Set custom PS1 prompt in .bashrc configuration file --- 0.18s\nosism.commons.operator : Gather variables for each operating system ----- 0.16s\nosism.commons.operator : Check number of SSH authorized keys ------------ 0.15s\nosism.commons.operator : Set custom environment variables in .bashrc configuration file --- 0.15s\nosism.commons.operator : Delete ssh authorized keys --------------------- 0.14s\nosism.commons.operator : Delete authorized GitHub accounts -------------- 0.14s\nosism.commons.operator : Set authorized GitHub accounts ----------------- 0.13s\nosism.commons.operator : Set operator_groups variable to default value --- 0.13s\n+ osism apply --environment custom facts\n\u001b[32m2026-02-24 00:25:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTrying to run play facts in environment custom\u001b[0m\n\u001b[32m2026-02-24 00:26:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mPrepare task for execution of facts.\u001b[0m\n\u001b[32m2026-02-24 00:26:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b9478762-dd77-4d27-951c-36d6391ab37d (facts) was prepared for execution.\u001b[0m\n\u001b[32m2026-02-24 00:26:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task b9478762-dd77-4d27-951c-36d6391ab37d (facts) has been started and output is visible here.\u001b[0m\n\nPLAY [Copy custom network devices fact] ****************************************\n\nTASK [Create custom facts directory] *******************************************\nTuesday 24 February 2026  00:26:08 +0000 (0:00:00.068)       0:00:00.068 ****** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [Copy fact file] **********************************************************\nTuesday 24 February 2026  00:26:09 +0000 (0:00:01.405)       0:00:01.474 ****** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\nPLAY [Copy custom ceph devices facts] ******************************************\n\nTASK [osism.commons.repository : Gather variables for each operating system] ***\nTuesday 24 February 2026  00:26:11 +0000 (0:00:01.219)       0:00:02.694 ****** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [osism.commons.repository : Set repository_default fact to default value] ***\nTuesday 24 February 2026  00:26:11 +0000 (0:00:00.088)       0:00:02.782 ****** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [osism.commons.repository : Set repositories to default] ******************\nTuesday 24 February 2026  00:26:11 +0000 (0:00:00.191)       0:00:02.974 ****** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [osism.commons.repository : Include distribution specific repository tasks] ***\nTuesday 24 February 2026  00:26:11 +0000 (0:00:00.216)       0:00:03.191 ****** \n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/repository/tasks/Ubuntu.yml for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m\n\nTASK [osism.commons.repository : Create /etc/apt/sources.list.d directory] *****\nTuesday 24 February 2026  00:26:11 +0000 (0:00:00.137)       0:00:03.328 ****** \n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [osism.commons.repository : Include tasks for Ubuntu < 24.04] *************\nTuesday 24 February 2026  00:26:12 +0000 (0:00:00.432)       0:00:03.761 ****** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [osism.commons.repository : Copy 99osism apt configuration] ***************\nTuesday 24 February 2026  00:26:12 +0000 (0:00:00.152)       0:00:03.914 ****** \n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\nTASK [osism.commons.repository : Remove sources.list file] *********************\nTuesday 24 February 2026  00:26:13 +0000 (0:00:01.077)       0:00:04.991 ****** \n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [osism.commons.repository : Copy ubuntu.sources file] *********************\nTuesday 24 February 2026  00:26:13 +0000 (0:00:00.473)       0:00:05.465 ****** \n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nTASK [osism.commons.repository : Update package cache] *************************\nTuesday 24 February 2026  00:26:15 +0000 (0:00:01.062)       0:00:06.528 ****** \n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\nTASK [Install required packages (RedHat)] **************************************\nTuesday 24 February 2026  00:26:32 +0000 (0:00:17.185)       0:00:23.713 ****** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Install required packages (Debian)] **************************************\nTuesday 24 February 2026  00:26:32 +0000 (0:00:00.095)       0:00:23.809 ****** \n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\nTASK [Create custom facts directory] *******************************************\nTuesday 24 February 2026  00:26:40 +0000 (0:00:08.327)       0:00:32.136 ****** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [Copy fact files] *********************************************************\nTuesday 24 February 2026  00:26:41 +0000 (0:00:00.466)       0:00:32.602 ****** \n\u001b[0;33mchanged: [testbed-node-4] => (item=testbed_ceph_devices)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item=testbed_ceph_devices)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=testbed_ceph_devices)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=testbed_ceph_devices_all)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item=testbed_ceph_devices_all)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=testbed_ceph_devices_all)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=testbed_ceph_osd_devices)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=testbed_ceph_osd_devices)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item=testbed_ceph_osd_devices)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=testbed_ceph_osd_devices_all)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=testbed_ceph_osd_devices_all)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item=testbed_ceph_osd_devices_all)\u001b[0m\n\nRUNNING HANDLER [osism.commons.repository : Force update of package cache] *****\nTuesday 24 February 2026  00:26:44 +0000 (0:00:03.556)       0:00:36.159 ****** \n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\nPLAY [Gather facts for all hosts] **********************************************\n\nTASK [Gathers facts about hosts] ***********************************************\nTuesday 24 February 2026  00:26:45 +0000 (0:00:01.345)       0:00:37.505 ****** \n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;32mtestbed-manager\u001b[0m            : \u001b[0;32mok=3   \u001b[0m changed=0    unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=3   \u001b[0m \u001b[0;33mchanged=2   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=3   \u001b[0m \u001b[0;33mchanged=2   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=3   \u001b[0m \u001b[0;33mchanged=2   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-3\u001b[0m             : \u001b[0;32mok=16  \u001b[0m \u001b[0;33mchanged=7   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-4\u001b[0m             : \u001b[0;32mok=16  \u001b[0m \u001b[0;33mchanged=7   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-5\u001b[0m             : \u001b[0;32mok=16  \u001b[0m \u001b[0;33mchanged=7   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nTuesday 24 February 2026  00:26:49 +0000 (0:00:03.769)       0:00:41.275 ****** \n=============================================================================== \nosism.commons.repository : Update package cache ------------------------ 17.19s\nInstall required packages (Debian) -------------------------------------- 8.33s\nGathers facts about hosts ----------------------------------------------- 3.77s\nCopy fact files --------------------------------------------------------- 3.56s\nCreate custom facts directory ------------------------------------------- 1.41s\nosism.commons.repository : Force update of package cache ---------------- 1.35s\nCopy fact file ---------------------------------------------------------- 1.22s\nosism.commons.repository : Copy 99osism apt configuration --------------- 1.08s\nosism.commons.repository : Copy ubuntu.sources file --------------------- 1.06s\nosism.commons.repository : Remove sources.list file --------------------- 0.47s\nCreate custom facts directory ------------------------------------------- 0.47s\nosism.commons.repository : Create /etc/apt/sources.list.d directory ----- 0.43s\nosism.commons.repository : Set repositories to default ------------------ 0.22s\nosism.commons.repository : Set repository_default fact to default value --- 0.19s\nosism.commons.repository : Include tasks for Ubuntu < 24.04 ------------- 0.15s\nosism.commons.repository : Include distribution specific repository tasks --- 0.14s\nInstall required packages (RedHat) -------------------------------------- 0.10s\nosism.commons.repository : Gather variables for each operating system --- 0.09s\n+ osism apply bootstrap\n\u001b[32m2026-02-24 00:27:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mPrepare task for execution of bootstrap.\u001b[0m\n\u001b[32m2026-02-24 00:27:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2991ae07-7a8f-4705-9073-594027f6153c (bootstrap) was prepared for execution.\u001b[0m\n\u001b[32m2026-02-24 00:27:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task 2991ae07-7a8f-4705-9073-594027f6153c (bootstrap) has been started and output is visible here.\u001b[0m\n\nPLAY [Group hosts based on state bootstrap] ************************************\n\nTASK [Group hosts based on state bootstrap] ************************************\nTuesday 24 February 2026  00:27:06 +0000 (0:00:00.142)       0:00:00.142 ****** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nPLAY [Gather facts for all hosts] **********************************************\n\nTASK [Gathers facts about hosts] ***********************************************\nTuesday 24 February 2026  00:27:06 +0000 (0:00:00.249)       0:00:00.392 ****** \n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\nPLAY [Gather facts for all hosts (if using --limit)] ***************************\n\nTASK [Gathers facts about hosts] ***********************************************\nTuesday 24 February 2026  00:27:10 +0000 (0:00:03.716)       0:00:04.109 ****** \n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-3) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-4) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=testbed-node-3) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-5) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=testbed-node-4) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=testbed-node-3) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-manager) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=testbed-node-5) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-0) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=testbed-node-4) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=testbed-manager) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=testbed-node-5) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-1) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=testbed-node-0) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager] => (item=testbed-node-3) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=testbed-node-1) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=testbed-manager) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=testbed-node-2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=testbed-node-0) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager] => (item=testbed-node-4) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-3) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=testbed-node-1) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-4) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager] => (item=testbed-node-5) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=testbed-node-2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-5) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=testbed-node-3) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-manager) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager] => (item=testbed-manager) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=testbed-node-4) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-0) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager] => (item=testbed-node-0) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=testbed-node-3) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-1) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=testbed-node-5) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager] => (item=testbed-node-1) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=testbed-node-4) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-manager] => (item=testbed-node-2) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=testbed-node-5) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=testbed-manager) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=testbed-node-0) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=testbed-manager) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=testbed-node-1) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=testbed-node-0) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=testbed-node-2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=testbed-node-1) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=testbed-node-2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nPLAY [Apply bootstrap roles part 1] ********************************************\n\nTASK [osism.commons.hostname : Set hostname] ***********************************\nTuesday 24 February 2026  00:27:11 +0000 (0:00:00.473)       0:00:04.583 ****** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\nTASK [osism.commons.hostname : Copy /etc/hostname] *****************************\nTuesday 24 February 2026  00:27:12 +0000 (0:00:01.188)       0:00:05.771 ****** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [osism.commons.hosts : Include type specific tasks] ***********************\nTuesday 24 February 2026  00:27:13 +0000 (0:00:01.228)       0:00:07.000 ****** \n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/hosts/tasks/type-template.yml for testbed-node-3, testbed-node-4, testbed-node-5, testbed-manager, testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [osism.commons.hosts : Copy /etc/hosts file] ******************************\nTuesday 24 February 2026  00:27:13 +0000 (0:00:00.244)       0:00:07.244 ****** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [osism.commons.proxy : Include distribution specific tasks] ***************\nTuesday 24 February 2026  00:27:15 +0000 (0:00:01.931)       0:00:09.176 ****** \n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/proxy/tasks/Debian-family.yml for testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [osism.commons.proxy : Configure proxy parameters for apt] ****************\nTuesday 24 February 2026  00:27:16 +0000 (0:00:00.315)       0:00:09.491 ****** \n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [osism.commons.proxy : Set system wide settings in environment file] ******\nTuesday 24 February 2026  00:27:17 +0000 (0:00:01.104)       0:00:10.596 ****** \n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [osism.commons.proxy : Remove system wide settings in environment file] ***\nTuesday 24 February 2026  00:27:17 +0000 (0:00:00.553)       0:00:11.149 ****** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\nTASK [osism.commons.resolvconf : Check minimum and maximum number of name servers] ***\nTuesday 24 February 2026  00:27:18 +0000 (0:00:00.489)       0:00:11.639 ****** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [osism.commons.resolvconf : Include resolvconf tasks] *********************\nTuesday 24 February 2026  00:27:18 +0000 (0:00:00.226)       0:00:11.866 ****** \n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/resolvconf/tasks/configure-resolv.yml for testbed-node-3, testbed-node-4, testbed-node-5, testbed-manager, testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [osism.commons.resolvconf : Include distribution specific installation tasks] ***\nTuesday 24 February 2026  00:27:18 +0000 (0:00:00.288)       0:00:12.154 ****** \n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/resolvconf/tasks/install-Debian-family.yml for testbed-node-3, testbed-node-4, testbed-node-5, testbed-manager, testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [osism.commons.resolvconf : Remove packages configuring /etc/resolv.conf] ***\nTuesday 24 February 2026  00:27:19 +0000 (0:00:00.440)       0:00:12.595 ****** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\nTASK [osism.commons.resolvconf : Install package systemd-resolved] *************\nTuesday 24 February 2026  00:27:20 +0000 (0:00:01.431)       0:00:14.027 ****** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [osism.commons.resolvconf : Retrieve file status of /etc/resolv.conf] *****\nTuesday 24 February 2026  00:27:20 +0000 (0:00:00.239)       0:00:14.267 ****** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [osism.commons.resolvconf : Archive existing file /etc/resolv.conf] *******\nTuesday 24 February 2026  00:27:21 +0000 (0:00:00.535)       0:00:14.802 ****** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [osism.commons.resolvconf : Link /run/systemd/resolve/stub-resolv.conf to /etc/resolv.conf] ***\nTuesday 24 February 2026  00:27:21 +0000 (0:00:00.219)       0:00:15.022 ****** \n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [osism.commons.resolvconf : Copy configuration files] *********************\nTuesday 24 February 2026  00:27:22 +0000 (0:00:00.654)       0:00:15.676 ****** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\nTASK [osism.commons.resolvconf : Start/enable systemd-resolved service] ********\nTuesday 24 February 2026  00:27:23 +0000 (0:00:01.139)       0:00:16.815 ****** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [osism.commons.resolvconf : Include distribution specific configuration tasks] ***\nTuesday 24 February 2026  00:27:24 +0000 (0:00:01.032)       0:00:17.848 ****** \n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/resolvconf/tasks/configure-Debian-family.yml for testbed-node-3, testbed-node-4, testbed-node-5, testbed-manager, testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [osism.commons.resolvconf : Restart systemd-resolved service] *************\nTuesday 24 February 2026  00:27:24 +0000 (0:00:00.347)       0:00:18.195 ****** \n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\nTASK [osism.commons.repository : Gather variables for each operating system] ***\nTuesday 24 February 2026  00:27:26 +0000 (0:00:01.231)       0:00:19.427 ****** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [osism.commons.repository : Set repository_default fact to default value] ***\nTuesday 24 February 2026  00:27:26 +0000 (0:00:00.261)       0:00:19.689 ****** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [osism.commons.repository : Set repositories to default] ******************\nTuesday 24 February 2026  00:27:26 +0000 (0:00:00.245)       0:00:19.934 ****** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [osism.commons.repository : Include distribution specific repository tasks] ***\nTuesday 24 February 2026  00:27:26 +0000 (0:00:00.229)       0:00:20.164 ****** \n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/repository/tasks/Ubuntu.yml for testbed-node-3, testbed-node-4, testbed-node-5, testbed-manager, testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [osism.commons.repository : Create /etc/apt/sources.list.d directory] *****\nTuesday 24 February 2026  00:27:27 +0000 (0:00:00.314)       0:00:20.478 ****** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [osism.commons.repository : Include tasks for Ubuntu < 24.04] *************\nTuesday 24 February 2026  00:27:27 +0000 (0:00:00.528)       0:00:21.007 ****** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [osism.commons.repository : Copy 99osism apt configuration] ***************\nTuesday 24 February 2026  00:27:27 +0000 (0:00:00.235)       0:00:21.242 ****** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\nTASK [osism.commons.repository : Remove sources.list file] *********************\nTuesday 24 February 2026  00:27:29 +0000 (0:00:01.189)       0:00:22.431 ****** \n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [osism.commons.repository : Copy ubuntu.sources file] *********************\nTuesday 24 February 2026  00:27:29 +0000 (0:00:00.569)       0:00:23.000 ****** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\nTASK [osism.commons.repository : Update package cache] *************************\nTuesday 24 February 2026  00:27:30 +0000 (0:00:01.184)       0:00:24.184 ****** \n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\nTASK [osism.services.rsyslog : Gather variables for each operating system] *****\nTuesday 24 February 2026  00:27:47 +0000 (0:00:17.061)       0:00:41.246 ****** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [osism.services.rsyslog : Set rsyslog_user variable to default value] *****\nTuesday 24 February 2026  00:27:48 +0000 (0:00:00.236)       0:00:41.483 ****** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [osism.services.rsyslog : Set rsyslog_workdir variable to default value] ***\nTuesday 24 February 2026  00:27:48 +0000 (0:00:00.233)       0:00:41.716 ****** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [osism.services.rsyslog : Include distribution specific install tasks] ****\nTuesday 24 February 2026  00:27:48 +0000 (0:00:00.210)       0:00:41.927 ****** \n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/services/roles/rsyslog/tasks/install-Debian-family.yml for testbed-node-3, testbed-node-4, testbed-node-5, testbed-manager, testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [osism.services.rsyslog : Install rsyslog package] ************************\nTuesday 24 February 2026  00:27:48 +0000 (0:00:00.275)       0:00:42.202 ****** \n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\nTASK [osism.services.rsyslog : Copy rsyslog.conf configuration file] ***********\nTuesday 24 February 2026  00:27:50 +0000 (0:00:01.470)       0:00:43.673 ****** \n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\nTASK [osism.services.rsyslog : Manage rsyslog service] *************************\nTuesday 24 February 2026  00:27:51 +0000 (0:00:01.084)       0:00:44.757 ****** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [osism.services.rsyslog : Include fluentd tasks] **************************\nTuesday 24 February 2026  00:27:52 +0000 (0:00:00.805)       0:00:45.563 ****** \n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/services/roles/rsyslog/tasks/fluentd.yml for testbed-node-3, testbed-node-4, testbed-node-5, testbed-manager, testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [osism.services.rsyslog : Forward syslog message to local fluentd daemon] ***\nTuesday 24 February 2026  00:27:52 +0000 (0:00:00.302)       0:00:45.866 ****** \n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [osism.services.rsyslog : Include additional log server tasks] ************\nTuesday 24 February 2026  00:27:53 +0000 (0:00:00.937)       0:00:46.803 ****** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [osism.services.rsyslog : Include logrotate tasks] ************************\nTuesday 24 February 2026  00:27:53 +0000 (0:00:00.251)       0:00:47.054 ****** \n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/services/roles/rsyslog/tasks/logrotate.yml for testbed-node-3, testbed-node-4, testbed-node-5, testbed-manager, testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [osism.services.rsyslog : Ensure logrotate package is installed] **********\nTuesday 24 February 2026  00:27:53 +0000 (0:00:00.332)       0:00:47.387 ****** \n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\nTASK [osism.services.rsyslog : Configure logrotate for rsyslog] ****************\nTuesday 24 February 2026  00:27:55 +0000 (0:00:01.760)       0:00:49.147 ****** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [osism.commons.systohc : Install util-linux-extra package] ****************\nTuesday 24 February 2026  00:27:56 +0000 (0:00:01.235)       0:00:50.383 ****** \n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nTASK [osism.commons.systohc : Sync hardware clock] *****************************\nTuesday 24 February 2026  00:28:08 +0000 (0:00:11.127)       0:01:01.511 ****** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\nTASK [osism.commons.configfs : Start sys-kernel-config mount] ******************\nTuesday 24 February 2026  00:28:09 +0000 (0:00:01.552)       0:01:03.064 ****** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\nTASK [osism.commons.packages : Gather variables for each operating system] *****\nTuesday 24 February 2026  00:28:10 +0000 (0:00:00.798)       0:01:03.862 ****** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [osism.commons.packages : Set required_packages_distribution variable to default value] ***\nTuesday 24 February 2026  00:28:10 +0000 (0:00:00.217)       0:01:04.080 ****** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [osism.commons.packages : Include distribution specific package tasks] ****\nTuesday 24 February 2026  00:28:10 +0000 (0:00:00.210)       0:01:04.291 ****** \n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/packages/tasks/package-Debian-family.yml for testbed-node-3, testbed-node-4, testbed-node-5, testbed-manager, testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [osism.commons.packages : Install needrestart package] ********************\nTuesday 24 February 2026  00:28:11 +0000 (0:00:00.313)       0:01:04.604 ****** \n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\nTASK [osism.commons.packages : Set needrestart mode] ***************************\nTuesday 24 February 2026  00:28:12 +0000 (0:00:01.389)       0:01:05.993 ****** \n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\nTASK [osism.commons.packages : Set apt_cache_valid_time variable to default value] ***\nTuesday 24 February 2026  00:28:13 +0000 (0:00:00.550)       0:01:06.544 ****** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [osism.commons.packages : Update package cache] ***************************\nTuesday 24 February 2026  00:28:13 +0000 (0:00:00.250)       0:01:06.794 ****** \n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\nTASK [osism.commons.packages : Download upgrade packages] **********************\nTuesday 24 February 2026  00:28:14 +0000 (0:00:01.016)       0:01:07.810 ****** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\nTASK [osism.commons.packages : Upgrade packages] *******************************\nTuesday 24 February 2026  00:28:16 +0000 (0:00:01.625)       0:01:09.436 ****** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\nTASK [osism.commons.packages : Download required packages] *********************\nTuesday 24 February 2026  00:28:18 +0000 (0:00:02.763)       0:01:12.199 ****** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [osism.commons.packages : Install required packages] **********************\nTuesday 24 February 2026  00:29:06 +0000 (0:00:48.155)       0:02:00.355 ****** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\nTASK [osism.commons.packages : Remove useless packages from the cache] *********\nTuesday 24 February 2026  00:30:32 +0000 (0:01:25.767)       0:03:26.122 ****** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\nTASK [osism.commons.packages : Remove dependencies that are no longer required] ***\nTuesday 24 February 2026  00:30:34 +0000 (0:00:02.126)       0:03:28.248 ****** \n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nTASK [osism.commons.sysctl : Include sysctl tasks] *****************************\nTuesday 24 February 2026  00:30:45 +0000 (0:00:11.002)       0:03:39.251 ****** \n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/sysctl/tasks/sysctl.yml for testbed-node-3, testbed-node-4, testbed-node-5, testbed-manager, testbed-node-0, testbed-node-1, testbed-node-2 => (item={'key': 'elasticsearch', 'value': [{'name': 'vm.max_map_count', 'value': 262144}]})\u001b[0m\n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/sysctl/tasks/sysctl.yml for testbed-node-3, testbed-node-4, testbed-node-5, testbed-manager, testbed-node-0, testbed-node-1, testbed-node-2 => (item={'key': 'rabbitmq', 'value': [{'name': 'net.ipv4.tcp_keepalive_time', 'value': 6}, {'name': 'net.ipv4.tcp_keepalive_intvl', 'value': 3}, {'name': 'net.ipv4.tcp_keepalive_probes', 'value': 3}, {'name': 'net.core.wmem_max', 'value': 16777216}, {'name': 'net.core.rmem_max', 'value': 16777216}, {'name': 'net.ipv4.tcp_fin_timeout', 'value': 20}, {'name': 'net.ipv4.tcp_tw_reuse', 'value': 1}, {'name': 'net.core.somaxconn', 'value': 4096}, {'name': 'net.ipv4.tcp_syncookies', 'value': 0}, {'name': 'net.ipv4.tcp_max_syn_backlog', 'value': 8192}]})\u001b[0m\n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/sysctl/tasks/sysctl.yml for testbed-node-3, testbed-node-4, testbed-node-5, testbed-manager, testbed-node-0, testbed-node-1, testbed-node-2 => (item={'key': 'generic', 'value': [{'name': 'vm.swappiness', 'value': 1}]})\u001b[0m\n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/sysctl/tasks/sysctl.yml for testbed-node-3, testbed-node-4, testbed-node-5, testbed-manager, testbed-node-0, testbed-node-1, testbed-node-2 => (item={'key': 'compute', 'value': [{'name': 'net.netfilter.nf_conntrack_max', 'value': 1048576}]})\u001b[0m\n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/sysctl/tasks/sysctl.yml for testbed-node-3, testbed-node-4, testbed-node-5, testbed-manager, testbed-node-0, testbed-node-1, testbed-node-2 => (item={'key': 'network', 'value': [{'name': 'net.netfilter.nf_conntrack_max', 'value': 1048576}]})\u001b[0m\n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/sysctl/tasks/sysctl.yml for testbed-node-3, testbed-node-4, testbed-node-5, testbed-manager, testbed-node-0, testbed-node-1, testbed-node-2 => (item={'key': 'k3s_node', 'value': [{'name': 'fs.inotify.max_user_instances', 'value': 1024}]})\u001b[0m\n\nTASK [osism.commons.sysctl : Set sysctl parameters on elasticsearch] ***********\nTuesday 24 February 2026  00:30:46 +0000 (0:00:00.381)       0:03:39.632 ****** \n\u001b[0;36mskipping: [testbed-node-3] => (item={'name': 'vm.max_map_count', 'value': 262144}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'name': 'vm.max_map_count', 'value': 262144}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'name': 'vm.max_map_count', 'value': 262144}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-manager] => (item={'name': 'vm.max_map_count', 'value': 262144}) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'name': 'vm.max_map_count', 'value': 262144})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'name': 'vm.max_map_count', 'value': 262144})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'name': 'vm.max_map_count', 'value': 262144})\u001b[0m\n\nTASK [osism.commons.sysctl : Set sysctl parameters on rabbitmq] ****************\nTuesday 24 February 2026  00:30:47 +0000 (0:00:00.849)       0:03:40.482 ****** \n\u001b[0;36mskipping: [testbed-node-3] => (item={'name': 'net.ipv4.tcp_keepalive_time', 'value': 6}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'name': 'net.ipv4.tcp_keepalive_intvl', 'value': 3}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'name': 'net.ipv4.tcp_keepalive_probes', 'value': 3}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'name': 'net.core.wmem_max', 'value': 16777216}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'name': 'net.core.rmem_max', 'value': 16777216}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'name': 'net.ipv4.tcp_fin_timeout', 'value': 20}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'name': 'net.ipv4.tcp_tw_reuse', 'value': 1}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'name': 'net.core.somaxconn', 'value': 4096}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'name': 'net.ipv4.tcp_keepalive_time', 'value': 6}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'name': 'net.ipv4.tcp_syncookies', 'value': 0}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'name': 'net.ipv4.tcp_keepalive_intvl', 'value': 3}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'name': 'net.ipv4.tcp_max_syn_backlog', 'value': 8192}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'name': 'net.ipv4.tcp_keepalive_probes', 'value': 3}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'name': 'net.core.wmem_max', 'value': 16777216}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'name': 'net.core.rmem_max', 'value': 16777216}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'name': 'net.ipv4.tcp_fin_timeout', 'value': 20}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'name': 'net.ipv4.tcp_tw_reuse', 'value': 1}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'name': 'net.core.somaxconn', 'value': 4096}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'name': 'net.ipv4.tcp_syncookies', 'value': 0}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'name': 'net.ipv4.tcp_keepalive_time', 'value': 6}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'name': 'net.ipv4.tcp_max_syn_backlog', 'value': 8192}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'name': 'net.ipv4.tcp_keepalive_intvl', 'value': 3}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'name': 'net.ipv4.tcp_keepalive_probes', 'value': 3}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'name': 'net.core.wmem_max', 'value': 16777216}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'name': 'net.core.rmem_max', 'value': 16777216}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'name': 'net.ipv4.tcp_fin_timeout', 'value': 20}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'name': 'net.ipv4.tcp_tw_reuse', 'value': 1}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'name': 'net.core.somaxconn', 'value': 4096}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'name': 'net.ipv4.tcp_syncookies', 'value': 0}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'name': 'net.ipv4.tcp_max_syn_backlog', 'value': 8192}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-manager] => (item={'name': 'net.ipv4.tcp_keepalive_time', 'value': 6}) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager] => (item={'name': 'net.ipv4.tcp_keepalive_intvl', 'value': 3}) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager] => (item={'name': 'net.ipv4.tcp_keepalive_probes', 'value': 3}) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager] => (item={'name': 'net.core.wmem_max', 'value': 16777216}) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager] => (item={'name': 'net.core.rmem_max', 'value': 16777216}) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager] => (item={'name': 'net.ipv4.tcp_fin_timeout', 'value': 20}) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager] => (item={'name': 'net.ipv4.tcp_tw_reuse', 'value': 1}) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager] => (item={'name': 'net.core.somaxconn', 'value': 4096}) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager] => (item={'name': 'net.ipv4.tcp_syncookies', 'value': 0}) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager] => (item={'name': 'net.ipv4.tcp_max_syn_backlog', 'value': 8192}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'name': 'net.ipv4.tcp_keepalive_time', 'value': 6})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'name': 'net.ipv4.tcp_keepalive_time', 'value': 6})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'name': 'net.ipv4.tcp_keepalive_time', 'value': 6})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'name': 'net.ipv4.tcp_keepalive_intvl', 'value': 3})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'name': 'net.ipv4.tcp_keepalive_intvl', 'value': 3})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'name': 'net.ipv4.tcp_keepalive_intvl', 'value': 3})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'name': 'net.ipv4.tcp_keepalive_probes', 'value': 3})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'name': 'net.ipv4.tcp_keepalive_probes', 'value': 3})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'name': 'net.ipv4.tcp_keepalive_probes', 'value': 3})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'name': 'net.core.wmem_max', 'value': 16777216})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'name': 'net.core.wmem_max', 'value': 16777216})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'name': 'net.core.wmem_max', 'value': 16777216})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'name': 'net.core.rmem_max', 'value': 16777216})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'name': 'net.core.rmem_max', 'value': 16777216})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'name': 'net.core.rmem_max', 'value': 16777216})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'name': 'net.ipv4.tcp_fin_timeout', 'value': 20})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'name': 'net.ipv4.tcp_fin_timeout', 'value': 20})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'name': 'net.ipv4.tcp_tw_reuse', 'value': 1})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'name': 'net.ipv4.tcp_tw_reuse', 'value': 1})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'name': 'net.core.somaxconn', 'value': 4096})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'name': 'net.core.somaxconn', 'value': 4096})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'name': 'net.ipv4.tcp_fin_timeout', 'value': 20})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'name': 'net.ipv4.tcp_syncookies', 'value': 0})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'name': 'net.ipv4.tcp_syncookies', 'value': 0})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'name': 'net.ipv4.tcp_tw_reuse', 'value': 1})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'name': 'net.ipv4.tcp_max_syn_backlog', 'value': 8192})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'name': 'net.ipv4.tcp_max_syn_backlog', 'value': 8192})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'name': 'net.core.somaxconn', 'value': 4096})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'name': 'net.ipv4.tcp_syncookies', 'value': 0})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'name': 'net.ipv4.tcp_max_syn_backlog', 'value': 8192})\u001b[0m\n\nTASK [osism.commons.sysctl : Set sysctl parameters on generic] *****************\nTuesday 24 February 2026  00:30:51 +0000 (0:00:04.767)       0:03:45.249 ****** \n\u001b[0;33mchanged: [testbed-manager] => (item={'name': 'vm.swappiness', 'value': 1})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'name': 'vm.swappiness', 'value': 1})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'name': 'vm.swappiness', 'value': 1})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'name': 'vm.swappiness', 'value': 1})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'name': 'vm.swappiness', 'value': 1})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'name': 'vm.swappiness', 'value': 1})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'name': 'vm.swappiness', 'value': 1})\u001b[0m\n\nTASK [osism.commons.sysctl : Set sysctl parameters on compute] *****************\nTuesday 24 February 2026  00:30:52 +0000 (0:00:00.563)       0:03:45.812 ****** \n\u001b[0;36mskipping: [testbed-manager] => (item={'name': 'net.netfilter.nf_conntrack_max', 'value': 1048576}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'name': 'net.netfilter.nf_conntrack_max', 'value': 1048576}) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'name': 'net.netfilter.nf_conntrack_max', 'value': 1048576}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'name': 'net.netfilter.nf_conntrack_max', 'value': 1048576}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'name': 'net.netfilter.nf_conntrack_max', 'value': 1048576})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'name': 'net.netfilter.nf_conntrack_max', 'value': 1048576})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'name': 'net.netfilter.nf_conntrack_max', 'value': 1048576})\u001b[0m\n\nTASK [osism.commons.sysctl : Set sysctl parameters on network] *****************\nTuesday 24 February 2026  00:30:52 +0000 (0:00:00.457)       0:03:46.269 ****** \n\u001b[0;36mskipping: [testbed-node-3] => (item={'name': 'net.netfilter.nf_conntrack_max', 'value': 1048576}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'name': 'net.netfilter.nf_conntrack_max', 'value': 1048576}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'name': 'net.netfilter.nf_conntrack_max', 'value': 1048576}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-manager] => (item={'name': 'net.netfilter.nf_conntrack_max', 'value': 1048576}) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'name': 'net.netfilter.nf_conntrack_max', 'value': 1048576})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'name': 'net.netfilter.nf_conntrack_max', 'value': 1048576})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'name': 'net.netfilter.nf_conntrack_max', 'value': 1048576})\u001b[0m\n\nTASK [osism.commons.sysctl : Set sysctl parameters on k3s_node] ****************\nTuesday 24 February 2026  00:30:53 +0000 (0:00:00.610)       0:03:46.880 ****** \n\u001b[0;36mskipping: [testbed-manager] => (item={'name': 'fs.inotify.max_user_instances', 'value': 1024}) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'name': 'fs.inotify.max_user_instances', 'value': 1024}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'name': 'fs.inotify.max_user_instances', 'value': 1024}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'name': 'fs.inotify.max_user_instances', 'value': 1024}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'name': 'fs.inotify.max_user_instances', 'value': 1024})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'name': 'fs.inotify.max_user_instances', 'value': 1024})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'name': 'fs.inotify.max_user_instances', 'value': 1024})\u001b[0m\n\nTASK [osism.commons.limits : Include limits tasks] *****************************\nTuesday 24 February 2026  00:30:53 +0000 (0:00:00.533)       0:03:47.413 ****** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [osism.commons.services : Populate service facts] *************************\nTuesday 24 February 2026  00:30:54 +0000 (0:00:00.270)       0:03:47.684 ****** \n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\nTASK [osism.commons.services : Check services] *********************************\nTuesday 24 February 2026  00:30:59 +0000 (0:00:05.383)       0:03:53.067 ****** \n\u001b[0;36mskipping: [testbed-node-3] => (item=nscd) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=nscd) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=nscd) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-manager] => (item=nscd) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=nscd) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=nscd) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=nscd) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [osism.commons.services : Start/enable required services] *****************\nTuesday 24 February 2026  00:30:59 +0000 (0:00:00.282)       0:03:53.350 ****** \n\u001b[0;32mok: [testbed-node-4] => (item=cron)\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item=cron)\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item=cron)\u001b[0m\n\u001b[0;32mok: [testbed-manager] => (item=cron)\u001b[0m\n\u001b[0;32mok: [testbed-node-0] => (item=cron)\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item=cron)\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item=cron)\u001b[0m\n\nTASK [osism.commons.motd : Include distribution specific configure tasks] ******\nTuesday 24 February 2026  00:31:01 +0000 (0:00:01.078)       0:03:54.429 ****** \n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/motd/tasks/configure-Debian-family.yml for testbed-node-3, testbed-node-4, testbed-node-5, testbed-manager, testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [osism.commons.motd : Remove update-motd package] *************************\nTuesday 24 February 2026  00:31:01 +0000 (0:00:00.405)       0:03:54.834 ****** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\nTASK [osism.commons.motd : Check if /etc/default/motd-news exists] *************\nTuesday 24 February 2026  00:31:02 +0000 (0:00:01.437)       0:03:56.272 ****** \n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [osism.commons.motd : Disable the dynamic motd-news service] **************\nTuesday 24 February 2026  00:31:03 +0000 (0:00:00.628)       0:03:56.901 ****** \n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\nTASK [osism.commons.motd : Get all configuration files in /etc/pam.d] **********\nTuesday 24 February 2026  00:31:04 +0000 (0:00:00.613)       0:03:57.514 ****** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [osism.commons.motd : Remove pam_motd.so rule] ****************************\nTuesday 24 February 2026  00:31:04 +0000 (0:00:00.569)       0:03:58.084 ****** \n\u001b[0;33mchanged: [testbed-manager] => (item={'path': '/etc/pam.d/sshd', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 2133, 'inode': 567, 'dev': 2049, 'nlink': 1, 'atime': 1771891429.738088, 'mtime': 1740432309.0, 'ctime': 1743685035.2598536, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'path': '/etc/pam.d/sshd', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 2133, 'inode': 567, 'dev': 2049, 'nlink': 1, 'atime': 1771891470.7879937, 'mtime': 1740432309.0, 'ctime': 1743685035.2598536, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'path': '/etc/pam.d/sshd', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 2133, 'inode': 567, 'dev': 2049, 'nlink': 1, 'atime': 1771891463.073259, 'mtime': 1740432309.0, 'ctime': 1743685035.2598536, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'path': '/etc/pam.d/sshd', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 2133, 'inode': 567, 'dev': 2049, 'nlink': 1, 'atime': 1771891450.9726114, 'mtime': 1740432309.0, 'ctime': 1743685035.2598536, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'path': '/etc/pam.d/sshd', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 2133, 'inode': 567, 'dev': 2049, 'nlink': 1, 'atime': 1771891457.1204278, 'mtime': 1740432309.0, 'ctime': 1743685035.2598536, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'path': '/etc/pam.d/sshd', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 2133, 'inode': 567, 'dev': 2049, 'nlink': 1, 'atime': 1771891464.438678, 'mtime': 1740432309.0, 'ctime': 1743685035.2598536, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'path': '/etc/pam.d/sshd', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 2133, 'inode': 567, 'dev': 2049, 'nlink': 1, 'atime': 1771891461.3313303, 'mtime': 1740432309.0, 'ctime': 1743685035.2598536, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item={'path': '/etc/pam.d/login', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 4118, 'inode': 554, 'dev': 2049, 'nlink': 1, 'atime': 1743684808.8363404, 'mtime': 1712646062.0, 'ctime': 1743685035.2588537, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'path': '/etc/pam.d/login', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 4118, 'inode': 554, 'dev': 2049, 'nlink': 1, 'atime': 1743684808.8363404, 'mtime': 1712646062.0, 'ctime': 1743685035.2588537, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'path': '/etc/pam.d/login', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 4118, 'inode': 554, 'dev': 2049, 'nlink': 1, 'atime': 1743684808.8363404, 'mtime': 1712646062.0, 'ctime': 1743685035.2588537, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'path': '/etc/pam.d/login', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 4118, 'inode': 554, 'dev': 2049, 'nlink': 1, 'atime': 1743684808.8363404, 'mtime': 1712646062.0, 'ctime': 1743685035.2588537, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'path': '/etc/pam.d/login', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 4118, 'inode': 554, 'dev': 2049, 'nlink': 1, 'atime': 1743684808.8363404, 'mtime': 1712646062.0, 'ctime': 1743685035.2588537, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'path': '/etc/pam.d/login', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 4118, 'inode': 554, 'dev': 2049, 'nlink': 1, 'atime': 1743684808.8363404, 'mtime': 1712646062.0, 'ctime': 1743685035.2588537, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'path': '/etc/pam.d/login', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 4118, 'inode': 554, 'dev': 2049, 'nlink': 1, 'atime': 1743684808.8363404, 'mtime': 1712646062.0, 'ctime': 1743685035.2588537, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})\u001b[0m\n\nTASK [osism.commons.motd : Copy motd file] *************************************\nTuesday 24 February 2026  00:31:05 +0000 (0:00:00.941)       0:03:59.026 ****** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\nTASK [osism.commons.motd : Copy issue file] ************************************\nTuesday 24 February 2026  00:31:06 +0000 (0:00:01.148)       0:04:00.175 ****** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\nTASK [osism.commons.motd : Copy issue.net file] ********************************\nTuesday 24 February 2026  00:31:07 +0000 (0:00:01.120)       0:04:01.295 ****** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\nTASK [osism.commons.motd : Configure SSH to print the motd] ********************\nTuesday 24 February 2026  00:31:09 +0000 (0:00:01.210)       0:04:02.506 ****** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [osism.commons.motd : Configure SSH to not print the motd] ****************\nTuesday 24 February 2026  00:31:09 +0000 (0:00:00.215)       0:04:02.721 ****** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\nTASK [osism.services.rng : Include distribution specific install tasks] ********\nTuesday 24 February 2026  00:31:09 +0000 (0:00:00.649)       0:04:03.370 ****** \n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/services/roles/rng/tasks/install-Debian-family.yml for testbed-node-3, testbed-node-4, testbed-node-5, testbed-manager, testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [osism.services.rng : Install rng package] ********************************\nTuesday 24 February 2026  00:31:10 +0000 (0:00:00.337)       0:04:03.707 ****** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\nTASK [osism.services.rng : Remove haveged package] *****************************\nTuesday 24 February 2026  00:31:18 +0000 (0:00:08.681)       0:04:12.389 ****** \n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\nTASK [osism.services.rng : Manage rng service] *********************************\nTuesday 24 February 2026  00:31:20 +0000 (0:00:01.231)       0:04:13.620 ****** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\nTASK [osism.commons.cleanup : Gather variables for each operating system] ******\nTuesday 24 February 2026  00:31:21 +0000 (0:00:01.107)       0:04:14.727 ****** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [osism.commons.cleanup : Set cleanup_packages_distribution variable to default value] ***\nTuesday 24 February 2026  00:31:21 +0000 (0:00:00.284)       0:04:15.011 ****** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [osism.commons.cleanup : Set cleanup_services_distribution variable to default value] ***\nTuesday 24 February 2026  00:31:21 +0000 (0:00:00.289)       0:04:15.301 ****** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [osism.commons.cleanup : Populate service facts] **************************\nTuesday 24 February 2026  00:31:22 +0000 (0:00:00.272)       0:04:15.574 ****** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\nTASK [osism.commons.cleanup : Include distribution specific timer tasks] *******\nTuesday 24 February 2026  00:31:28 +0000 (0:00:06.083)       0:04:21.657 ****** \n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/cleanup/tasks/timers-Debian-family.yml for testbed-node-3, testbed-node-4, testbed-node-5, testbed-manager, testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [osism.commons.cleanup : Disable apt-daily timers] ************************\nTuesday 24 February 2026  00:31:28 +0000 (0:00:00.406)       0:04:22.064 ****** \n\u001b[0;36mskipping: [testbed-node-3] => (item=apt-daily-upgrade) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=apt-daily) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=apt-daily-upgrade) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=apt-daily) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=apt-daily-upgrade) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=apt-daily) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-manager] => (item=apt-daily-upgrade) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager] => (item=apt-daily) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=apt-daily-upgrade) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=apt-daily) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=apt-daily-upgrade) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=apt-daily) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=apt-daily-upgrade) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=apt-daily) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [osism.commons.cleanup : Include service tasks] ***************************\nTuesday 24 February 2026  00:31:28 +0000 (0:00:00.326)       0:04:22.390 ****** \n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/cleanup/tasks/services-Debian-family.yml for testbed-node-3, testbed-node-4, testbed-node-5, testbed-manager, testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [osism.commons.cleanup : Cleanup services] ********************************\nTuesday 24 February 2026  00:31:29 +0000 (0:00:00.391)       0:04:22.781 ****** \n\u001b[0;36mskipping: [testbed-node-3] => (item=ModemManager.service) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=ModemManager.service) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=ModemManager.service) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-manager] => (item=ModemManager.service) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=ModemManager.service) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=ModemManager.service) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=ModemManager.service) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [osism.commons.cleanup : Include packages tasks] **************************\nTuesday 24 February 2026  00:31:29 +0000 (0:00:00.322)       0:04:23.104 ****** \n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/cleanup/tasks/packages-Debian-family.yml for testbed-node-3, testbed-node-4, testbed-node-5, testbed-manager, testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [osism.commons.cleanup : Cleanup installed packages] **********************\nTuesday 24 February 2026  00:31:30 +0000 (0:00:00.377)       0:04:23.481 ****** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\nTASK [osism.commons.cleanup : Remove cloudinit package] ************************\nTuesday 24 February 2026  00:32:06 +0000 (0:00:36.193)       0:04:59.675 ****** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\nTASK [osism.commons.cleanup : Uninstall unattended-upgrades package] ***********\nTuesday 24 February 2026  00:32:15 +0000 (0:00:09.248)       0:05:08.924 ****** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\nTASK [osism.commons.cleanup : Remove useless packages from the cache] **********\nTuesday 24 February 2026  00:32:24 +0000 (0:00:08.711)       0:05:17.636 ****** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\nTASK [osism.commons.cleanup : Remove dependencies that are no longer required] ***\nTuesday 24 February 2026  00:32:26 +0000 (0:00:02.094)       0:05:19.730 ****** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\nTASK [osism.commons.cleanup : Include cloudinit tasks] *************************\nTuesday 24 February 2026  00:32:33 +0000 (0:00:06.841)       0:05:26.572 ****** \n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/cleanup/tasks/cloudinit.yml for testbed-node-3, testbed-node-4, testbed-node-5, testbed-manager, testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [osism.commons.cleanup : Remove cloud-init configuration directory] *******\nTuesday 24 February 2026  00:32:33 +0000 (0:00:00.367)       0:05:26.939 ****** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\nTASK [osism.commons.timezone : Install tzdata package] *************************\nTuesday 24 February 2026  00:32:34 +0000 (0:00:00.644)       0:05:27.584 ****** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\nTASK [osism.commons.timezone : Set timezone to UTC] ****************************\nTuesday 24 February 2026  00:32:36 +0000 (0:00:02.125)       0:05:29.709 ****** \n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [osism.commons.timezone : Create /etc/adjtime file] ***********************\nTuesday 24 February 2026  00:32:37 +0000 (0:00:00.766)       0:05:30.476 ****** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [osism.commons.timezone : Ensure UTC in /etc/adjtime] *********************\nTuesday 24 February 2026  00:32:37 +0000 (0:00:00.257)       0:05:30.734 ****** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [osism.services.docker : Gather variables for each operating system] ******\nTuesday 24 February 2026  00:32:37 +0000 (0:00:00.374)       0:05:31.109 ****** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [osism.services.docker : Set docker_version variable to default value] ****\nTuesday 24 February 2026  00:32:37 +0000 (0:00:00.279)       0:05:31.388 ****** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [osism.services.docker : Set docker_cli_version variable to default value] ***\nTuesday 24 February 2026  00:32:38 +0000 (0:00:00.293)       0:05:31.682 ****** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [osism.services.docker : Print used docker version] ***********************\nTuesday 24 February 2026  00:32:38 +0000 (0:00:00.283)       0:05:31.966 ****** \n\u001b[0;32mok: [testbed-node-3] => \u001b[0m\n\u001b[0;32m  docker_version: 5:27.5.1\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => \u001b[0m\n\u001b[0;32m  docker_version: 5:27.5.1\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => \u001b[0m\n\u001b[0;32m  docker_version: 5:27.5.1\u001b[0m\n\u001b[0;32mok: [testbed-manager] => \u001b[0m\n\u001b[0;32m  docker_version: 5:27.5.1\u001b[0m\n\u001b[0;32mok: [testbed-node-0] => \u001b[0m\n\u001b[0;32m  docker_version: 5:27.5.1\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => \u001b[0m\n\u001b[0;32m  docker_version: 5:27.5.1\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => \u001b[0m\n\u001b[0;32m  docker_version: 5:27.5.1\u001b[0m\n\nTASK [osism.services.docker : Print used docker cli version] *******************\nTuesday 24 February 2026  00:32:38 +0000 (0:00:00.258)       0:05:32.224 ****** \n\u001b[0;32mok: [testbed-node-3] => \u001b[0m\n\u001b[0;32m  docker_cli_version: 5:27.5.1\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => \u001b[0m\n\u001b[0;32m  docker_cli_version: 5:27.5.1\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => \u001b[0m\n\u001b[0;32m  docker_cli_version: 5:27.5.1\u001b[0m\n\u001b[0;32mok: [testbed-manager] => \u001b[0m\n\u001b[0;32m  docker_cli_version: 5:27.5.1\u001b[0m\n\u001b[0;32mok: [testbed-node-0] => \u001b[0m\n\u001b[0;32m  docker_cli_version: 5:27.5.1\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => \u001b[0m\n\u001b[0;32m  docker_cli_version: 5:27.5.1\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => \u001b[0m\n\u001b[0;32m  docker_cli_version: 5:27.5.1\u001b[0m\n\nTASK [osism.services.docker : Include block storage tasks] *********************\nTuesday 24 February 2026  00:32:39 +0000 (0:00:00.297)       0:05:32.521 ****** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [osism.services.docker : Include zram storage tasks] **********************\nTuesday 24 February 2026  00:32:39 +0000 (0:00:00.257)       0:05:32.778 ****** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [osism.services.docker : Include docker install tasks] ********************\nTuesday 24 February 2026  00:32:39 +0000 (0:00:00.410)       0:05:33.189 ****** \n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/services/roles/docker/tasks/install-docker-Debian-family.yml for testbed-node-3, testbed-node-4, testbed-node-5, testbed-manager, testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [osism.services.docker : Remove old architecture-dependent repository] ****\nTuesday 24 February 2026  00:32:40 +0000 (0:00:00.410)       0:05:33.600 ****** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [osism.services.docker : Gather package facts] ****************************\nTuesday 24 February 2026  00:32:41 +0000 (0:00:00.828)       0:05:34.428 ****** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\nTASK [osism.services.docker : Check whether packages are installed that should not be installed] ***\nTuesday 24 February 2026  00:32:43 +0000 (0:00:02.913)       0:05:37.342 ****** \n\u001b[0;36mskipping: [testbed-node-3] => (item=containerd) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=docker.io) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=docker-engine) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=containerd) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=docker.io) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=docker-engine) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=containerd) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=docker.io) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=docker-engine) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-manager] => (item=containerd) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager] => (item=docker.io) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-manager] => (item=docker-engine) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=containerd) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=docker.io) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=docker-engine) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=containerd) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=docker.io) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=docker-engine) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=containerd) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=docker.io) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=docker-engine) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [osism.services.docker : Install apt-transport-https package] *************\nTuesday 24 February 2026  00:32:44 +0000 (0:00:00.643)       0:05:37.986 ****** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\nTASK [osism.services.docker : Add repository gpg key] **************************\nTuesday 24 February 2026  00:32:52 +0000 (0:00:07.739)       0:05:45.725 ****** \n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [osism.services.docker : Add repository] **********************************\nTuesday 24 February 2026  00:32:53 +0000 (0:00:01.122)       0:05:46.848 ****** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\nTASK [osism.services.docker : Update package cache] ****************************\nTuesday 24 February 2026  00:33:02 +0000 (0:00:09.260)       0:05:56.108 ****** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\nTASK [osism.services.docker : Pin docker package version] **********************\nTuesday 24 February 2026  00:33:06 +0000 (0:00:03.809)       0:05:59.918 ****** \n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\nTASK [osism.services.docker : Pin docker-cli package version] ******************\nTuesday 24 February 2026  00:33:08 +0000 (0:00:01.532)       0:06:01.451 ****** \n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [osism.services.docker : Unlock containerd package] ***********************\nTuesday 24 February 2026  00:33:09 +0000 (0:00:01.315)       0:06:02.767 ****** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nTASK [osism.services.docker : Install containerd package] **********************\nTuesday 24 February 2026  00:33:10 +0000 (0:00:00.794)       0:06:03.561 ****** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\nTASK [osism.services.docker : Lock containerd package] *************************\nTuesday 24 February 2026  00:33:20 +0000 (0:00:10.380)       0:06:13.942 ****** \n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [osism.services.docker : Install docker-cli package] **********************\nTuesday 24 February 2026  00:33:21 +0000 (0:00:00.951)       0:06:14.894 ****** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\nTASK [osism.services.docker : Install docker package] **************************\nTuesday 24 February 2026  00:33:32 +0000 (0:00:10.585)       0:06:25.479 ****** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [osism.services.docker : Unblock installation of python docker packages] ***\nTuesday 24 February 2026  00:33:44 +0000 (0:00:12.126)       0:06:37.606 ****** \n\u001b[0;32mok: [testbed-node-3] => (item=python3-docker)\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item=python3-docker)\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item=python3-docker)\u001b[0m\n\u001b[0;32mok: [testbed-manager] => (item=python3-docker)\u001b[0m\n\u001b[0;32mok: [testbed-node-0] => (item=python3-docker)\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item=python-docker)\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item=python3-docker)\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item=python-docker)\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item=python3-docker)\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item=python-docker)\u001b[0m\n\u001b[0;32mok: [testbed-manager] => (item=python-docker)\u001b[0m\n\u001b[0;32mok: [testbed-node-0] => (item=python-docker)\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item=python-docker)\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item=python-docker)\u001b[0m\n\nTASK [osism.services.docker : Install python3 docker package] ******************\nTuesday 24 February 2026  00:33:45 +0000 (0:00:01.194)       0:06:38.800 ****** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [osism.services.docker : Install python3 docker package from Debian Sid] ***\nTuesday 24 February 2026  00:33:45 +0000 (0:00:00.527)       0:06:39.328 ****** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\nTASK [osism.services.docker : Remove python docker packages (install python bindings from pip)] ***\nTuesday 24 February 2026  00:33:50 +0000 (0:00:04.196)       0:06:43.525 ****** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [osism.services.docker : Block installation of python docker packages (install python bindings from pip)] ***\nTuesday 24 February 2026  00:33:50 +0000 (0:00:00.685)       0:06:44.211 ****** \n\u001b[0;36mskipping: [testbed-node-3] => (item=python3-docker) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=python-docker) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=python3-docker) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=python-docker) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=python3-docker) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=python-docker) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-manager] => (item=python3-docker) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager] => (item=python-docker) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=python3-docker) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=python-docker) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=python3-docker) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=python-docker) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=python3-docker) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=python-docker) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [osism.services.docker : Install python3-pip package (install python bindings from pip)] ***\nTuesday 24 February 2026  00:33:51 +0000 (0:00:00.544)       0:06:44.755 ****** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [osism.services.docker : Install docker packages (install python bindings from pip)] ***\nTuesday 24 February 2026  00:33:51 +0000 (0:00:00.529)       0:06:45.284 ****** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [osism.services.docker : Install packages required by docker login] *******\nTuesday 24 February 2026  00:33:52 +0000 (0:00:00.497)       0:06:45.781 ****** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [osism.services.docker : Ensure that some packages are not installed] *****\nTuesday 24 February 2026  00:33:52 +0000 (0:00:00.536)       0:06:46.318 ****** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [osism.services.docker : Include config tasks] ****************************\nTuesday 24 February 2026  00:33:55 +0000 (0:00:02.246)       0:06:48.565 ****** \n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/services/roles/docker/tasks/config.yml for testbed-node-3, testbed-node-4, testbed-node-5, testbed-manager, testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [osism.services.docker : Create plugins directory] ************************\nTuesday 24 February 2026  00:33:56 +0000 (0:00:00.948)       0:06:49.513 ****** \n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [osism.services.docker : Create systemd overlay directory] ****************\nTuesday 24 February 2026  00:33:57 +0000 (0:00:00.978)       0:06:50.492 ****** \n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [osism.services.docker : Copy systemd overlay file] ***********************\nTuesday 24 February 2026  00:33:58 +0000 (0:00:01.060)       0:06:51.553 ****** \n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [osism.services.docker : Reload systemd daemon if systemd overlay file is changed] ***\nTuesday 24 February 2026  00:33:59 +0000 (0:00:01.451)       0:06:53.004 ****** \n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [osism.services.docker : Copy limits configuration file] ******************\nTuesday 24 February 2026  00:34:01 +0000 (0:00:01.481)       0:06:54.486 ****** \n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\nTASK [osism.services.docker : Copy daemon.json configuration file] *************\nTuesday 24 February 2026  00:34:02 +0000 (0:00:01.426)       0:06:55.913 ****** \n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\nTASK [osism.services.docker : Include service tasks] ***************************\nTuesday 24 February 2026  00:34:03 +0000 (0:00:01.437)       0:06:57.350 ****** \n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/services/roles/docker/tasks/service.yml for testbed-node-3, testbed-node-4, testbed-node-5, testbed-manager, testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [osism.services.docker : Reload systemd daemon] ***************************\nTuesday 24 February 2026  00:34:05 +0000 (0:00:01.132)       0:06:58.483 ****** \n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [osism.services.docker : Manage service] **********************************\nTuesday 24 February 2026  00:34:06 +0000 (0:00:01.410)       0:06:59.894 ****** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\nTASK [osism.services.docker : Manage docker socket service] ********************\nTuesday 24 February 2026  00:34:07 +0000 (0:00:01.238)       0:07:01.133 ****** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [osism.services.docker : Manage containerd service] ***********************\nTuesday 24 February 2026  00:34:08 +0000 (0:00:01.188)       0:07:02.321 ****** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [osism.services.docker : Include bootstrap tasks] *************************\nTuesday 24 February 2026  00:34:10 +0000 (0:00:01.555)       0:07:03.877 ****** \n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/services/roles/docker/tasks/bootstrap.yml for testbed-node-3, testbed-node-4, testbed-node-5, testbed-manager, testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [osism.services.docker : Flush handlers] **********************************\nTuesday 24 February 2026  00:34:11 +0000 (0:00:00.950)       0:07:04.828 ****** \n\nTASK [osism.services.docker : Flush handlers] **********************************\nTuesday 24 February 2026  00:34:11 +0000 (0:00:00.040)       0:07:04.868 ****** \n\nTASK [osism.services.docker : Flush handlers] **********************************\nTuesday 24 February 2026  00:34:11 +0000 (0:00:00.039)       0:07:04.908 ****** \n\nTASK [osism.services.docker : Flush handlers] **********************************\nTuesday 24 February 2026  00:34:11 +0000 (0:00:00.052)       0:07:04.960 ****** \n\nTASK [osism.services.docker : Flush handlers] **********************************\nTuesday 24 February 2026  00:34:11 +0000 (0:00:00.040)       0:07:05.001 ****** \n\nTASK [osism.services.docker : Flush handlers] **********************************\nTuesday 24 February 2026  00:34:11 +0000 (0:00:00.038)       0:07:05.040 ****** \n\nTASK [osism.services.docker : Flush handlers] **********************************\nTuesday 24 February 2026  00:34:11 +0000 (0:00:00.049)       0:07:05.089 ****** \n\nRUNNING HANDLER [osism.commons.repository : Force update of package cache] *****\nTuesday 24 February 2026  00:34:11 +0000 (0:00:00.039)       0:07:05.129 ****** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\nRUNNING HANDLER [osism.services.rsyslog : Restart rsyslog service] *************\nTuesday 24 February 2026  00:34:13 +0000 (0:00:01.325)       0:07:06.454 ****** \n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\nRUNNING HANDLER [osism.services.rsyslog : Restart logrotate service] ***********\nTuesday 24 February 2026  00:34:14 +0000 (0:00:01.566)       0:07:08.021 ****** \n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [osism.services.docker : Restart docker service] ***************\nTuesday 24 February 2026  00:34:15 +0000 (0:00:01.182)       0:07:09.203 ****** \n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\nRUNNING HANDLER [osism.services.docker : Wait after docker service restart] ****\nTuesday 24 February 2026  00:34:18 +0000 (0:00:02.420)       0:07:11.624 ****** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [osism.services.docker : Add user to docker group] ************************\nTuesday 24 February 2026  00:34:18 +0000 (0:00:00.095)       0:07:11.720 ****** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [osism.services.docker : Log into private registry and force re-authorization] ***\nTuesday 24 February 2026  00:34:19 +0000 (0:00:00.994)       0:07:12.715 ****** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [osism.services.docker : Include facts tasks] *****************************\nTuesday 24 February 2026  00:34:20 +0000 (0:00:00.809)       0:07:13.524 ****** \n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/services/roles/docker/tasks/facts.yml for testbed-node-3, testbed-node-4, testbed-node-5, testbed-manager, testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [osism.services.docker : Create facts directory] **************************\nTuesday 24 February 2026  00:34:20 +0000 (0:00:00.858)       0:07:14.383 ****** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [osism.services.docker : Copy docker fact files] **************************\nTuesday 24 February 2026  00:34:21 +0000 (0:00:00.804)       0:07:15.188 ****** \n\u001b[0;33mchanged: [testbed-node-3] => (item=docker_containers)\u001b[0m\n\u001b[0;32mok: [testbed-manager] => (item=docker_containers)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=docker_containers)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=docker_containers)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=docker_containers)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=docker_containers)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=docker_containers)\u001b[0m\n\u001b[0;32mok: [testbed-manager] => (item=docker_images)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=docker_images)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item=docker_images)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=docker_images)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=docker_images)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=docker_images)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=docker_images)\u001b[0m\n\nTASK [osism.commons.docker_compose : This install type is not supported] *******\nTuesday 24 February 2026  00:34:24 +0000 (0:00:02.694)       0:07:17.882 ****** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [osism.commons.docker_compose : Include distribution specific install tasks] ***\nTuesday 24 February 2026  00:34:24 +0000 (0:00:00.514)       0:07:18.397 ****** \n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/docker_compose/tasks/install-Debian-family.yml for testbed-node-3, testbed-node-4, testbed-node-5, testbed-manager, testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [osism.commons.docker_compose : Remove docker-compose apt preferences file] ***\nTuesday 24 February 2026  00:34:25 +0000 (0:00:00.822)       0:07:19.219 ****** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [osism.commons.docker_compose : Get checksum of docker-compose file] ******\nTuesday 24 February 2026  00:34:26 +0000 (0:00:00.888)       0:07:20.108 ****** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [osism.commons.docker_compose : Remove docker-compose binary] *************\nTuesday 24 February 2026  00:34:27 +0000 (0:00:01.100)       0:07:21.208 ****** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [osism.commons.docker_compose : Uninstall docker-compose package] *********\nTuesday 24 February 2026  00:34:28 +0000 (0:00:00.524)       0:07:21.733 ****** \n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\nTASK [osism.commons.docker_compose : Copy docker-compose script] ***************\nTuesday 24 February 2026  00:34:29 +0000 (0:00:01.522)       0:07:23.256 ****** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [osism.commons.docker_compose : Install docker-compose-plugin package] ****\nTuesday 24 February 2026  00:34:30 +0000 (0:00:00.510)       0:07:23.766 ****** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\nTASK [osism.commons.docker_compose : Copy osism.target systemd file] ***********\nTuesday 24 February 2026  00:34:38 +0000 (0:00:08.545)       0:07:32.312 ****** \n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\nTASK [osism.commons.docker_compose : Enable osism.target] **********************\nTuesday 24 February 2026  00:34:40 +0000 (0:00:01.374)       0:07:33.686 ****** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [osism.commons.docker_compose : Copy docker-compose systemd unit file] ****\nTuesday 24 February 2026  00:34:41 +0000 (0:00:01.718)       0:07:35.405 ****** \n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [osism.commons.facts : Create custom facts directory] *********************\nTuesday 24 February 2026  00:34:43 +0000 (0:00:01.696)       0:07:37.101 ****** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [osism.commons.facts : Copy fact files] ***********************************\nTuesday 24 February 2026  00:34:44 +0000 (0:00:01.089)       0:07:38.191 ****** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [osism.services.chrony : Check minimum and maximum number of servers] *****\nTuesday 24 February 2026  00:34:45 +0000 (0:00:00.817)       0:07:39.008 ****** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [osism.services.chrony : Gather variables for each operating system] ******\nTuesday 24 February 2026  00:34:46 +0000 (0:00:00.514)       0:07:39.523 ****** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [osism.services.chrony : Set chrony_conf_file variable to default value] ***\nTuesday 24 February 2026  00:34:46 +0000 (0:00:00.557)       0:07:40.080 ****** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [osism.services.chrony : Set chrony_key_file variable to default value] ***\nTuesday 24 February 2026  00:34:47 +0000 (0:00:00.716)       0:07:40.796 ****** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [osism.services.chrony : Populate service facts] **************************\nTuesday 24 February 2026  00:34:47 +0000 (0:00:00.520)       0:07:41.317 ****** \n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\nTASK [osism.services.chrony : Manage timesyncd service] ************************\nTuesday 24 February 2026  00:34:53 +0000 (0:00:05.588)       0:07:46.906 ****** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [osism.services.chrony : Include distribution specific install tasks] *****\nTuesday 24 February 2026  00:34:54 +0000 (0:00:00.612)       0:07:47.518 ****** \n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/services/roles/chrony/tasks/install-Debian-family.yml for testbed-node-3, testbed-node-4, testbed-node-5, testbed-manager, testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [osism.services.chrony : Install package] *********************************\nTuesday 24 February 2026  00:34:55 +0000 (0:00:01.241)       0:07:48.760 ****** \n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\nTASK [osism.services.chrony : Manage chrony service] ***************************\nTuesday 24 February 2026  00:34:57 +0000 (0:00:02.207)       0:07:50.967 ****** \n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [osism.services.chrony : Check if configuration file exists] **************\nTuesday 24 February 2026  00:34:58 +0000 (0:00:01.225)       0:07:52.192 ****** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [osism.services.chrony : Copy configuration file] *************************\nTuesday 24 February 2026  00:34:59 +0000 (0:00:00.913)       0:07:53.106 ****** \n\u001b[0;33mchanged: [testbed-node-4] => (item=/usr/share/ansible/collections/ansible_collections/osism/services/roles/chrony/templates/chrony.conf.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item=/usr/share/ansible/collections/ansible_collections/osism/services/roles/chrony/templates/chrony.conf.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item=/usr/share/ansible/collections/ansible_collections/osism/services/roles/chrony/templates/chrony.conf.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=/usr/share/ansible/collections/ansible_collections/osism/services/roles/chrony/templates/chrony.conf.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=/usr/share/ansible/collections/ansible_collections/osism/services/roles/chrony/templates/chrony.conf.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=/usr/share/ansible/collections/ansible_collections/osism/services/roles/chrony/templates/chrony.conf.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=/usr/share/ansible/collections/ansible_collections/osism/services/roles/chrony/templates/chrony.conf.j2)\u001b[0m\n\nTASK [osism.services.lldpd : Include distribution specific install tasks] ******\nTuesday 24 February 2026  00:35:01 +0000 (0:00:01.978)       0:07:55.084 ****** \n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/services/roles/lldpd/tasks/install-Debian-family.yml for testbed-node-3, testbed-node-4, testbed-node-5, testbed-manager, testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [osism.services.lldpd : Install lldpd package] ****************************\nTuesday 24 February 2026  00:35:02 +0000 (0:00:00.857)       0:07:55.942 ****** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\nTASK [osism.services.lldpd : Manage lldpd service] *****************************\nTuesday 24 February 2026  00:35:12 +0000 (0:00:10.392)       0:08:06.335 ****** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [osism.commons.docker_compose : Reload systemd daemon] *********\nTuesday 24 February 2026  00:35:14 +0000 (0:00:01.864)       0:08:08.199 ****** \n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\nRUNNING HANDLER [osism.services.chrony : Restart chrony service] ***************\nTuesday 24 February 2026  00:35:16 +0000 (0:00:01.297)       0:08:09.496 ****** \n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nPLAY [Apply bootstrap role part 2] *********************************************\n\nTASK [Include hardening role] **************************************************\nTuesday 24 February 2026  00:35:17 +0000 (0:00:01.360)       0:08:10.857 ****** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nPLAY [Apply bootstrap roles part 3] ********************************************\n\nTASK [osism.services.journald : Copy configuration file] ***********************\nTuesday 24 February 2026  00:35:17 +0000 (0:00:00.491)       0:08:11.349 ****** \n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [osism.services.journald : Manage journald service] ***********************\nTuesday 24 February 2026  00:35:19 +0000 (0:00:01.335)       0:08:12.684 ****** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [Include auditd role] *****************************************************\nTuesday 24 February 2026  00:35:20 +0000 (0:00:01.375)       0:08:14.060 ****** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [Include smartd role] *****************************************************\nTuesday 24 February 2026  00:35:21 +0000 (0:00:00.586)       0:08:14.646 ****** \n\u001b[0;36mincluded: osism.services.smartd for testbed-node-3, testbed-node-4, testbed-node-5, testbed-manager, testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [osism.services.smartd : Include distribution specific install tasks] *****\nTuesday 24 February 2026  00:35:22 +0000 (0:00:00.832)       0:08:15.479 ****** \n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/services/roles/smartd/tasks/install-Debian-family.yml for testbed-node-3, testbed-node-4, testbed-node-5, testbed-manager, testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [osism.services.smartd : Install smartmontools package] *******************\nTuesday 24 February 2026  00:35:22 +0000 (0:00:00.779)       0:08:16.258 ****** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\nTASK [osism.services.smartd : Create /var/log/smartd directory] ****************\nTuesday 24 February 2026  00:35:32 +0000 (0:00:09.728)       0:08:25.986 ****** \n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [osism.services.smartd : Copy smartmontools configuration file] ***********\nTuesday 24 February 2026  00:35:33 +0000 (0:00:00.874)       0:08:26.861 ****** \n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\nTASK [osism.services.smartd : Manage smartd service] ***************************\nTuesday 24 February 2026  00:35:34 +0000 (0:00:01.387)       0:08:28.248 ****** \n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [osism.services.journald : Restart journald service] ***********\nTuesday 24 February 2026  00:35:36 +0000 (0:00:02.020)       0:08:30.269 ****** \n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [osism.services.smartd : Restart smartd service] ***************\nTuesday 24 February 2026  00:35:38 +0000 (0:00:01.227)       0:08:31.496 ****** \n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\nPLAY [Set state bootstrap] *****************************************************\n\nTASK [Set osism.bootstrap.status fact] *****************************************\nTuesday 24 February 2026  00:35:39 +0000 (0:00:01.125)       0:08:32.621 ****** \n\u001b[0;36mincluded: osism.commons.state for testbed-node-3, testbed-node-4, testbed-node-5, testbed-manager, testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [osism.commons.state : Create custom facts directory] *********************\nTuesday 24 February 2026  00:35:40 +0000 (0:00:00.992)       0:08:33.614 ****** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [osism.commons.state : Write state into file] *****************************\nTuesday 24 February 2026  00:35:41 +0000 (0:00:00.885)       0:08:34.499 ****** \n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [Set osism.bootstrap.timestamp fact] **************************************\nTuesday 24 February 2026  00:35:42 +0000 (0:00:01.117)       0:08:35.617 ****** \n\u001b[0;36mincluded: osism.commons.state for testbed-node-3, testbed-node-4, testbed-node-5, testbed-manager, testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [osism.commons.state : Create custom facts directory] *********************\nTuesday 24 February 2026  00:35:43 +0000 (0:00:01.032)       0:08:36.649 ****** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [osism.commons.state : Write state into file] *****************************\nTuesday 24 February 2026  00:35:44 +0000 (0:00:00.849)       0:08:37.498 ****** \n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;33mtestbed-manager\u001b[0m            : \u001b[0;32mok=168 \u001b[0m \u001b[0;33mchanged=40  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=42  \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=177 \u001b[0m \u001b[0;33mchanged=69  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=36  \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=177 \u001b[0m \u001b[0;33mchanged=69  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=36  \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=177 \u001b[0m \u001b[0;33mchanged=69  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=36  \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-3\u001b[0m             : \u001b[0;32mok=175 \u001b[0m \u001b[0;33mchanged=65  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=38  \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-4\u001b[0m             : \u001b[0;32mok=175 \u001b[0m \u001b[0;33mchanged=65  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=37  \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-5\u001b[0m             : \u001b[0;32mok=175 \u001b[0m \u001b[0;33mchanged=65  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=37  \u001b[0m rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nTuesday 24 February 2026  00:35:45 +0000 (0:00:01.108)       0:08:38.607 ****** \n=============================================================================== \nosism.commons.packages : Install required packages --------------------- 85.77s\nosism.commons.packages : Download required packages -------------------- 48.16s\nosism.commons.cleanup : Cleanup installed packages --------------------- 36.19s\nosism.commons.repository : Update package cache ------------------------ 17.06s\nosism.services.docker : Install docker package ------------------------- 12.13s\nosism.commons.systohc : Install util-linux-extra package --------------- 11.13s\nosism.commons.packages : Remove dependencies that are no longer required -- 11.00s\nosism.services.docker : Install docker-cli package --------------------- 10.59s\nosism.services.lldpd : Install lldpd package --------------------------- 10.39s\nosism.services.docker : Install containerd package --------------------- 10.38s\nosism.services.smartd : Install smartmontools package ------------------- 9.73s\nosism.services.docker : Add repository ---------------------------------- 9.26s\nosism.commons.cleanup : Remove cloudinit package ------------------------ 9.25s\nosism.commons.cleanup : Uninstall unattended-upgrades package ----------- 8.71s\nosism.services.rng : Install rng package -------------------------------- 8.68s\nosism.commons.docker_compose : Install docker-compose-plugin package ---- 8.55s\nosism.services.docker : Install apt-transport-https package ------------- 7.74s\nosism.commons.cleanup : Remove dependencies that are no longer required --- 6.84s\nosism.commons.cleanup : Populate service facts -------------------------- 6.08s\nosism.services.chrony : Populate service facts -------------------------- 5.59s\n+ osism apply fail2ban\n\u001b[32m2026-02-24 00:35:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mPrepare task for execution of fail2ban.\u001b[0m\n\u001b[32m2026-02-24 00:35:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c6792215-2f8a-4696-9f5d-fb0127fc1bb6 (fail2ban) was prepared for execution.\u001b[0m\n\u001b[32m2026-02-24 00:35:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task c6792215-2f8a-4696-9f5d-fb0127fc1bb6 (fail2ban) has been started and output is visible here.\u001b[0m\n\nPLAY [Apply role fail2ban] *****************************************************\n\nTASK [osism.services.fail2ban : Include distribution specific install tasks] ***\nTuesday 24 February 2026  00:36:02 +0000 (0:00:00.239)       0:00:00.239 ****** \n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/services/roles/fail2ban/tasks/install-Debian-family.yml for testbed-manager, testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m\n\nTASK [osism.services.fail2ban : Install fail2ban package] **********************\nTuesday 24 February 2026  00:36:03 +0000 (0:00:01.180)       0:00:01.419 ****** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\nTASK [osism.services.fail2ban : Copy configuration files] **********************\nTuesday 24 February 2026  00:36:15 +0000 (0:00:11.893)       0:00:13.313 ****** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nTASK [osism.services.fail2ban : Manage fail2ban service] ***********************\nTuesday 24 February 2026  00:36:17 +0000 (0:00:01.509)       0:00:14.822 ****** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [osism.services.fail2ban : Reload fail2ban configuration] *****************\nTuesday 24 February 2026  00:36:18 +0000 (0:00:01.538)       0:00:16.361 ****** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;33mtestbed-manager\u001b[0m            : \u001b[0;32mok=5   \u001b[0m \u001b[0;33mchanged=3   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=5   \u001b[0m \u001b[0;33mchanged=3   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=5   \u001b[0m \u001b[0;33mchanged=3   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=5   \u001b[0m \u001b[0;33mchanged=3   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-3\u001b[0m             : \u001b[0;32mok=5   \u001b[0m \u001b[0;33mchanged=3   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-4\u001b[0m             : \u001b[0;32mok=5   \u001b[0m \u001b[0;33mchanged=3   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-5\u001b[0m             : \u001b[0;32mok=5   \u001b[0m \u001b[0;33mchanged=3   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nTuesday 24 February 2026  00:36:20 +0000 (0:00:01.717)       0:00:18.079 ****** \n=============================================================================== \nosism.services.fail2ban : Install fail2ban package --------------------- 11.89s\nosism.services.fail2ban : Reload fail2ban configuration ----------------- 1.72s\nosism.services.fail2ban : Manage fail2ban service ----------------------- 1.54s\nosism.services.fail2ban : Copy configuration files ---------------------- 1.51s\nosism.services.fail2ban : Include distribution specific install tasks --- 1.18s\n+ [[ -e /etc/redhat-release ]]\n+ osism apply network\n\u001b[32m2026-02-24 00:36:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mPrepare task for execution of network.\u001b[0m\n\u001b[32m2026-02-24 00:36:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ee38c4cf-b985-4420-8dba-d9d7697ff09e (network) was prepared for execution.\u001b[0m\n\u001b[32m2026-02-24 00:36:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task ee38c4cf-b985-4420-8dba-d9d7697ff09e (network) has been started and output is visible here.\u001b[0m\n\nPLAY [Apply role network] ******************************************************\n\nTASK [osism.commons.network : Gather variables for each operating system] ******\nTuesday 24 February 2026  00:36:37 +0000 (0:00:00.257)       0:00:00.257 ****** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [osism.commons.network : Include type specific tasks] *********************\nTuesday 24 February 2026  00:36:38 +0000 (0:00:00.693)       0:00:00.951 ****** \n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/network/tasks/netplan-Debian-family.yml for testbed-manager, testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m\n\nTASK [osism.commons.network : Install required packages] ***********************\nTuesday 24 February 2026  00:36:39 +0000 (0:00:01.208)       0:00:02.159 ****** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\nTASK [osism.commons.network : Remove ifupdown package] *************************\nTuesday 24 February 2026  00:36:41 +0000 (0:00:02.126)       0:00:04.286 ****** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [osism.commons.network : Create required directories] *********************\nTuesday 24 February 2026  00:36:43 +0000 (0:00:01.850)       0:00:06.136 ****** \n\u001b[0;32mok: [testbed-manager] => (item=/etc/netplan)\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item=/etc/netplan)\u001b[0m\n\u001b[0;32mok: [testbed-node-0] => (item=/etc/netplan)\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item=/etc/netplan)\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item=/etc/netplan)\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item=/etc/netplan)\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item=/etc/netplan)\u001b[0m\n\nTASK [osism.commons.network : Prepare netplan configuration template] **********\nTuesday 24 February 2026  00:36:44 +0000 (0:00:01.015)       0:00:07.152 ****** \n\u001b[0;32mok: [testbed-node-1 -> localhost]\u001b[0m\n\u001b[0;32mok: [testbed-node-2 -> localhost]\u001b[0m\n\u001b[0;32mok: [testbed-manager -> localhost]\u001b[0m\n\u001b[0;32mok: [testbed-node-5 -> localhost]\u001b[0m\n\u001b[0;32mok: [testbed-node-0 -> localhost]\u001b[0m\n\u001b[0;32mok: [testbed-node-3 -> localhost]\u001b[0m\n\u001b[0;32mok: [testbed-node-4 -> localhost]\u001b[0m\n\nTASK [osism.commons.network : Copy netplan configuration] **********************\nTuesday 24 February 2026  00:36:48 +0000 (0:00:03.559)       0:00:10.712 ****** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nTASK [osism.commons.network : Remove netplan configuration template] ***********\nTuesday 24 February 2026  00:36:50 +0000 (0:00:01.654)       0:00:12.366 ****** \n\u001b[0;32mok: [testbed-node-0 -> localhost]\u001b[0m\n\u001b[0;32mok: [testbed-manager -> localhost]\u001b[0m\n\u001b[0;32mok: [testbed-node-2 -> localhost]\u001b[0m\n\u001b[0;32mok: [testbed-node-1 -> localhost]\u001b[0m\n\u001b[0;32mok: [testbed-node-4 -> localhost]\u001b[0m\n\u001b[0;32mok: [testbed-node-3 -> localhost]\u001b[0m\n\u001b[0;32mok: [testbed-node-5 -> localhost]\u001b[0m\n\nTASK [osism.commons.network : Check if path for interface file exists] *********\nTuesday 24 February 2026  00:36:52 +0000 (0:00:01.949)       0:00:14.316 ****** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [osism.commons.network : Copy interfaces file] ****************************\nTuesday 24 February 2026  00:36:53 +0000 (0:00:01.137)       0:00:15.454 ****** \n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [osism.commons.network : Install package networkd-dispatcher] *************\nTuesday 24 February 2026  00:36:53 +0000 (0:00:00.714)       0:00:16.168 ****** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [osism.commons.network : Copy dispatcher scripts] *************************\nTuesday 24 February 2026  00:36:56 +0000 (0:00:02.427)       0:00:18.596 ****** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item={'dest': 'routable.d/iptables.sh', 'src': '/opt/configuration/network/iptables.sh'})\u001b[0m\n\nTASK [osism.commons.network : Manage service networkd-dispatcher] **************\nTuesday 24 February 2026  00:36:57 +0000 (0:00:00.912)       0:00:19.508 ****** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nTASK [osism.commons.network : Include cleanup tasks] ***************************\nTuesday 24 February 2026  00:36:58 +0000 (0:00:01.693)       0:00:21.201 ****** \n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/network/tasks/cleanup-netplan.yml for testbed-manager, testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m\n\nTASK [osism.commons.network : List existing configuration files] ***************\nTuesday 24 February 2026  00:37:00 +0000 (0:00:01.280)       0:00:22.482 ****** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [osism.commons.network : Set network_configured_files fact] ***************\nTuesday 24 February 2026  00:37:01 +0000 (0:00:01.169)       0:00:23.651 ****** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [osism.commons.network : Remove unused configuration files] ***************\nTuesday 24 February 2026  00:37:02 +0000 (0:00:00.706)       0:00:24.357 ****** \n\u001b[0;36mskipping: [testbed-manager] => (item=/etc/netplan/01-osism.yaml) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=/etc/netplan/01-osism.yaml) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=/etc/netplan/01-osism.yaml) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=/etc/netplan/01-osism.yaml) \u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item=/etc/netplan/50-cloud-init.yaml)\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=/etc/netplan/01-osism.yaml) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=/etc/netplan/01-osism.yaml) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=/etc/netplan/50-cloud-init.yaml)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=/etc/netplan/50-cloud-init.yaml)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=/etc/netplan/50-cloud-init.yaml)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item=/etc/netplan/50-cloud-init.yaml)\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=/etc/netplan/01-osism.yaml) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=/etc/netplan/50-cloud-init.yaml)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=/etc/netplan/50-cloud-init.yaml)\u001b[0m\n\nTASK [osism.commons.network : Include dummy interfaces] ************************\nTuesday 24 February 2026  00:37:03 +0000 (0:00:01.222)       0:00:25.579 ****** \n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [osism.commons.network : Include vxlan interfaces] ************************\nTuesday 24 February 2026  00:37:03 +0000 (0:00:00.603)       0:00:26.183 ****** \n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/network/tasks/vxlan-interfaces.yml for testbed-node-4, testbed-manager, testbed-node-1, testbed-node-0, testbed-node-2, testbed-node-3, testbed-node-5\u001b[0m\n\nTASK [osism.commons.network : Create systemd networkd netdev files] ************\nTuesday 24 February 2026  00:37:08 +0000 (0:00:04.481)       0:00:30.664 ****** \n\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'vxlan0', 'value': {'addresses': ['192.168.112.5/20'], 'dests': ['192.168.16.10', '192.168.16.11', '192.168.16.12', '192.168.16.13', '192.168.16.14', '192.168.16.15'], 'local_ip': '192.168.16.5', 'mtu': 1350, 'vni': 42}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'vxlan0', 'value': {'addresses': [], 'dests': ['192.168.16.11', '192.168.16.12', '192.168.16.13', '192.168.16.14', '192.168.16.15', '192.168.16.5'], 'local_ip': '192.168.16.10', 'mtu': 1350, 'vni': 42}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'vxlan0', 'value': {'addresses': [], 'dests': ['192.168.16.10', '192.168.16.11', '192.168.16.13', '192.168.16.14', '192.168.16.15', '192.168.16.5'], 'local_ip': '192.168.16.12', 'mtu': 1350, 'vni': 42}})\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'vxlan1', 'value': {'addresses': ['192.168.128.5/20'], 'dests': ['192.168.16.10', '192.168.16.11', '192.168.16.12', '192.168.16.13', '192.168.16.14', '192.168.16.15'], 'local_ip': '192.168.16.5', 'mtu': 1350, 'vni': 23}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'vxlan0', 'value': {'addresses': [], 'dests': ['192.168.16.10', '192.168.16.12', '192.168.16.13', '192.168.16.14', '192.168.16.15', '192.168.16.5'], 'local_ip': '192.168.16.11', 'mtu': 1350, 'vni': 42}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'vxlan0', 'value': {'addresses': [], 'dests': ['192.168.16.10', '192.168.16.11', '192.168.16.12', '192.168.16.13', '192.168.16.14', '192.168.16.5'], 'local_ip': '192.168.16.15', 'mtu': 1350, 'vni': 42}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'vxlan0', 'value': {'addresses': [], 'dests': ['192.168.16.10', '192.168.16.11', '192.168.16.12', '192.168.16.14', '192.168.16.15', '192.168.16.5'], 'local_ip': '192.168.16.13', 'mtu': 1350, 'vni': 42}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'vxlan1', 'value': {'addresses': ['192.168.128.10/20'], 'dests': ['192.168.16.11', '192.168.16.12', '192.168.16.13', '192.168.16.14', '192.168.16.15', '192.168.16.5'], 'local_ip': '192.168.16.10', 'mtu': 1350, 'vni': 23}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'vxlan0', 'value': {'addresses': [], 'dests': ['192.168.16.10', '192.168.16.11', '192.168.16.12', '192.168.16.13', '192.168.16.15', '192.168.16.5'], 'local_ip': '192.168.16.14', 'mtu': 1350, 'vni': 42}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'vxlan1', 'value': {'addresses': ['192.168.128.12/20'], 'dests': ['192.168.16.10', '192.168.16.11', '192.168.16.13', '192.168.16.14', '192.168.16.15', '192.168.16.5'], 'local_ip': '192.168.16.12', 'mtu': 1350, 'vni': 23}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'vxlan1', 'value': {'addresses': ['192.168.128.11/20'], 'dests': ['192.168.16.10', '192.168.16.12', '192.168.16.13', '192.168.16.14', '192.168.16.15', '192.168.16.5'], 'local_ip': '192.168.16.11', 'mtu': 1350, 'vni': 23}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'vxlan1', 'value': {'addresses': ['192.168.128.15/20'], 'dests': ['192.168.16.10', '192.168.16.11', '192.168.16.12', '192.168.16.13', '192.168.16.14', '192.168.16.5'], 'local_ip': '192.168.16.15', 'mtu': 1350, 'vni': 23}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'vxlan1', 'value': {'addresses': ['192.168.128.13/20'], 'dests': ['192.168.16.10', '192.168.16.11', '192.168.16.12', '192.168.16.14', '192.168.16.15', '192.168.16.5'], 'local_ip': '192.168.16.13', 'mtu': 1350, 'vni': 23}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'vxlan1', 'value': {'addresses': ['192.168.128.14/20'], 'dests': ['192.168.16.10', '192.168.16.11', '192.168.16.12', '192.168.16.13', '192.168.16.15', '192.168.16.5'], 'local_ip': '192.168.16.14', 'mtu': 1350, 'vni': 23}})\u001b[0m\n\nTASK [osism.commons.network : Create systemd networkd network files] ***********\nTuesday 24 February 2026  00:37:13 +0000 (0:00:05.206)       0:00:35.871 ****** \n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'vxlan0', 'value': {'addresses': [], 'dests': ['192.168.16.10', '192.168.16.11', '192.168.16.13', '192.168.16.14', '192.168.16.15', '192.168.16.5'], 'local_ip': '192.168.16.12', 'mtu': 1350, 'vni': 42}})\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'vxlan0', 'value': {'addresses': ['192.168.112.5/20'], 'dests': ['192.168.16.10', '192.168.16.11', '192.168.16.12', '192.168.16.13', '192.168.16.14', '192.168.16.15'], 'local_ip': '192.168.16.5', 'mtu': 1350, 'vni': 42}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'vxlan0', 'value': {'addresses': [], 'dests': ['192.168.16.11', '192.168.16.12', '192.168.16.13', '192.168.16.14', '192.168.16.15', '192.168.16.5'], 'local_ip': '192.168.16.10', 'mtu': 1350, 'vni': 42}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'vxlan0', 'value': {'addresses': [], 'dests': ['192.168.16.10', '192.168.16.12', '192.168.16.13', '192.168.16.14', '192.168.16.15', '192.168.16.5'], 'local_ip': '192.168.16.11', 'mtu': 1350, 'vni': 42}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'vxlan0', 'value': {'addresses': [], 'dests': ['192.168.16.10', '192.168.16.11', '192.168.16.12', '192.168.16.13', '192.168.16.14', '192.168.16.5'], 'local_ip': '192.168.16.15', 'mtu': 1350, 'vni': 42}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'vxlan0', 'value': {'addresses': [], 'dests': ['192.168.16.10', '192.168.16.11', '192.168.16.12', '192.168.16.14', '192.168.16.15', '192.168.16.5'], 'local_ip': '192.168.16.13', 'mtu': 1350, 'vni': 42}})\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'vxlan1', 'value': {'addresses': ['192.168.128.5/20'], 'dests': ['192.168.16.10', '192.168.16.11', '192.168.16.12', '192.168.16.13', '192.168.16.14', '192.168.16.15'], 'local_ip': '192.168.16.5', 'mtu': 1350, 'vni': 23}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'vxlan1', 'value': {'addresses': ['192.168.128.12/20'], 'dests': ['192.168.16.10', '192.168.16.11', '192.168.16.13', '192.168.16.14', '192.168.16.15', '192.168.16.5'], 'local_ip': '192.168.16.12', 'mtu': 1350, 'vni': 23}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'vxlan0', 'value': {'addresses': [], 'dests': ['192.168.16.10', '192.168.16.11', '192.168.16.12', '192.168.16.13', '192.168.16.15', '192.168.16.5'], 'local_ip': '192.168.16.14', 'mtu': 1350, 'vni': 42}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'vxlan1', 'value': {'addresses': ['192.168.128.10/20'], 'dests': ['192.168.16.11', '192.168.16.12', '192.168.16.13', '192.168.16.14', '192.168.16.15', '192.168.16.5'], 'local_ip': '192.168.16.10', 'mtu': 1350, 'vni': 23}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'vxlan1', 'value': {'addresses': ['192.168.128.11/20'], 'dests': ['192.168.16.10', '192.168.16.12', '192.168.16.13', '192.168.16.14', '192.168.16.15', '192.168.16.5'], 'local_ip': '192.168.16.11', 'mtu': 1350, 'vni': 23}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'vxlan1', 'value': {'addresses': ['192.168.128.15/20'], 'dests': ['192.168.16.10', '192.168.16.11', '192.168.16.12', '192.168.16.13', '192.168.16.14', '192.168.16.5'], 'local_ip': '192.168.16.15', 'mtu': 1350, 'vni': 23}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'vxlan1', 'value': {'addresses': ['192.168.128.13/20'], 'dests': ['192.168.16.10', '192.168.16.11', '192.168.16.12', '192.168.16.14', '192.168.16.15', '192.168.16.5'], 'local_ip': '192.168.16.13', 'mtu': 1350, 'vni': 23}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'vxlan1', 'value': {'addresses': ['192.168.128.14/20'], 'dests': ['192.168.16.10', '192.168.16.11', '192.168.16.12', '192.168.16.13', '192.168.16.15', '192.168.16.5'], 'local_ip': '192.168.16.14', 'mtu': 1350, 'vni': 23}})\u001b[0m\n\nTASK [osism.commons.network : Include networkd cleanup tasks] ******************\nTuesday 24 February 2026  00:37:18 +0000 (0:00:05.271)       0:00:41.142 ****** \n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/network/tasks/cleanup-networkd.yml for testbed-manager, testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m\n\nTASK [osism.commons.network : List existing configuration files] ***************\nTuesday 24 February 2026  00:37:20 +0000 (0:00:01.223)       0:00:42.366 ****** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [osism.commons.network : Remove unused configuration files] ***************\nTuesday 24 February 2026  00:37:20 +0000 (0:00:00.877)       0:00:43.243 ****** \n\u001b[0;36mskipping: [testbed-manager] => (item=/etc/systemd/network/30-vxlan1.network) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager] => (item=/etc/systemd/network/30-vxlan0.network) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager] => (item=/etc/systemd/network/30-vxlan1.netdev) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager] => (item=/etc/systemd/network/30-vxlan0.netdev) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=/etc/systemd/network/30-vxlan1.network) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=/etc/systemd/network/30-vxlan0.network) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=/etc/systemd/network/30-vxlan1.netdev) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=/etc/systemd/network/30-vxlan0.netdev) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=/etc/systemd/network/30-vxlan1.network) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=/etc/systemd/network/30-vxlan0.network) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=/etc/systemd/network/30-vxlan1.netdev) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=/etc/systemd/network/30-vxlan0.netdev) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=/etc/systemd/network/30-vxlan1.network) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=/etc/systemd/network/30-vxlan0.network) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=/etc/systemd/network/30-vxlan1.netdev) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=/etc/systemd/network/30-vxlan0.netdev) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=/etc/systemd/network/30-vxlan1.network) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=/etc/systemd/network/30-vxlan0.network) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=/etc/systemd/network/30-vxlan1.netdev) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=/etc/systemd/network/30-vxlan0.netdev) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=/etc/systemd/network/30-vxlan1.network) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=/etc/systemd/network/30-vxlan0.network) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=/etc/systemd/network/30-vxlan1.netdev) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=/etc/systemd/network/30-vxlan0.netdev) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=/etc/systemd/network/30-vxlan1.network) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=/etc/systemd/network/30-vxlan0.network) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=/etc/systemd/network/30-vxlan1.netdev) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=/etc/systemd/network/30-vxlan0.netdev) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [osism.commons.network : Include network extra init] **********************\nTuesday 24 February 2026  00:37:21 +0000 (0:00:00.969)       0:00:44.212 ****** \n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/network/tasks/network-extra-init.yml for testbed-manager, testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m\n\nTASK [osism.commons.network : Deploy network-extra-init script] ****************\nTuesday 24 February 2026  00:37:23 +0000 (0:00:01.312)       0:00:45.525 ****** \n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [osism.commons.network : Deploy network-extra-init systemd service] *******\nTuesday 24 February 2026  00:37:23 +0000 (0:00:00.727)       0:00:46.253 ****** \n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [osism.commons.network : Enable and start network-extra-init service] *****\nTuesday 24 February 2026  00:37:24 +0000 (0:00:00.912)       0:00:47.165 ****** \n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [osism.commons.network : Disable and stop network-extra-init service] *****\nTuesday 24 February 2026  00:37:25 +0000 (0:00:00.665)       0:00:47.831 ****** \n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [osism.commons.network : Remove network-extra-init systemd service] *******\nTuesday 24 February 2026  00:37:27 +0000 (0:00:01.812)       0:00:49.643 ****** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [osism.commons.network : Remove network-extra-init script] ****************\nTuesday 24 February 2026  00:37:28 +0000 (0:00:01.000)       0:00:50.643 ****** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nRUNNING HANDLER [osism.commons.network : Reload systemd-networkd] **************\nTuesday 24 February 2026  00:37:30 +0000 (0:00:02.417)       0:00:53.061 ****** \n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nRUNNING HANDLER [osism.commons.network : Netplan configuration changed] ********\nTuesday 24 February 2026  00:37:31 +0000 (0:00:01.084)       0:00:54.145 ****** \n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;33mtestbed-manager\u001b[0m            : \u001b[0;32mok=25  \u001b[0m \u001b[0;33mchanged=5   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=8   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=24  \u001b[0m \u001b[0;33mchanged=5   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=9   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=24  \u001b[0m \u001b[0;33mchanged=5   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=9   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=24  \u001b[0m \u001b[0;33mchanged=5   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=9   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-3\u001b[0m             : \u001b[0;32mok=24  \u001b[0m \u001b[0;33mchanged=5   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=9   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-4\u001b[0m             : \u001b[0;32mok=24  \u001b[0m \u001b[0;33mchanged=5   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=9   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-5\u001b[0m             : \u001b[0;32mok=24  \u001b[0m \u001b[0;33mchanged=5   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=9   \u001b[0m rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nTuesday 24 February 2026  00:37:32 +0000 (0:00:00.570)       0:00:54.716 ****** \n=============================================================================== \nosism.commons.network : Create systemd networkd network files ----------- 5.27s\nosism.commons.network : Create systemd networkd netdev files ------------ 5.21s\nosism.commons.network : Include vxlan interfaces ------------------------ 4.48s\nosism.commons.network : Prepare netplan configuration template ---------- 3.56s\nosism.commons.network : Install package networkd-dispatcher ------------- 2.43s\nosism.commons.network : Remove network-extra-init script ---------------- 2.42s\nosism.commons.network : Install required packages ----------------------- 2.13s\nosism.commons.network : Remove netplan configuration template ----------- 1.95s\nosism.commons.network : Remove ifupdown package ------------------------- 1.85s\nosism.commons.network : Disable and stop network-extra-init service ----- 1.81s\nosism.commons.network : Manage service networkd-dispatcher -------------- 1.69s\nosism.commons.network : Copy netplan configuration ---------------------- 1.65s\nosism.commons.network : Include network extra init ---------------------- 1.31s\nosism.commons.network : Include cleanup tasks --------------------------- 1.28s\nosism.commons.network : Include networkd cleanup tasks ------------------ 1.22s\nosism.commons.network : Remove unused configuration files --------------- 1.22s\nosism.commons.network : Include type specific tasks --------------------- 1.21s\nosism.commons.network : List existing configuration files --------------- 1.17s\nosism.commons.network : Check if path for interface file exists --------- 1.14s\nosism.commons.network : Reload systemd-networkd ------------------------- 1.08s\n+ osism apply wireguard\n\u001b[32m2026-02-24 00:37:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mPrepare task for execution of wireguard.\u001b[0m\n\u001b[32m2026-02-24 00:37:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1a3ae285-433a-47d0-8887-3b85c5dc6177 (wireguard) was prepared for execution.\u001b[0m\n\u001b[32m2026-02-24 00:37:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task 1a3ae285-433a-47d0-8887-3b85c5dc6177 (wireguard) has been started and output is visible here.\u001b[0m\n\nPLAY [Apply role wireguard] ****************************************************\n\nTASK [osism.services.wireguard : Install iptables package] *********************\nTuesday 24 February 2026  00:37:49 +0000 (0:00:00.243)       0:00:00.243 ****** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\nTASK [osism.services.wireguard : Install wireguard package] ********************\nTuesday 24 February 2026  00:37:51 +0000 (0:00:01.628)       0:00:01.872 ****** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nTASK [osism.services.wireguard : Create public and private key - server] *******\nTuesday 24 February 2026  00:37:58 +0000 (0:00:07.002)       0:00:08.875 ****** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nTASK [osism.services.wireguard : Create preshared key] *************************\nTuesday 24 February 2026  00:37:58 +0000 (0:00:00.549)       0:00:09.424 ****** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nTASK [osism.services.wireguard : Get preshared key] ****************************\nTuesday 24 February 2026  00:37:59 +0000 (0:00:00.461)       0:00:09.885 ****** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\nTASK [osism.services.wireguard : Get public key - server] **********************\nTuesday 24 February 2026  00:38:00 +0000 (0:00:00.720)       0:00:10.606 ****** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\nTASK [osism.services.wireguard : Get private key - server] *********************\nTuesday 24 February 2026  00:38:00 +0000 (0:00:00.424)       0:00:11.031 ****** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\nTASK [osism.services.wireguard : Copy wg0.conf configuration file] *************\nTuesday 24 February 2026  00:38:01 +0000 (0:00:00.508)       0:00:11.539 ****** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nTASK [osism.services.wireguard : Copy client configuration files] **************\nTuesday 24 February 2026  00:38:02 +0000 (0:00:01.341)       0:00:12.881 ****** \n\u001b[0;33mchanged: [testbed-manager] => (item=None)\u001b[0m\n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nTASK [osism.services.wireguard : Manage wg-quick@wg0.service service] **********\nTuesday 24 February 2026  00:38:03 +0000 (0:00:01.027)       0:00:13.908 ****** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nRUNNING HANDLER [osism.services.wireguard : Restart wg0 service] ***************\nTuesday 24 February 2026  00:38:05 +0000 (0:00:01.749)       0:00:15.658 ****** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;33mtestbed-manager\u001b[0m            : \u001b[0;32mok=11  \u001b[0m \u001b[0;33mchanged=7   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nTuesday 24 February 2026  00:38:06 +0000 (0:00:01.055)       0:00:16.714 ****** \n=============================================================================== \nosism.services.wireguard : Install wireguard package -------------------- 7.00s\nosism.services.wireguard : Manage wg-quick@wg0.service service ---------- 1.75s\nosism.services.wireguard : Install iptables package --------------------- 1.63s\nosism.services.wireguard : Copy wg0.conf configuration file ------------- 1.34s\nosism.services.wireguard : Restart wg0 service -------------------------- 1.06s\nosism.services.wireguard : Copy client configuration files -------------- 1.03s\nosism.services.wireguard : Get preshared key ---------------------------- 0.72s\nosism.services.wireguard : Create public and private key - server ------- 0.55s\nosism.services.wireguard : Get private key - server --------------------- 0.51s\nosism.services.wireguard : Create preshared key ------------------------- 0.46s\nosism.services.wireguard : Get public key - server ---------------------- 0.42s\n+ sh -c /opt/configuration/scripts/prepare-wireguard-configuration.sh\n  % Total    % Received % Xferd  Average Speed   Time    Time     Time  Current\n                                 Dload  Upload   Total   Spent    Left  Speed\n\r  0     0    0     0    0     0      0      0 --:--:-- --:--:-- --:--:--     0\r100    15  100    15    0     0    195      0 --:--:-- --:--:-- --:--:--   197\n+ osism apply --environment custom workarounds\n\u001b[32m2026-02-24 00:38:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTrying to run play workarounds in environment custom\u001b[0m\n\u001b[32m2026-02-24 00:38:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mPrepare task for execution of workarounds.\u001b[0m\n\u001b[32m2026-02-24 00:38:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8e00bff8-7c24-4924-8aee-b1d560d35af1 (workarounds) was prepared for execution.\u001b[0m\n\u001b[32m2026-02-24 00:38:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task 8e00bff8-7c24-4924-8aee-b1d560d35af1 (workarounds) has been started and output is visible here.\u001b[0m\n\nPLAY [Group hosts based on configuration] **************************************\n\nTASK [Group hosts based on virtualization_role] ********************************\nTuesday 24 February 2026  00:38:23 +0000 (0:00:00.132)       0:00:00.132 ****** \n\u001b[0;33mchanged: [testbed-node-0] => (item=virtualization_role_guest)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=virtualization_role_guest)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=virtualization_role_guest)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item=virtualization_role_guest)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=virtualization_role_guest)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=virtualization_role_guest)\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item=virtualization_role_guest)\u001b[0m\n\nPLAY [Apply netplan configuration on the manager node] *************************\n\nTASK [Apply netplan configuration] *********************************************\nTuesday 24 February 2026  00:38:24 +0000 (0:00:00.851)       0:00:00.984 ****** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\nPLAY [Apply netplan configuration on all other nodes] **************************\n\nTASK [Apply netplan configuration] *********************************************\nTuesday 24 February 2026  00:38:27 +0000 (0:00:02.538)       0:00:03.522 ****** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nPLAY [Add custom CA certificates to non-manager nodes] *************************\n\nTASK [Copy custom CA certificates] *********************************************\nTuesday 24 February 2026  00:38:29 +0000 (0:00:01.863)       0:00:05.386 ****** \n\u001b[0;33mchanged: [testbed-node-2] => (item=/opt/configuration/environments/kolla/certificates/ca/testbed.crt)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item=/opt/configuration/environments/kolla/certificates/ca/testbed.crt)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=/opt/configuration/environments/kolla/certificates/ca/testbed.crt)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=/opt/configuration/environments/kolla/certificates/ca/testbed.crt)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=/opt/configuration/environments/kolla/certificates/ca/testbed.crt)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=/opt/configuration/environments/kolla/certificates/ca/testbed.crt)\u001b[0m\n\nTASK [Run update-ca-certificates] **********************************************\nTuesday 24 February 2026  00:38:30 +0000 (0:00:01.581)       0:00:06.967 ****** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\nTASK [Run update-ca-trust] *****************************************************\nTuesday 24 February 2026  00:38:34 +0000 (0:00:03.867)       0:00:10.834 ****** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nPLAY [Add a workaround service] ************************************************\n\nTASK [Copy workarounds.sh scripts] *********************************************\nTuesday 24 February 2026  00:38:35 +0000 (0:00:00.711)       0:00:11.546 ****** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nTASK [Copy workarounds systemd unit file] **************************************\nTuesday 24 February 2026  00:38:36 +0000 (0:00:01.525)       0:00:13.071 ****** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nTASK [Reload systemd daemon] ***************************************************\nTuesday 24 February 2026  00:38:38 +0000 (0:00:01.604)       0:00:14.676 ****** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\nTASK [Enable workarounds.service (Debian)] *************************************\nTuesday 24 February 2026  00:38:39 +0000 (0:00:01.550)       0:00:16.226 ****** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nTASK [Enable and start workarounds.service (RedHat)] ***************************\nTuesday 24 February 2026  00:38:41 +0000 (0:00:01.706)       0:00:17.933 ****** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\nPLAY [On Ubuntu 24.04 install python3-docker from Debian Sid] ******************\n\nTASK [Install python3-docker] **************************************************\nTuesday 24 February 2026  00:38:42 +0000 (0:00:00.608)       0:00:18.541 ****** \n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;33mtestbed-manager\u001b[0m            : \u001b[0;32mok=7   \u001b[0m \u001b[0;33mchanged=4   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=1   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=9   \u001b[0m \u001b[0;33mchanged=6   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=9   \u001b[0m \u001b[0;33mchanged=6   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=9   \u001b[0m \u001b[0;33mchanged=6   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-3\u001b[0m             : \u001b[0;32mok=9   \u001b[0m \u001b[0;33mchanged=6   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-4\u001b[0m             : \u001b[0;32mok=9   \u001b[0m \u001b[0;33mchanged=6   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-5\u001b[0m             : \u001b[0;32mok=9   \u001b[0m \u001b[0;33mchanged=6   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nTuesday 24 February 2026  00:38:45 +0000 (0:00:02.839)       0:00:21.381 ****** \n=============================================================================== \nRun update-ca-certificates ---------------------------------------------- 3.87s\nInstall python3-docker -------------------------------------------------- 2.84s\nApply netplan configuration --------------------------------------------- 2.54s\nApply netplan configuration --------------------------------------------- 1.86s\nEnable workarounds.service (Debian) ------------------------------------- 1.71s\nCopy workarounds systemd unit file -------------------------------------- 1.60s\nCopy custom CA certificates --------------------------------------------- 1.58s\nReload systemd daemon --------------------------------------------------- 1.55s\nCopy workarounds.sh scripts --------------------------------------------- 1.53s\nGroup hosts based on virtualization_role -------------------------------- 0.85s\nRun update-ca-trust ----------------------------------------------------- 0.71s\nEnable and start workarounds.service (RedHat) --------------------------- 0.61s\n+ osism apply reboot -l testbed-nodes -e ireallymeanit=yes\n\u001b[32m2026-02-24 00:38:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mPrepare task for execution of reboot.\u001b[0m\n\u001b[32m2026-02-24 00:38:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 11252132-c821-496b-81ce-f5f47a776033 (reboot) was prepared for execution.\u001b[0m\n\u001b[32m2026-02-24 00:38:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task 11252132-c821-496b-81ce-f5f47a776033 (reboot) has been started and output is visible here.\u001b[0m\n\nPLAY [Reboot systems] **********************************************************\n\nTASK [Exit playbook, if user did not mean to reboot systems] *******************\nTuesday 24 February 2026  00:39:01 +0000 (0:00:00.149)       0:00:00.149 ****** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [Reboot system - do not wait for the reboot to complete] ******************\nTuesday 24 February 2026  00:39:01 +0000 (0:00:00.086)       0:00:00.235 ****** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [Reboot system - wait for the reboot to complete] *************************\nTuesday 24 February 2026  00:39:02 +0000 (0:00:00.923)       0:00:01.159 ****** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nPLAY [Reboot systems] **********************************************************\n\nTASK [Exit playbook, if user did not mean to reboot systems] *******************\nTuesday 24 February 2026  00:39:02 +0000 (0:00:00.090)       0:00:01.249 ****** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [Reboot system - do not wait for the reboot to complete] ******************\nTuesday 24 February 2026  00:39:03 +0000 (0:00:00.081)       0:00:01.331 ****** \n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\nTASK [Reboot system - wait for the reboot to complete] *************************\nTuesday 24 February 2026  00:39:03 +0000 (0:00:00.656)       0:00:01.987 ****** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nPLAY [Reboot systems] **********************************************************\n\nTASK [Exit playbook, if user did not mean to reboot systems] *******************\nTuesday 24 February 2026  00:39:03 +0000 (0:00:00.087)       0:00:02.075 ****** \n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [Reboot system - do not wait for the reboot to complete] ******************\nTuesday 24 February 2026  00:39:03 +0000 (0:00:00.125)       0:00:02.201 ****** \n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [Reboot system - wait for the reboot to complete] *************************\nTuesday 24 February 2026  00:39:04 +0000 (0:00:00.652)       0:00:02.853 ****** \n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nPLAY [Reboot systems] **********************************************************\n\nTASK [Exit playbook, if user did not mean to reboot systems] *******************\nTuesday 24 February 2026  00:39:04 +0000 (0:00:00.119)       0:00:02.972 ****** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Reboot system - do not wait for the reboot to complete] ******************\nTuesday 24 February 2026  00:39:04 +0000 (0:00:00.088)       0:00:03.061 ****** \n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\nTASK [Reboot system - wait for the reboot to complete] *************************\nTuesday 24 February 2026  00:39:05 +0000 (0:00:00.694)       0:00:03.755 ****** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nPLAY [Reboot systems] **********************************************************\n\nTASK [Exit playbook, if user did not mean to reboot systems] *******************\nTuesday 24 February 2026  00:39:05 +0000 (0:00:00.101)       0:00:03.857 ****** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Reboot system - do not wait for the reboot to complete] ******************\nTuesday 24 February 2026  00:39:05 +0000 (0:00:00.096)       0:00:03.954 ****** \n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\nTASK [Reboot system - wait for the reboot to complete] *************************\nTuesday 24 February 2026  00:39:06 +0000 (0:00:00.686)       0:00:04.640 ****** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nPLAY [Reboot systems] **********************************************************\n\nTASK [Exit playbook, if user did not mean to reboot systems] *******************\nTuesday 24 February 2026  00:39:06 +0000 (0:00:00.110)       0:00:04.751 ****** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Reboot system - do not wait for the reboot to complete] ******************\nTuesday 24 February 2026  00:39:06 +0000 (0:00:00.086)       0:00:04.838 ****** \n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nTASK [Reboot system - wait for the reboot to complete] *************************\nTuesday 24 February 2026  00:39:07 +0000 (0:00:00.720)       0:00:05.559 ****** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=1   \u001b[0m \u001b[0;33mchanged=1   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=1   \u001b[0m \u001b[0;33mchanged=1   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=1   \u001b[0m \u001b[0;33mchanged=1   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-3\u001b[0m             : \u001b[0;32mok=1   \u001b[0m \u001b[0;33mchanged=1   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-4\u001b[0m             : \u001b[0;32mok=1   \u001b[0m \u001b[0;33mchanged=1   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-5\u001b[0m             : \u001b[0;32mok=1   \u001b[0m \u001b[0;33mchanged=1   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nTuesday 24 February 2026  00:39:07 +0000 (0:00:00.047)       0:00:05.606 ****** \n=============================================================================== \nReboot system - do not wait for the reboot to complete ------------------ 4.33s\nExit playbook, if user did not mean to reboot systems ------------------- 0.57s\nReboot system - wait for the reboot to complete ------------------------- 0.56s\n+ osism apply wait-for-connection -l testbed-nodes -e ireallymeanit=yes\n\u001b[32m2026-02-24 00:39:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mPrepare task for execution of wait-for-connection.\u001b[0m\n\u001b[32m2026-02-24 00:39:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4909c938-92d1-403b-8766-57378ac3da4b (wait-for-connection) was prepared for execution.\u001b[0m\n\u001b[32m2026-02-24 00:39:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task 4909c938-92d1-403b-8766-57378ac3da4b (wait-for-connection) has been started and output is visible here.\u001b[0m\n\nPLAY [Wait until remote systems are reachable] *********************************\n\nTASK [Wait until remote system is reachable] ***********************************\nTuesday 24 February 2026  00:39:25 +0000 (0:00:00.215)       0:00:00.215 ****** \n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;32mtestbed-node-0\u001b[0m             : \u001b[0;32mok=1   \u001b[0m changed=0    unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\u001b[0;32mtestbed-node-1\u001b[0m             : \u001b[0;32mok=1   \u001b[0m changed=0    unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\u001b[0;32mtestbed-node-2\u001b[0m             : \u001b[0;32mok=1   \u001b[0m changed=0    unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\u001b[0;32mtestbed-node-3\u001b[0m             : \u001b[0;32mok=1   \u001b[0m changed=0    unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\u001b[0;32mtestbed-node-4\u001b[0m             : \u001b[0;32mok=1   \u001b[0m changed=0    unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\u001b[0;32mtestbed-node-5\u001b[0m             : \u001b[0;32mok=1   \u001b[0m changed=0    unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nTuesday 24 February 2026  00:39:36 +0000 (0:00:11.472)       0:00:11.688 ****** \n=============================================================================== \nWait until remote system is reachable ---------------------------------- 11.47s\n+ osism apply hddtemp\n\u001b[32m2026-02-24 00:39:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mPrepare task for execution of hddtemp.\u001b[0m\n\u001b[32m2026-02-24 00:39:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3aa2872d-3529-407f-83f1-db28cd516501 (hddtemp) was prepared for execution.\u001b[0m\n\u001b[32m2026-02-24 00:39:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task 3aa2872d-3529-407f-83f1-db28cd516501 (hddtemp) has been started and output is visible here.\u001b[0m\n\nPLAY [Apply role hddtemp] ******************************************************\n\nTASK [osism.services.hddtemp : Gather variables for each operating system] *****\nTuesday 24 February 2026  00:39:53 +0000 (0:00:00.273)       0:00:00.273 ****** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [osism.services.hddtemp : Include distribution specific install tasks] ****\nTuesday 24 February 2026  00:39:54 +0000 (0:00:00.795)       0:00:01.069 ****** \n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/services/roles/hddtemp/tasks/install-Debian-family.yml for testbed-manager, testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m\n\nTASK [osism.services.hddtemp : Remove hddtemp package] *************************\nTuesday 24 February 2026  00:39:55 +0000 (0:00:01.217)       0:00:02.286 ****** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\nTASK [osism.services.hddtemp : Enable Kernel Module drivetemp] *****************\nTuesday 24 February 2026  00:39:57 +0000 (0:00:02.025)       0:00:04.311 ****** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nTASK [osism.services.hddtemp : Check if drivetemp module is available] *********\nTuesday 24 February 2026  00:39:59 +0000 (0:00:01.285)       0:00:05.597 ****** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\nTASK [osism.services.hddtemp : Load Kernel Module drivetemp] *******************\nTuesday 24 February 2026  00:40:00 +0000 (0:00:01.214)       0:00:06.812 ****** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [osism.services.hddtemp : Install lm-sensors] *****************************\nTuesday 24 February 2026  00:40:01 +0000 (0:00:00.869)       0:00:07.681 ****** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\nTASK [osism.services.hddtemp : Include distribution specific service tasks] ****\nTuesday 24 February 2026  00:40:16 +0000 (0:00:14.768)       0:00:22.449 ****** \n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/services/roles/hddtemp/tasks/service-Debian-family.yml for testbed-manager, testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m\n\nTASK [osism.services.hddtemp : Manage lm-sensors service] **********************\nTuesday 24 February 2026  00:40:17 +0000 (0:00:01.273)       0:00:23.723 ****** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;33mtestbed-manager\u001b[0m            : \u001b[0;32mok=9   \u001b[0m \u001b[0;33mchanged=4   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=8   \u001b[0m \u001b[0;33mchanged=3   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=1   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=8   \u001b[0m \u001b[0;33mchanged=3   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=1   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=8   \u001b[0m \u001b[0;33mchanged=3   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=1   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-3\u001b[0m             : \u001b[0;32mok=8   \u001b[0m \u001b[0;33mchanged=3   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=1   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-4\u001b[0m             : \u001b[0;32mok=8   \u001b[0m \u001b[0;33mchanged=3   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=1   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-5\u001b[0m             : \u001b[0;32mok=8   \u001b[0m \u001b[0;33mchanged=3   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=1   \u001b[0m rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nTuesday 24 February 2026  00:40:19 +0000 (0:00:01.942)       0:00:25.666 ****** \n=============================================================================== \nosism.services.hddtemp : Install lm-sensors ---------------------------- 14.77s\nosism.services.hddtemp : Remove hddtemp package ------------------------- 2.03s\nosism.services.hddtemp : Manage lm-sensors service ---------------------- 1.94s\nosism.services.hddtemp : Enable Kernel Module drivetemp ----------------- 1.29s\nosism.services.hddtemp : Include distribution specific service tasks ---- 1.27s\nosism.services.hddtemp : Include distribution specific install tasks ---- 1.22s\nosism.services.hddtemp : Check if drivetemp module is available --------- 1.21s\nosism.services.hddtemp : Load Kernel Module drivetemp ------------------- 0.87s\nosism.services.hddtemp : Gather variables for each operating system ----- 0.80s\n++ semver latest 7.1.1\n+ [[ -1 -ge 0 ]]\n+ [[ latest == \\l\\a\\t\\e\\s\\t ]]\n+ sudo systemctl restart manager.service\n+ [[ ceph-ansible == \\c\\e\\p\\h\\-\\a\\n\\s\\i\\b\\l\\e ]]\n+ wait_for_container_healthy 60 ceph-ansible\n+ local max_attempts=60\n+ local name=ceph-ansible\n+ local attempt_num=1\n++ /usr/bin/docker inspect -f '{{.State.Health.Status}}' ceph-ansible\n+ [[ unhealthy == \\h\\e\\a\\l\\t\\h\\y ]]\n+ ((  attempt_num++ == max_attempts  ))\n+ sleep 5\n++ /usr/bin/docker inspect -f '{{.State.Health.Status}}' ceph-ansible\n+ [[ unhealthy == \\h\\e\\a\\l\\t\\h\\y ]]\n+ ((  attempt_num++ == max_attempts  ))\n+ sleep 5\n++ /usr/bin/docker inspect -f '{{.State.Health.Status}}' ceph-ansible\n+ [[ unhealthy == \\h\\e\\a\\l\\t\\h\\y ]]\n+ ((  attempt_num++ == max_attempts  ))\n+ sleep 5\n++ /usr/bin/docker inspect -f '{{.State.Health.Status}}' ceph-ansible\n+ [[ unhealthy == \\h\\e\\a\\l\\t\\h\\y ]]\n+ ((  attempt_num++ == max_attempts  ))\n+ sleep 5\n++ /usr/bin/docker inspect -f '{{.State.Health.Status}}' ceph-ansible\n+ [[ unhealthy == \\h\\e\\a\\l\\t\\h\\y ]]\n+ ((  attempt_num++ == max_attempts  ))\n+ sleep 5\n++ /usr/bin/docker inspect -f '{{.State.Health.Status}}' ceph-ansible\n+ [[ unhealthy == \\h\\e\\a\\l\\t\\h\\y ]]\n+ ((  attempt_num++ == max_attempts  ))\n+ sleep 5\n++ /usr/bin/docker inspect -f '{{.State.Health.Status}}' ceph-ansible\n+ [[ unhealthy == \\h\\e\\a\\l\\t\\h\\y ]]\n+ ((  attempt_num++ == max_attempts  ))\n+ sleep 5\n++ /usr/bin/docker inspect -f '{{.State.Health.Status}}' ceph-ansible\n+ [[ starting == \\h\\e\\a\\l\\t\\h\\y ]]\n+ ((  attempt_num++ == max_attempts  ))\n+ sleep 5\n++ /usr/bin/docker inspect -f '{{.State.Health.Status}}' ceph-ansible\n+ [[ starting == \\h\\e\\a\\l\\t\\h\\y ]]\n+ ((  attempt_num++ == max_attempts  ))\n+ sleep 5\n++ /usr/bin/docker inspect -f '{{.State.Health.Status}}' ceph-ansible\n+ [[ starting == \\h\\e\\a\\l\\t\\h\\y ]]\n+ ((  attempt_num++ == max_attempts  ))\n+ sleep 5\n++ /usr/bin/docker inspect -f '{{.State.Health.Status}}' ceph-ansible\n+ [[ starting == \\h\\e\\a\\l\\t\\h\\y ]]\n+ ((  attempt_num++ == max_attempts  ))\n+ sleep 5\n++ /usr/bin/docker inspect -f '{{.State.Health.Status}}' ceph-ansible\n+ [[ starting == \\h\\e\\a\\l\\t\\h\\y ]]\n+ ((  attempt_num++ == max_attempts  ))\n+ sleep 5\n++ /usr/bin/docker inspect -f '{{.State.Health.Status}}' ceph-ansible\n+ [[ starting == \\h\\e\\a\\l\\t\\h\\y ]]\n+ ((  attempt_num++ == max_attempts  ))\n+ sleep 5\n++ /usr/bin/docker inspect -f '{{.State.Health.Status}}' ceph-ansible\n+ [[ healthy == \\h\\e\\a\\l\\t\\h\\y ]]\n+ wait_for_container_healthy 60 kolla-ansible\n+ local max_attempts=60\n+ local name=kolla-ansible\n+ local attempt_num=1\n++ /usr/bin/docker inspect -f '{{.State.Health.Status}}' kolla-ansible\n+ [[ healthy == \\h\\e\\a\\l\\t\\h\\y ]]\n+ wait_for_container_healthy 60 osism-ansible\n+ local max_attempts=60\n+ local name=osism-ansible\n+ local attempt_num=1\n++ /usr/bin/docker inspect -f '{{.State.Health.Status}}' osism-ansible\n+ [[ healthy == \\h\\e\\a\\l\\t\\h\\y ]]\n+ [[ true == \\t\\r\\u\\e ]]\n+ sh -c /opt/configuration/scripts/disable-ara.sh\nARA in ceph-ansible already disabled.\nARA in kolla-ansible already disabled.\nARA in osism-ansible already disabled.\nARA in osism-kubernetes already disabled.\n+ osism apply gather-facts\n\u001b[32m2026-02-24 00:41:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mPrepare task for execution of gather-facts.\u001b[0m\n\u001b[32m2026-02-24 00:41:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 09f896ba-61bc-4bff-955b-0f3f7aeaadc4 (gather-facts) was prepared for execution.\u001b[0m\n\u001b[32m2026-02-24 00:41:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task 09f896ba-61bc-4bff-955b-0f3f7aeaadc4 (gather-facts) has been started and output is visible here.\u001b[0m\n\nPLAY [Gather facts for all hosts] **********************************************\n\nTASK [Gathers facts about hosts] ***********************************************\nTuesday 24 February 2026  00:41:56 +0000 (0:00:00.197)       0:00:00.197 ****** \n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nPLAY [Gather facts for all hosts if using --limit] *****************************\n\nTASK [Gather facts for all hosts] **********************************************\nTuesday 24 February 2026  00:42:04 +0000 (0:00:08.546)       0:00:08.743 ****** \n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;32mtestbed-manager\u001b[0m            : \u001b[0;32mok=1   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=1   \u001b[0m rescued=0    ignored=0   \n\u001b[0;32mtestbed-node-0\u001b[0m             : \u001b[0;32mok=1   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=1   \u001b[0m rescued=0    ignored=0   \n\u001b[0;32mtestbed-node-1\u001b[0m             : \u001b[0;32mok=1   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=1   \u001b[0m rescued=0    ignored=0   \n\u001b[0;32mtestbed-node-2\u001b[0m             : \u001b[0;32mok=1   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=1   \u001b[0m rescued=0    ignored=0   \n\u001b[0;32mtestbed-node-3\u001b[0m             : \u001b[0;32mok=1   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=1   \u001b[0m rescued=0    ignored=0   \n\u001b[0;32mtestbed-node-4\u001b[0m             : \u001b[0;32mok=1   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=1   \u001b[0m rescued=0    ignored=0   \n\u001b[0;32mtestbed-node-5\u001b[0m             : \u001b[0;32mok=1   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=1   \u001b[0m rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nTuesday 24 February 2026  00:42:05 +0000 (0:00:00.444)       0:00:09.187 ****** \n=============================================================================== \nGathers facts about hosts ----------------------------------------------- 8.55s\nGather facts for all hosts ---------------------------------------------- 0.44s\n+ sudo ln -sf /opt/configuration/scripts/deploy/001-helpers.sh /usr/local/bin/deploy-helper\n+ sudo ln -sf /opt/configuration/scripts/deploy/100-ceph-with-ansible.sh /usr/local/bin/deploy-ceph-with-ansible\n+ sudo ln -sf /opt/configuration/scripts/deploy/100-ceph-with-rook.sh /usr/local/bin/deploy-ceph-with-rook\n+ sudo ln -sf /opt/configuration/scripts/deploy/200-infrastructure.sh /usr/local/bin/deploy-infrastructure\n+ sudo ln -sf /opt/configuration/scripts/deploy/300-openstack.sh /usr/local/bin/deploy-openstack\n+ sudo ln -sf /opt/configuration/scripts/deploy/320-openstack-minimal.sh /usr/local/bin/deploy-openstack-minimal\n+ sudo ln -sf /opt/configuration/scripts/deploy/400-monitoring.sh /usr/local/bin/deploy-monitoring\n+ sudo ln -sf /opt/configuration/scripts/deploy/500-kubernetes.sh /usr/local/bin/deploy-kubernetes\n+ sudo ln -sf /opt/configuration/scripts/deploy/510-clusterapi.sh /usr/local/bin/deploy-kubernetes-clusterapi\n+ sudo ln -sf /opt/configuration/scripts/upgrade-manager.sh /usr/local/bin/upgrade-manager\n+ sudo ln -sf /opt/configuration/scripts/upgrade/100-ceph-with-ansible.sh /usr/local/bin/upgrade-ceph-with-ansible\n+ sudo ln -sf /opt/configuration/scripts/upgrade/100-ceph-with-rook.sh /usr/local/bin/upgrade-ceph-with-rook\n+ sudo ln -sf /opt/configuration/scripts/upgrade/200-infrastructure.sh /usr/local/bin/upgrade-infrastructure\n+ sudo ln -sf /opt/configuration/scripts/upgrade/300-openstack.sh /usr/local/bin/upgrade-openstack\n+ sudo ln -sf /opt/configuration/scripts/upgrade/320-openstack-minimal.sh /usr/local/bin/upgrade-openstack-minimal\n+ sudo ln -sf /opt/configuration/scripts/upgrade/400-monitoring.sh /usr/local/bin/upgrade-monitoring\n+ sudo ln -sf /opt/configuration/scripts/upgrade/500-kubernetes.sh /usr/local/bin/upgrade-kubernetes\n+ sudo ln -sf /opt/configuration/scripts/upgrade/510-clusterapi.sh /usr/local/bin/upgrade-kubernetes-clusterapi\n+ sudo ln -sf /opt/configuration/scripts/bootstrap/300-openstack.sh /usr/local/bin/bootstrap-openstack\n+ sudo ln -sf /opt/configuration/scripts/bootstrap/301-openstack-octavia-amhpora-image.sh /usr/local/bin/bootstrap-octavia\n+ sudo ln -sf /opt/configuration/scripts/bootstrap/302-openstack-k8s-clusterapi-images.sh /usr/local/bin/bootstrap-clusterapi\n+ sudo ln -sf /opt/configuration/scripts/disable-local-registry.sh /usr/local/bin/disable-local-registry\n+ sudo ln -sf /opt/configuration/scripts/pull-images.sh /usr/local/bin/pull-images\n+ [[ false == \\t\\r\\u\\e ]]",
                            "stdout_lines": [
                                "",
                                "# DEPLOY MANAGER",
                                "",
                                "+ set -e",
                                "+ echo",
                                "+ echo '# DEPLOY MANAGER'",
                                "+ echo",
                                "+ cat /opt/manager-vars.sh",
                                "export NUMBER_OF_NODES=6",
                                "",
                                "export CEPH_VERSION=reef",
                                "export CONFIGURATION_VERSION=main",
                                "export MANAGER_VERSION=latest",
                                "export OPENSTACK_VERSION=2024.2",
                                "",
                                "export ARA=false",
                                "export DEPLOY_MODE=manager",
                                "export TEMPEST=true",
                                "export IS_ZUUL=true",
                                "",
                                "export MANAGER_PUBLIC_IP_ADDRESS=81.163.192.195",
                                "export EXTERNAL_API=false",
                                "",
                                "export IMAGE_USER=ubuntu",
                                "export IMAGE_NODE_USER=ubuntu",
                                "",
                                "export CEPH_STACK=ceph-ansible",
                                "",
                                "+ echo",
                                "+ source /opt/configuration/scripts/include.sh",
                                "++ export INTERACTIVE=false",
                                "++ INTERACTIVE=false",
                                "++ export OSISM_APPLY_RETRY=1",
                                "++ OSISM_APPLY_RETRY=1",
                                "+ source /opt/manager-vars.sh",
                                "++ export NUMBER_OF_NODES=6",
                                "++ NUMBER_OF_NODES=6",
                                "++ export CEPH_VERSION=reef",
                                "++ CEPH_VERSION=reef",
                                "++ export CONFIGURATION_VERSION=main",
                                "++ CONFIGURATION_VERSION=main",
                                "++ export MANAGER_VERSION=latest",
                                "++ MANAGER_VERSION=latest",
                                "++ export OPENSTACK_VERSION=2024.2",
                                "++ OPENSTACK_VERSION=2024.2",
                                "++ export ARA=false",
                                "++ ARA=false",
                                "++ export DEPLOY_MODE=manager",
                                "++ DEPLOY_MODE=manager",
                                "++ export TEMPEST=true",
                                "++ TEMPEST=true",
                                "++ export IS_ZUUL=true",
                                "++ IS_ZUUL=true",
                                "++ export MANAGER_PUBLIC_IP_ADDRESS=81.163.192.195",
                                "++ MANAGER_PUBLIC_IP_ADDRESS=81.163.192.195",
                                "++ export EXTERNAL_API=false",
                                "++ EXTERNAL_API=false",
                                "++ export IMAGE_USER=ubuntu",
                                "++ IMAGE_USER=ubuntu",
                                "++ export IMAGE_NODE_USER=ubuntu",
                                "++ IMAGE_NODE_USER=ubuntu",
                                "++ export CEPH_STACK=ceph-ansible",
                                "++ CEPH_STACK=ceph-ansible",
                                "+ sudo ln -sf /opt/configuration/contrib/semver2.sh /usr/local/bin/semver",
                                "+ docker version",
                                "Client: Docker Engine - Community",
                                " Version:           27.5.1",
                                " API version:       1.47",
                                " Go version:        go1.22.11",
                                " Git commit:        9f9e405",
                                " Built:             Wed Jan 22 13:41:48 2025",
                                " OS/Arch:           linux/amd64",
                                " Context:           default",
                                "",
                                "Server: Docker Engine - Community",
                                " Engine:",
                                "  Version:          27.5.1",
                                "  API version:      1.47 (minimum version 1.24)",
                                "  Go version:       go1.22.11",
                                "  Git commit:       4c9b3b0",
                                "  Built:            Wed Jan 22 13:41:48 2025",
                                "  OS/Arch:          linux/amd64",
                                "  Experimental:     false",
                                " containerd:",
                                "  Version:          v2.2.1",
                                "  GitCommit:        dea7da592f5d1d2b7755e3a161be07f43fad8f75",
                                " runc:",
                                "  Version:          1.3.4",
                                "  GitCommit:        v1.3.4-0-gd6d73eb8",
                                " docker-init:",
                                "  Version:          0.19.0",
                                "  GitCommit:        de40ad0",
                                "+ sh -c /opt/configuration/scripts/deploy/000-manager.sh",
                                "+ set -e",
                                "+ source /opt/manager-vars.sh",
                                "++ export NUMBER_OF_NODES=6",
                                "++ NUMBER_OF_NODES=6",
                                "++ export CEPH_VERSION=reef",
                                "++ CEPH_VERSION=reef",
                                "++ export CONFIGURATION_VERSION=main",
                                "++ CONFIGURATION_VERSION=main",
                                "++ export MANAGER_VERSION=latest",
                                "++ MANAGER_VERSION=latest",
                                "++ export OPENSTACK_VERSION=2024.2",
                                "++ OPENSTACK_VERSION=2024.2",
                                "++ export ARA=false",
                                "++ ARA=false",
                                "++ export DEPLOY_MODE=manager",
                                "++ DEPLOY_MODE=manager",
                                "++ export TEMPEST=true",
                                "++ TEMPEST=true",
                                "++ export IS_ZUUL=true",
                                "++ IS_ZUUL=true",
                                "++ export MANAGER_PUBLIC_IP_ADDRESS=81.163.192.195",
                                "++ MANAGER_PUBLIC_IP_ADDRESS=81.163.192.195",
                                "++ export EXTERNAL_API=false",
                                "++ EXTERNAL_API=false",
                                "++ export IMAGE_USER=ubuntu",
                                "++ IMAGE_USER=ubuntu",
                                "++ export IMAGE_NODE_USER=ubuntu",
                                "++ IMAGE_NODE_USER=ubuntu",
                                "++ export CEPH_STACK=ceph-ansible",
                                "++ CEPH_STACK=ceph-ansible",
                                "+ source /opt/configuration/scripts/include.sh",
                                "++ export INTERACTIVE=false",
                                "++ INTERACTIVE=false",
                                "++ export OSISM_APPLY_RETRY=1",
                                "++ OSISM_APPLY_RETRY=1",
                                "+ [[ latest != \\l\\a\\t\\e\\s\\t ]]",
                                "+ [[ latest == \\l\\a\\t\\e\\s\\t ]]",
                                "+ /opt/configuration/scripts/set-ceph-version.sh reef",
                                "+ set -e",
                                "+ VERSION=reef",
                                "++ grep '^ceph_version:' /opt/configuration/environments/manager/configuration.yml",
                                "+ [[ -n ceph_version: reef ]]",
                                "+ sed -i 's/ceph_version: .*/ceph_version: reef/g' /opt/configuration/environments/manager/configuration.yml",
                                "+ /opt/configuration/scripts/set-openstack-version.sh 2024.2",
                                "+ set -e",
                                "+ VERSION=2024.2",
                                "++ grep '^openstack_version:' /opt/configuration/environments/manager/configuration.yml",
                                "+ [[ -n openstack_version: 2024.2 ]]",
                                "+ sed -i 's/openstack_version: .*/openstack_version: 2024.2/g' /opt/configuration/environments/manager/configuration.yml",
                                "+ [[ ceph-ansible == \\r\\o\\o\\k ]]",
                                "++ semver latest 7.0.0",
                                "+ [[ -1 -ge 0 ]]",
                                "+ [[ latest == \\l\\a\\t\\e\\s\\t ]]",
                                "+ echo 'enable_osism_kubernetes: true'",
                                "++ semver latest 10.0.0-0",
                                "+ [[ -1 -ge 0 ]]",
                                "++ semver 2024.2 2025.1",
                                "+ [[ -1 -ge 0 ]]",
                                "+ /opt/configuration/scripts/enable-resource-nodes.sh",
                                "+ [[ -e /opt/venv/bin/activate ]]",
                                "+ source /opt/venv/bin/activate",
                                "++ deactivate nondestructive",
                                "++ '[' -n '' ']'",
                                "++ '[' -n '' ']'",
                                "++ hash -r",
                                "++ '[' -n '' ']'",
                                "++ unset VIRTUAL_ENV",
                                "++ unset VIRTUAL_ENV_PROMPT",
                                "++ '[' '!' nondestructive = nondestructive ']'",
                                "++ '[' linux-gnu = cygwin ']'",
                                "++ '[' linux-gnu = msys ']'",
                                "++ export VIRTUAL_ENV=/opt/venv",
                                "++ VIRTUAL_ENV=/opt/venv",
                                "++ _OLD_VIRTUAL_PATH=/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/usr/games:/usr/local/games:/snap/bin",
                                "++ PATH=/opt/venv/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/usr/games:/usr/local/games:/snap/bin",
                                "++ export PATH",
                                "++ '[' -n '' ']'",
                                "++ '[' -z '' ']'",
                                "++ _OLD_VIRTUAL_PS1=",
                                "++ PS1='(venv) '",
                                "++ export PS1",
                                "++ VIRTUAL_ENV_PROMPT='(venv) '",
                                "++ export VIRTUAL_ENV_PROMPT",
                                "++ hash -r",
                                "+ ansible-playbook -i testbed-manager, --vault-password-file /opt/configuration/environments/.vault_pass /opt/configuration/ansible/manager-part-3.yml",
                                "",
                                "PLAY [Copy custom facts] *******************************************************",
                                "",
                                "TASK [Create custom facts directory] *******************************************",
                                "ok: [testbed-manager]",
                                "",
                                "TASK [Copy fact files] *********************************************************",
                                "changed: [testbed-manager]",
                                "",
                                "PLAY [Before the deployment of the manager] ************************************",
                                "",
                                "TASK [Gathering Facts] *********************************************************",
                                "ok: [testbed-manager]",
                                "",
                                "TASK [Get /opt/manager-vars.sh] ************************************************",
                                "ok: [testbed-manager]",
                                "",
                                "TASK [Add ara_server_mariadb_volume_type parameter] ****************************",
                                "changed: [testbed-manager]",
                                "",
                                "TASK [Add netbox_enable parameter] *********************************************",
                                "skipping: [testbed-manager]",
                                "",
                                "TASK [Install HWE kernel package on Ubuntu] ************************************",
                                "changed: [testbed-manager]",
                                "",
                                "TASK [Check if /etc/OTC_region exist] ******************************************",
                                "ok: [testbed-manager]",
                                "",
                                "TASK [Add nova_compute_virt_type parameter] ************************************",
                                "skipping: [testbed-manager]",
                                "",
                                "PLAY [Apply role traefik] ******************************************************",
                                "",
                                "TASK [Gathering Facts] *********************************************************",
                                "ok: [testbed-manager]",
                                "",
                                "TASK [Apply traefik role] ******************************************************",
                                "included: osism.services.traefik for testbed-manager",
                                "",
                                "TASK [osism.services.traefik : Include config tasks] ***************************",
                                "included: /usr/share/ansible/collections/ansible_collections/osism/services/roles/traefik/tasks/config.yml for testbed-manager",
                                "",
                                "TASK [osism.services.traefik : Create required directories] ********************",
                                "changed: [testbed-manager] => (item=/opt/traefik)",
                                "changed: [testbed-manager] => (item=/opt/traefik/certificates)",
                                "changed: [testbed-manager] => (item=/opt/traefik/configuration)",
                                "",
                                "TASK [osism.services.traefik : Copy configuration files] ***********************",
                                "changed: [testbed-manager] => (item=traefik.yml)",
                                "changed: [testbed-manager] => (item=traefik.env)",
                                "changed: [testbed-manager] => (item=certificates.yml)",
                                "",
                                "TASK [osism.services.traefik : Copy certificate cert files] ********************",
                                "changed: [testbed-manager] => (item=None)",
                                "changed: [testbed-manager]",
                                "",
                                "TASK [osism.services.traefik : Copy certificate key files] *********************",
                                "changed: [testbed-manager] => (item=None)",
                                "changed: [testbed-manager]",
                                "",
                                "TASK [osism.services.traefik : Copy dynamic configuration] *********************",
                                "skipping: [testbed-manager]",
                                "",
                                "TASK [osism.services.traefik : Remove dynamic configuration] *******************",
                                "ok: [testbed-manager]",
                                "",
                                "TASK [osism.services.traefik : Include service tasks] **************************",
                                "included: /usr/share/ansible/collections/ansible_collections/osism/services/roles/traefik/tasks/service.yml for testbed-manager",
                                "",
                                "TASK [osism.services.traefik : Create traefik external network] ****************",
                                "changed: [testbed-manager]",
                                "",
                                "TASK [osism.services.traefik : Copy docker-compose.yml file] *******************",
                                "changed: [testbed-manager]",
                                "",
                                "TASK [osism.services.traefik : Manage traefik service] *************************",
                                "changed: [testbed-manager]",
                                "",
                                "RUNNING HANDLER [osism.services.traefik : Restart traefik service] *************",
                                "skipping: [testbed-manager]",
                                "",
                                "PLAY [Deploy manager service] **************************************************",
                                "",
                                "TASK [Gathering Facts] *********************************************************",
                                "ok: [testbed-manager]",
                                "",
                                "TASK [Apply manager role] ******************************************************",
                                "included: osism.services.manager for testbed-manager",
                                "",
                                "TASK [osism.services.manager : Include install tasks] **************************",
                                "included: /usr/share/ansible/collections/ansible_collections/osism/services/roles/manager/tasks/install-Debian-family.yml for testbed-manager",
                                "",
                                "TASK [osism.services.manager : Install required packages] **********************",
                                "ok: [testbed-manager]",
                                "",
                                "TASK [osism.services.manager : Gather variables for each operating system] *****",
                                "ok: [testbed-manager]",
                                "",
                                "TASK [osism.services.manager : Include config tasks] ***************************",
                                "included: /usr/share/ansible/collections/ansible_collections/osism/services/roles/manager/tasks/config.yml for testbed-manager",
                                "",
                                "TASK [osism.services.manager : Create required directories] ********************",
                                "changed: [testbed-manager] => (item=/opt/ansible)",
                                "changed: [testbed-manager] => (item=/opt/archive)",
                                "changed: [testbed-manager] => (item=/opt/manager/configuration)",
                                "changed: [testbed-manager] => (item=/opt/manager/data)",
                                "ok: [testbed-manager] => (item=/opt/manager)",
                                "changed: [testbed-manager] => (item=/opt/manager/secrets)",
                                "changed: [testbed-manager] => (item=/opt/ansible/secrets)",
                                "changed: [testbed-manager] => (item=/opt/state)",
                                "",
                                "TASK [osism.services.manager : Copy all environment file] **********************",
                                "changed: [testbed-manager]",
                                "",
                                "TASK [osism.services.manager : Copy client environment file] *******************",
                                "changed: [testbed-manager]",
                                "",
                                "TASK [osism.services.manager : Include ara config tasks] ***********************",
                                "included: /usr/share/ansible/collections/ansible_collections/osism/services/roles/manager/tasks/config-ara.yml for testbed-manager",
                                "",
                                "TASK [osism.services.manager : Copy ARA environment files] *********************",
                                "changed: [testbed-manager] => (item=ara)",
                                "changed: [testbed-manager] => (item=ara-server)",
                                "",
                                "TASK [osism.services.manager : Copy MariaDB environment file] ******************",
                                "changed: [testbed-manager]",
                                "",
                                "TASK [osism.services.manager : Include vault config tasks] *********************",
                                "skipping: [testbed-manager]",
                                "",
                                "TASK [osism.services.manager : Include frontend config tasks] ******************",
                                "included: /usr/share/ansible/collections/ansible_collections/osism/services/roles/manager/tasks/config-frontend.yml for testbed-manager",
                                "",
                                "TASK [osism.services.manager : Copy frontend environment file] *****************",
                                "changed: [testbed-manager]",
                                "",
                                "TASK [osism.services.manager : Include ansible config tasks] *******************",
                                "included: /usr/share/ansible/collections/ansible_collections/osism/services/roles/manager/tasks/config-ansible.yml for testbed-manager",
                                "",
                                "TASK [osism.services.manager : Copy private ssh keys] **************************",
                                "changed: [testbed-manager] => (item=None)",
                                "changed: [testbed-manager] => (item=None)",
                                "changed: [testbed-manager]",
                                "",
                                "TASK [osism.services.manager : Copy ansible environment file] ******************",
                                "changed: [testbed-manager]",
                                "",
                                "TASK [osism.services.manager : Include netbox config tasks] ********************",
                                "skipping: [testbed-manager]",
                                "",
                                "TASK [osism.services.manager : Include celery config tasks] ********************",
                                "included: /usr/share/ansible/collections/ansible_collections/osism/services/roles/manager/tasks/config-celery.yml for testbed-manager",
                                "",
                                "TASK [osism.services.manager : Set fs.inotify.max_user_watches] ****************",
                                "changed: [testbed-manager]",
                                "",
                                "TASK [osism.services.manager : Set fs.inotify.max_user_instances] **************",
                                "changed: [testbed-manager]",
                                "",
                                "TASK [osism.services.manager : Copy celery environment files] ******************",
                                "changed: [testbed-manager] => (item=conductor)",
                                "changed: [testbed-manager] => (item=openstack)",
                                "",
                                "TASK [osism.services.manager : Copy listener environment file] *****************",
                                "changed: [testbed-manager]",
                                "",
                                "TASK [osism.services.manager : Check for conductor.yml] ************************",
                                "ok: [testbed-manager]",
                                "",
                                "TASK [osism.services.manager : Copy conductor configuration file] **************",
                                "changed: [testbed-manager]",
                                "",
                                "TASK [osism.services.manager : Copy empty conductor configuration file] ********",
                                "skipping: [testbed-manager]",
                                "",
                                "TASK [osism.services.manager : Include wrapper config tasks] *******************",
                                "included: /usr/share/ansible/collections/ansible_collections/osism/services/roles/manager/tasks/config-wrapper.yml for testbed-manager",
                                "",
                                "TASK [osism.services.manager : Include wrapper vars file] **********************",
                                "ok: [testbed-manager]",
                                "",
                                "TASK [osism.services.manager : Copy wrapper scripts] ***************************",
                                "changed: [testbed-manager] => (item=osism)",
                                "changed: [testbed-manager] => (item=osism-update-docker)",
                                "changed: [testbed-manager] => (item=osism-update-manager)",
                                "",
                                "TASK [osism.services.manager : Copy cilium wrapper script] *********************",
                                "changed: [testbed-manager]",
                                "",
                                "TASK [osism.services.manager : Copy hubble wrapper script] *********************",
                                "changed: [testbed-manager]",
                                "",
                                "TASK [osism.services.manager : Copy flux wrapper script] ***********************",
                                "changed: [testbed-manager]",
                                "",
                                "TASK [osism.services.manager : Include scripts config tasks] *******************",
                                "included: /usr/share/ansible/collections/ansible_collections/osism/services/roles/manager/tasks/config-scripts.yml for testbed-manager",
                                "",
                                "TASK [osism.services.manager : Include scripts vars file] **********************",
                                "ok: [testbed-manager]",
                                "",
                                "TASK [osism.services.manager : Copy scripts] ***********************************",
                                "changed: [testbed-manager] => (item=osism-include)",
                                "",
                                "TASK [osism.services.manager : Include service tasks] **************************",
                                "included: /usr/share/ansible/collections/ansible_collections/osism/services/roles/manager/tasks/service.yml for testbed-manager",
                                "",
                                "TASK [osism.services.manager : Copy manager systemd unit file] *****************",
                                "changed: [testbed-manager]",
                                "",
                                "TASK [osism.services.manager : Create traefik external network] ****************",
                                "ok: [testbed-manager]",
                                "",
                                "TASK [osism.services.manager : Set mariadb healthcheck for mariadb < 11.0.0] ***",
                                "skipping: [testbed-manager]",
                                "",
                                "TASK [osism.services.manager : Set mariadb healthcheck for mariadb >= 11.0.0] ***",
                                "ok: [testbed-manager]",
                                "",
                                "TASK [osism.services.manager : Copy docker-compose.yml file] *******************",
                                "changed: [testbed-manager]",
                                "",
                                "TASK [osism.services.manager : Pull container images] **************************",
                                "changed: [testbed-manager]",
                                "",
                                "TASK [osism.services.manager : Stop and disable old service docker-compose@manager] ***",
                                "ok: [testbed-manager]",
                                "",
                                "TASK [osism.services.manager : Do a manual start of the manager service] *******",
                                "skipping: [testbed-manager]",
                                "",
                                "TASK [osism.services.manager : Manage manager service] *************************",
                                "changed: [testbed-manager]",
                                "",
                                "TASK [osism.services.manager : Register that manager service was started] ******",
                                "ok: [testbed-manager]",
                                "",
                                "TASK [osism.services.manager : Flush handlers] *********************************",
                                "",
                                "RUNNING HANDLER [osism.services.manager : Restart manager service] *************",
                                "skipping: [testbed-manager]",
                                "",
                                "RUNNING HANDLER [osism.services.manager : Wait for manager service to start] ***",
                                "Pausing for 60 seconds",
                                "changed: [testbed-manager]",
                                "",
                                "RUNNING HANDLER [osism.services.manager : Ensure that all containers are up] ***",
                                "changed: [testbed-manager]",
                                "",
                                "RUNNING HANDLER [osism.services.manager : Wait for an healthy manager service] ***",
                                "FAILED - RETRYING: [testbed-manager]: Wait for an healthy manager service (50 retries left).",
                                "FAILED - RETRYING: [testbed-manager]: Wait for an healthy manager service (49 retries left).",
                                "changed: [testbed-manager]",
                                "",
                                "RUNNING HANDLER [osism.services.manager : Copy osismclient bash completion script] ***",
                                "changed: [testbed-manager]",
                                "",
                                "TASK [osism.services.manager : Include initialize tasks] ***********************",
                                "included: /usr/share/ansible/collections/ansible_collections/osism/services/roles/manager/tasks/initialize.yml for testbed-manager",
                                "",
                                "TASK [osism.services.manager : Flush handlers] *********************************",
                                "",
                                "TASK [osism.services.manager : Include vault initialize tasks] *****************",
                                "skipping: [testbed-manager]",
                                "",
                                "TASK [osism.services.manager : Include version verification tasks] *************",
                                "included: /usr/share/ansible/collections/ansible_collections/osism/services/roles/manager/tasks/verify-versions.yml for testbed-manager",
                                "",
                                "TASK [osism.services.manager : Deploy service manager version check script] ****",
                                "changed: [testbed-manager]",
                                "",
                                "TASK [osism.services.manager : Execute service manager version check] **********",
                                "ok: [testbed-manager]",
                                "",
                                "TASK [osism.services.manager : Display version check results] ******************",
                                "ok: [testbed-manager] => {",
                                "    \"version_check_result.stdout_lines\": [",
                                "        \"=== OSISM Container Version Check ===\",",
                                "        \"Checking running containers against expected versions...\",",
                                "        \"\",",
                                "        \"Checking service: inventory_reconciler (Inventory Reconciler Service)\",",
                                "        \"  Expected: registry.osism.tech/osism/inventory-reconciler:latest\",",
                                "        \"  Enabled:  true\",",
                                "        \"  Running:  registry.osism.tech/osism/inventory-reconciler:latest\",",
                                "        \"  Status:   \u2705 MATCH\",",
                                "        \"\",",
                                "        \"Checking service: osism-ansible (OSISM Ansible Service)\",",
                                "        \"  Expected: registry.osism.tech/osism/osism-ansible:latest\",",
                                "        \"  Enabled:  true\",",
                                "        \"  Running:  registry.osism.tech/osism/osism-ansible:latest\",",
                                "        \"  Status:   \u2705 MATCH\",",
                                "        \"\",",
                                "        \"Checking service: osism-kubernetes (Osism-Kubernetes Service)\",",
                                "        \"  Expected: registry.osism.tech/osism/osism-kubernetes:latest\",",
                                "        \"  Enabled:  true\",",
                                "        \"  Running:  registry.osism.tech/osism/osism-kubernetes:latest\",",
                                "        \"  Status:   \u2705 MATCH\",",
                                "        \"\",",
                                "        \"Checking service: ceph-ansible (Ceph-Ansible Service)\",",
                                "        \"  Expected: registry.osism.tech/osism/ceph-ansible:reef\",",
                                "        \"  Enabled:  true\",",
                                "        \"  Running:  registry.osism.tech/osism/ceph-ansible:reef\",",
                                "        \"  Status:   \u2705 MATCH\",",
                                "        \"\",",
                                "        \"Checking service: kolla-ansible (Kolla-Ansible Service)\",",
                                "        \"  Expected: registry.osism.tech/osism/kolla-ansible:2024.2\",",
                                "        \"  Enabled:  true\",",
                                "        \"  Running:  registry.osism.tech/osism/kolla-ansible:2024.2\",",
                                "        \"  Status:   \u2705 MATCH\",",
                                "        \"\",",
                                "        \"Checking service: osismclient (OSISM Client)\",",
                                "        \"  Expected: registry.osism.tech/osism/osism:latest\",",
                                "        \"  Enabled:  true\",",
                                "        \"  Running:  registry.osism.tech/osism/osism:latest\",",
                                "        \"  Status:   \u2705 MATCH\",",
                                "        \"\",",
                                "        \"Checking service: ara-server (ARA Server)\",",
                                "        \"  Expected: registry.osism.tech/osism/ara-server:1.7.3\",",
                                "        \"  Enabled:  true\",",
                                "        \"  Running:  registry.osism.tech/osism/ara-server:1.7.3\",",
                                "        \"  Status:   \u2705 MATCH\",",
                                "        \"\",",
                                "        \"Checking service: mariadb (MariaDB for ARA)\",",
                                "        \"  Expected: registry.osism.tech/dockerhub/library/mariadb:11.8.4\",",
                                "        \"  Enabled:  true\",",
                                "        \"  Running:  registry.osism.tech/dockerhub/library/mariadb:11.8.4\",",
                                "        \"  Status:   \u2705 MATCH\",",
                                "        \"\",",
                                "        \"Checking service: frontend (OSISM Frontend)\",",
                                "        \"  Expected: registry.osism.tech/osism/osism-frontend:latest\",",
                                "        \"  Enabled:  true\",",
                                "        \"  Running:  registry.osism.tech/osism/osism-frontend:latest\",",
                                "        \"  Status:   \u2705 MATCH\",",
                                "        \"\",",
                                "        \"Checking service: redis (Redis Cache)\",",
                                "        \"  Expected: registry.osism.tech/dockerhub/library/redis:7.4.7-alpine\",",
                                "        \"  Enabled:  true\",",
                                "        \"  Running:  registry.osism.tech/dockerhub/library/redis:7.4.7-alpine\",",
                                "        \"  Status:   \u2705 MATCH\",",
                                "        \"\",",
                                "        \"Checking service: api (OSISM API Service)\",",
                                "        \"  Expected: registry.osism.tech/osism/osism:latest\",",
                                "        \"  Enabled:  true\",",
                                "        \"  Running:  registry.osism.tech/osism/osism:latest\",",
                                "        \"  Status:   \u2705 MATCH\",",
                                "        \"\",",
                                "        \"Checking service: listener (OpenStack Event Listener)\",",
                                "        \"  Expected: registry.osism.tech/osism/osism:latest\",",
                                "        \"  Enabled:  true\",",
                                "        \"  Running:  registry.osism.tech/osism/osism:latest\",",
                                "        \"  Status:   \u2705 MATCH\",",
                                "        \"\",",
                                "        \"Checking service: openstack (OpenStack Integration)\",",
                                "        \"  Expected: registry.osism.tech/osism/osism:latest\",",
                                "        \"  Enabled:  true\",",
                                "        \"  Running:  registry.osism.tech/osism/osism:latest\",",
                                "        \"  Status:   \u2705 MATCH\",",
                                "        \"\",",
                                "        \"Checking service: beat (Celery Beat Scheduler)\",",
                                "        \"  Expected: registry.osism.tech/osism/osism:latest\",",
                                "        \"  Enabled:  true\",",
                                "        \"  Running:  registry.osism.tech/osism/osism:latest\",",
                                "        \"  Status:   \u2705 MATCH\",",
                                "        \"\",",
                                "        \"Checking service: flower (Celery Flower Monitor)\",",
                                "        \"  Expected: registry.osism.tech/osism/osism:latest\",",
                                "        \"  Enabled:  true\",",
                                "        \"  Running:  registry.osism.tech/osism/osism:latest\",",
                                "        \"  Status:   \u2705 MATCH\",",
                                "        \"\",",
                                "        \"=== Summary ===\",",
                                "        \"Errors (version mismatches): 0\",",
                                "        \"Warnings (expected containers not running): 0\",",
                                "        \"\",",
                                "        \"\u2705 All running containers match expected versions!\"",
                                "    ]",
                                "}",
                                "",
                                "TASK [osism.services.manager : Skip version check due to service configuration] ***",
                                "skipping: [testbed-manager]",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "testbed-manager            : ok=70   changed=37   unreachable=0    failed=0    skipped=12   rescued=0    ignored=0   ",
                                "",
                                "+ [[ -e /opt/venv/bin/activate ]]",
                                "+ deactivate",
                                "+ '[' -n /usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/usr/games:/usr/local/games:/snap/bin ']'",
                                "+ PATH=/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/usr/games:/usr/local/games:/snap/bin",
                                "+ export PATH",
                                "+ unset _OLD_VIRTUAL_PATH",
                                "+ '[' -n '' ']'",
                                "+ hash -r",
                                "+ '[' -n '' ']'",
                                "+ unset VIRTUAL_ENV",
                                "+ unset VIRTUAL_ENV_PROMPT",
                                "+ '[' '!' '' = nondestructive ']'",
                                "+ unset -f deactivate",
                                "+ cp /home/dragon/.ssh/id_rsa.pub /opt/ansible/secrets/id_rsa.operator.pub",
                                "+ [[ ceph-ansible == \\c\\e\\p\\h\\-\\a\\n\\s\\i\\b\\l\\e ]]",
                                "+ wait_for_container_healthy 60 ceph-ansible",
                                "+ local max_attempts=60",
                                "+ local name=ceph-ansible",
                                "+ local attempt_num=1",
                                "++ /usr/bin/docker inspect -f '{{.State.Health.Status}}' ceph-ansible",
                                "+ [[ healthy == \\h\\e\\a\\l\\t\\h\\y ]]",
                                "+ wait_for_container_healthy 60 kolla-ansible",
                                "+ local max_attempts=60",
                                "+ local name=kolla-ansible",
                                "+ local attempt_num=1",
                                "++ /usr/bin/docker inspect -f '{{.State.Health.Status}}' kolla-ansible",
                                "+ [[ healthy == \\h\\e\\a\\l\\t\\h\\y ]]",
                                "+ wait_for_container_healthy 60 osism-ansible",
                                "+ local max_attempts=60",
                                "+ local name=osism-ansible",
                                "+ local attempt_num=1",
                                "++ /usr/bin/docker inspect -f '{{.State.Health.Status}}' osism-ansible",
                                "+ [[ healthy == \\h\\e\\a\\l\\t\\h\\y ]]",
                                "+ [[ true == \\t\\r\\u\\e ]]",
                                "+ sh -c /opt/configuration/scripts/disable-ara.sh",
                                "+ docker compose --project-directory /opt/manager ps",
                                "NAME                             IMAGE                                                      COMMAND                  SERVICE                CREATED              STATUS                        PORTS",
                                "ceph-ansible                     registry.osism.tech/osism/ceph-ansible:reef                \"/entrypoint.sh osis\u2026\"   ceph-ansible           About a minute ago   Up About a minute (healthy)   ",
                                "kolla-ansible                    registry.osism.tech/osism/kolla-ansible:2024.2             \"/entrypoint.sh osis\u2026\"   kolla-ansible          About a minute ago   Up About a minute (healthy)   ",
                                "manager-api-1                    registry.osism.tech/osism/osism:latest                     \"/sbin/tini -- osism\u2026\"   api                    About a minute ago   Up About a minute (healthy)   192.168.16.5:8000->8000/tcp",
                                "manager-ara-server-1             registry.osism.tech/osism/ara-server:1.7.3                 \"sh -c '/wait && /ru\u2026\"   ara-server             About a minute ago   Up About a minute (healthy)   8000/tcp",
                                "manager-beat-1                   registry.osism.tech/osism/osism:latest                     \"/sbin/tini -- osism\u2026\"   beat                   About a minute ago   Up About a minute (healthy)   ",
                                "manager-flower-1                 registry.osism.tech/osism/osism:latest                     \"/sbin/tini -- osism\u2026\"   flower                 About a minute ago   Up About a minute (healthy)   ",
                                "manager-inventory_reconciler-1   registry.osism.tech/osism/inventory-reconciler:latest      \"/sbin/tini -- /entr\u2026\"   inventory_reconciler   About a minute ago   Up 56 seconds (healthy)       ",
                                "manager-listener-1               registry.osism.tech/osism/osism:latest                     \"/sbin/tini -- osism\u2026\"   listener               About a minute ago   Up About a minute (healthy)   ",
                                "manager-mariadb-1                registry.osism.tech/dockerhub/library/mariadb:11.8.4       \"docker-entrypoint.s\u2026\"   mariadb                About a minute ago   Up About a minute (healthy)   3306/tcp",
                                "manager-openstack-1              registry.osism.tech/osism/osism:latest                     \"/sbin/tini -- osism\u2026\"   openstack              About a minute ago   Up About a minute (healthy)   ",
                                "manager-redis-1                  registry.osism.tech/dockerhub/library/redis:7.4.7-alpine   \"docker-entrypoint.s\u2026\"   redis                  About a minute ago   Up About a minute (healthy)   6379/tcp",
                                "osism-ansible                    registry.osism.tech/osism/osism-ansible:latest             \"/entrypoint.sh osis\u2026\"   osism-ansible          About a minute ago   Up About a minute (healthy)   ",
                                "osism-frontend                   registry.osism.tech/osism/osism-frontend:latest            \"docker-entrypoint.s\u2026\"   frontend               About a minute ago   Up About a minute             192.168.16.5:3000->3000/tcp",
                                "osism-kubernetes                 registry.osism.tech/osism/osism-kubernetes:latest          \"/entrypoint.sh osis\u2026\"   osism-kubernetes       About a minute ago   Up About a minute (healthy)   ",
                                "osismclient                      registry.osism.tech/osism/osism:latest                     \"/sbin/tini -- sleep\u2026\"   osismclient            About a minute ago   Up About a minute (healthy)   ",
                                "++ semver latest 7.0.0",
                                "+ [[ -1 -ge 0 ]]",
                                "+ [[ latest == \\l\\a\\t\\e\\s\\t ]]",
                                "+ sed -i s/community.general.yaml/osism.commons.still_alive/ /opt/configuration/environments/ansible.cfg",
                                "+ osism apply resolvconf -l testbed-manager",
                                "\u001b[32m2026-02-24 00:22:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mPrepare task for execution of resolvconf.\u001b[0m",
                                "\u001b[32m2026-02-24 00:22:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5a9194ec-1896-415a-bfdb-560ac86dbab4 (resolvconf) was prepared for execution.\u001b[0m",
                                "\u001b[32m2026-02-24 00:22:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task 5a9194ec-1896-415a-bfdb-560ac86dbab4 (resolvconf) has been started and output is visible here.\u001b[0m",
                                "",
                                "PLAY [Apply role resolvconf] ***************************************************",
                                "",
                                "TASK [Gathering Facts] *********************************************************",
                                "Tuesday 24 February 2026  00:22:04 +0000 (0:00:00.101)       0:00:00.101 ****** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.commons.resolvconf : Check minimum and maximum number of name servers] ***",
                                "Tuesday 24 February 2026  00:22:07 +0000 (0:00:03.314)       0:00:03.416 ****** ",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.commons.resolvconf : Include resolvconf tasks] *********************",
                                "Tuesday 24 February 2026  00:22:07 +0000 (0:00:00.063)       0:00:03.480 ****** ",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/resolvconf/tasks/configure-resolv.yml for testbed-manager\u001b[0m",
                                "",
                                "TASK [osism.commons.resolvconf : Include distribution specific installation tasks] ***",
                                "Tuesday 24 February 2026  00:22:07 +0000 (0:00:00.089)       0:00:03.569 ****** ",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/resolvconf/tasks/install-Debian-family.yml for testbed-manager\u001b[0m",
                                "",
                                "TASK [osism.commons.resolvconf : Remove packages configuring /etc/resolv.conf] ***",
                                "Tuesday 24 February 2026  00:22:07 +0000 (0:00:00.066)       0:00:03.635 ****** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.commons.resolvconf : Install package systemd-resolved] *************",
                                "Tuesday 24 February 2026  00:22:08 +0000 (0:00:00.866)       0:00:04.502 ****** ",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.commons.resolvconf : Retrieve file status of /etc/resolv.conf] *****",
                                "Tuesday 24 February 2026  00:22:08 +0000 (0:00:00.061)       0:00:04.564 ****** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.commons.resolvconf : Archive existing file /etc/resolv.conf] *******",
                                "Tuesday 24 February 2026  00:22:09 +0000 (0:00:01.440)       0:00:06.004 ****** ",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.commons.resolvconf : Link /run/systemd/resolve/stub-resolv.conf to /etc/resolv.conf] ***",
                                "Tuesday 24 February 2026  00:22:10 +0000 (0:00:00.072)       0:00:06.077 ****** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.commons.resolvconf : Copy configuration files] *********************",
                                "Tuesday 24 February 2026  00:22:10 +0000 (0:00:00.485)       0:00:06.562 ****** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.commons.resolvconf : Start/enable systemd-resolved service] ********",
                                "Tuesday 24 February 2026  00:22:11 +0000 (0:00:00.956)       0:00:07.519 ****** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.commons.resolvconf : Include distribution specific configuration tasks] ***",
                                "Tuesday 24 February 2026  00:22:12 +0000 (0:00:00.874)       0:00:08.394 ****** ",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/resolvconf/tasks/configure-Debian-family.yml for testbed-manager\u001b[0m",
                                "",
                                "TASK [osism.commons.resolvconf : Restart systemd-resolved service] *************",
                                "Tuesday 24 February 2026  00:22:12 +0000 (0:00:00.077)       0:00:08.471 ****** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;33mtestbed-manager\u001b[0m            : \u001b[0;32mok=10  \u001b[0m \u001b[0;33mchanged=3   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=3   \u001b[0m rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Tuesday 24 February 2026  00:22:13 +0000 (0:00:01.099)       0:00:09.571 ****** ",
                                "=============================================================================== ",
                                "Gathering Facts --------------------------------------------------------- 3.31s",
                                "osism.commons.resolvconf : Retrieve file status of /etc/resolv.conf ----- 1.44s",
                                "osism.commons.resolvconf : Restart systemd-resolved service ------------- 1.10s",
                                "osism.commons.resolvconf : Copy configuration files --------------------- 0.96s",
                                "osism.commons.resolvconf : Start/enable systemd-resolved service -------- 0.87s",
                                "osism.commons.resolvconf : Remove packages configuring /etc/resolv.conf --- 0.87s",
                                "osism.commons.resolvconf : Link /run/systemd/resolve/stub-resolv.conf to /etc/resolv.conf --- 0.49s",
                                "osism.commons.resolvconf : Include resolvconf tasks --------------------- 0.09s",
                                "osism.commons.resolvconf : Include distribution specific configuration tasks --- 0.08s",
                                "osism.commons.resolvconf : Archive existing file /etc/resolv.conf ------- 0.07s",
                                "osism.commons.resolvconf : Include distribution specific installation tasks --- 0.07s",
                                "osism.commons.resolvconf : Check minimum and maximum number of name servers --- 0.06s",
                                "osism.commons.resolvconf : Install package systemd-resolved ------------- 0.06s",
                                "+ osism apply sshconfig",
                                "\u001b[32m2026-02-24 00:22:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mPrepare task for execution of sshconfig.\u001b[0m",
                                "\u001b[32m2026-02-24 00:22:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 0ea583c6-54c2-46df-adef-59f07b9f4681 (sshconfig) was prepared for execution.\u001b[0m",
                                "\u001b[32m2026-02-24 00:22:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task 0ea583c6-54c2-46df-adef-59f07b9f4681 (sshconfig) has been started and output is visible here.\u001b[0m",
                                "",
                                "PLAY [Apply role sshconfig] ****************************************************",
                                "",
                                "TASK [osism.commons.sshconfig : Get home directory of operator user] ***********",
                                "Tuesday 24 February 2026  00:22:29 +0000 (0:00:00.117)       0:00:00.117 ****** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.commons.sshconfig : Ensure .ssh/config.d exist] ********************",
                                "Tuesday 24 February 2026  00:22:30 +0000 (0:00:00.467)       0:00:00.584 ****** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.commons.sshconfig : Ensure config for each host exist] *************",
                                "Tuesday 24 February 2026  00:22:30 +0000 (0:00:00.365)       0:00:00.950 ****** ",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=testbed-node-0)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=testbed-node-1)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=testbed-node-2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=testbed-node-3)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=testbed-node-4)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=testbed-node-5)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=testbed-manager)\u001b[0m",
                                "",
                                "TASK [osism.commons.sshconfig : Add extra config] ******************************",
                                "Tuesday 24 February 2026  00:22:35 +0000 (0:00:05.158)       0:00:06.109 ****** ",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.commons.sshconfig : Assemble ssh config] ***************************",
                                "Tuesday 24 February 2026  00:22:35 +0000 (0:00:00.077)       0:00:06.186 ****** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;33mtestbed-manager\u001b[0m            : \u001b[0;32mok=4   \u001b[0m \u001b[0;33mchanged=3   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=1   \u001b[0m rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Tuesday 24 February 2026  00:22:36 +0000 (0:00:00.568)       0:00:06.755 ****** ",
                                "=============================================================================== ",
                                "osism.commons.sshconfig : Ensure config for each host exist ------------- 5.16s",
                                "osism.commons.sshconfig : Assemble ssh config --------------------------- 0.57s",
                                "osism.commons.sshconfig : Get home directory of operator user ----------- 0.47s",
                                "osism.commons.sshconfig : Ensure .ssh/config.d exist -------------------- 0.37s",
                                "osism.commons.sshconfig : Add extra config ------------------------------ 0.08s",
                                "+ osism apply known-hosts",
                                "\u001b[32m2026-02-24 00:22:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mPrepare task for execution of known-hosts.\u001b[0m",
                                "\u001b[32m2026-02-24 00:22:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 688403ab-c18b-4137-889d-96137831af39 (known-hosts) was prepared for execution.\u001b[0m",
                                "\u001b[32m2026-02-24 00:22:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task 688403ab-c18b-4137-889d-96137831af39 (known-hosts) has been started and output is visible here.\u001b[0m",
                                "",
                                "PLAY [Apply role known_hosts] **************************************************",
                                "",
                                "TASK [osism.commons.known_hosts : Run ssh-keyscan for all hosts with hostname] ***",
                                "Tuesday 24 February 2026  00:22:52 +0000 (0:00:00.166)       0:00:00.166 ****** ",
                                "\u001b[0;32mok: [testbed-manager] => (item=testbed-node-3)\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager] => (item=testbed-node-4)\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager] => (item=testbed-node-5)\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager] => (item=testbed-manager)\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager] => (item=testbed-node-0)\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager] => (item=testbed-node-1)\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager] => (item=testbed-node-2)\u001b[0m",
                                "",
                                "TASK [osism.commons.known_hosts : Write scanned known_hosts entries for all hosts with hostname] ***",
                                "Tuesday 24 February 2026  00:22:58 +0000 (0:00:05.860)       0:00:06.026 ****** ",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/known_hosts/tasks/write-scanned.yml for testbed-manager => (item=Scanned entries of testbed-node-3)\u001b[0m",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/known_hosts/tasks/write-scanned.yml for testbed-manager => (item=Scanned entries of testbed-node-4)\u001b[0m",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/known_hosts/tasks/write-scanned.yml for testbed-manager => (item=Scanned entries of testbed-node-5)\u001b[0m",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/known_hosts/tasks/write-scanned.yml for testbed-manager => (item=Scanned entries of testbed-manager)\u001b[0m",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/known_hosts/tasks/write-scanned.yml for testbed-manager => (item=Scanned entries of testbed-node-0)\u001b[0m",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/known_hosts/tasks/write-scanned.yml for testbed-manager => (item=Scanned entries of testbed-node-1)\u001b[0m",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/known_hosts/tasks/write-scanned.yml for testbed-manager => (item=Scanned entries of testbed-node-2)\u001b[0m",
                                "",
                                "TASK [osism.commons.known_hosts : Write scanned known_hosts entries] ***********",
                                "Tuesday 24 February 2026  00:22:58 +0000 (0:00:00.161)       0:00:06.188 ****** ",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=testbed-node-3 ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABgQDOWRdk1N8OIeC4uCbXt2ifzDinxCcJxpFg+cy8/O9qdNiU6uF6MYSNDWdTmj5LsQfBOv9B2tGl+oFtMoV6HFtFDsUgRunw1ZTuofhokH0eYBBrukiCLi7dxc7SnmFZ2ITcQjZB2BhCHKwtvOxTBksVGswVmjoPGjlciYKDTaeLZD3zYiLjKGP1UKtIfjzrxuwVuHbzGMwnExq8l2JeYF84/mrQP2FRBfH/UkCVyuvZnHhQAHRoURNmhrCwaepBbx2ZaZNUEdmxJnkJMWJtvpk3C7BPgiGEPwLlhM79cqKzgtFZkZ+7/qZCZACjPF9aaTds5utFpV74TDHtAQrUKM5KAI7dFIf9jeNVWw6r0vgL44maMvPIYQD6vinwzlujJKG0cBCq678Up0uTRF07q5NpCBf0oGWOC8LoxdycSfI2Vs3ljaUt+b3T4R0BO5sNT2JCHRdcUNZwFAGwZ4RW4RCGD0j5w26RMy9pp+ofRgsAUEYDHRkZ/4WHfMxIUMjwmU0=)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=testbed-node-3 ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBL2uknstYHYYZxd3ycHLBWlZpyvOW+d9YIVVFgEYz3Zxjh4zS4EdPoYhodFG1Qr3nQQ5Rvp4lzvZiJjxhFxTZ28=)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=testbed-node-3 ssh-ed25519 AAAAC3NzaC1lZDI1NTE5AAAAIN2iQH254D6gxizx/kDK2ek50uPnRRnQlJj02mWpt5cH)\u001b[0m",
                                "",
                                "TASK [osism.commons.known_hosts : Write scanned known_hosts entries] ***********",
                                "Tuesday 24 February 2026  00:22:59 +0000 (0:00:01.122)       0:00:07.311 ****** ",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=testbed-node-4 ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABgQCQjWzk/nNVedMF8uWXwcthkgBn54kNzWDVJ5eAPCCqD++s091Bnx29/cjP7Q+lEdxp8BnEg6BnF1cH9i1RmNbAgQCZEjLF7WwCK0M9ptx6Ebp7M9j0UtRuFhtNBOIefbKOwTKJw21yOvzNX2xzvH8/WS1ODjJeH0ZbhWhXHgdKisLQjwJmaxGD+nYxDFAhWFjcfBE5kb+STYmdxujCZ6b8pARZS4XJUEPFZZgx7U0jylvG8nLUN7CgCxy9breJpHMKtjk3Mm7zSrH0Nhop1eiz4c/v8/Zag21EPO9jo1R3fvIOeAYIvQ3eQdBfRULWc89Ijgdvq8z6+OeFRx0wCP6Zzcw9w3nE1if7IkikF4MDaCXIIlf/KW77/6r1iWWol+aOBNR64BA9d9VjstePAQAHlQ/a/qONP5kD1NRXAEhAwuJzdxjvX7RjyuntHNE6n/kcDEfacZKEMThOs/WyIVRVkAgvjgj9wSkSYMqp1kkhKY5QsfOocegMHtkKtqb4FB8=)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=testbed-node-4 ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBOly8FEun0/C8rfMAzTzUQzKUsbnR8IbBcxxSLcEEY/FPTQCBr4O94lHghhDYXhW3M0OoKHX8XqwcnbnJumkl4g=)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=testbed-node-4 ssh-ed25519 AAAAC3NzaC1lZDI1NTE5AAAAINkXgg1KUois1zDu7aBhB83ATBEugpnY7XNMswDNHl3w)\u001b[0m",
                                "",
                                "TASK [osism.commons.known_hosts : Write scanned known_hosts entries] ***********",
                                "Tuesday 24 February 2026  00:23:00 +0000 (0:00:01.017)       0:00:08.328 ****** ",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=testbed-node-5 ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBAeSrLLHw6GzSjU89slqhT47VMqLM9hsem1yhDpwlj3R+Q4qo3kzcxiGcXhAd6bD1fjqCK3Wfu1+cKO9fhEy1g4=)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=testbed-node-5 ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABgQCsN9mDKfSY623Rz4GeNVRKPyAQ2P/kyJ1M+ciS9jyXpN6jHd1oAzAqVipldMZnQkPqLtbfhsAgFdIVodbBCIlTURpF1N64E61lwTsipSvaqRW+gWwNykSp0Kd2tgjA8QDzk0CdLtTIR0Y3G2xepnzsMFF7ybQyNx80kovfbvbA/M5u8pekTSQ9hO27dVqafm/Rkg6Se7e7C3Va9Bunhj+O5UUJofrtdFJRpOMYUilHjqXgehDZpamdm9zIFDyxlXLqB/oJmcr3Zn9JFLgwMuZfSc+z8G/c5H1lJLYJr0Pd/0rsb826cqxKrFy//JDih1XP8Q/66TDP9wmvBJts2MrsmXe1d6zukaxINB436jntilYM+bzRL+8dH9eXexj/sUZZiBCyfGAlTJr+I4RARGU49FvPvKQtz1SQhQoUB1OR/6FooWx6yv1CA/Dw5PiZuSvECNDR++LE6ZvuLqAMvwVHm7zxgSx5q0VNtIqN+PTHN6yq9r2rnhErhLoOuQThUm8=)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=testbed-node-5 ssh-ed25519 AAAAC3NzaC1lZDI1NTE5AAAAIP4XvFEG+/G6UU398M8h8QkghZwwa+RxCDcBl6tu+MbK)\u001b[0m",
                                "",
                                "TASK [osism.commons.known_hosts : Write scanned known_hosts entries] ***********",
                                "Tuesday 24 February 2026  00:23:01 +0000 (0:00:01.000)       0:00:09.329 ****** ",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=testbed-manager ssh-ed25519 AAAAC3NzaC1lZDI1NTE5AAAAIDfU67w4g0t5ruLoohFtsOVj0FEJKCFA6yTUXHoTeiU+)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=testbed-manager ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABgQCUeJXF5lVxZV05KJ56am57/G4Pj6M8Ir07kqn5lXoOZXaKtAM034AYvC6FfkuxmYUsc69IbF5cYYNzWSTi7uigANnV3Gtkq8OlDelwVdV3vQDp+71dO6t30HlzvuEvk4qjeYkciisK/RE5RqVZCkqkMQIQR7TunEGspNpQyMP0Jb6CHkRMFAhPzIDv+Hcun0aWylP6WK+44TYcReX6RIOILPhPUXYV9H0ZFoC2L/hup+kkY1S85c8WaqH/OcR/XnOLwODXyRpMm2DX9kwhU0QeIeO+VqU4Jp6aXAfjSLbFbgAcm7LMAA7qdTd3liWDXJO22pgua2aAolsOe8gcFWZtMlD0AwtOUBd5c2MFrrn85aoaINKiXu/fOJxsV67SX729tb+1aXygYdUsjDxJBNHKdkmaLGpWBVgkz++0yy6DS6viWLi2nDL/3p80JUPbLv57s8zpev1x9ZMHJxbWkCjGgqjZ7HZ2WAsNklDWNgvZ03i/KiiUoaCCDjIofbR2R0E=)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=testbed-manager ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBAAEM4MNA1yZvjsAJIz6pihEIhOsw75Cu+tJPTJEbZDNCdZ3BprKWAIEwVr1UFlQTAVPSVDKMXHBd3XzGkopW84=)\u001b[0m",
                                "",
                                "TASK [osism.commons.known_hosts : Write scanned known_hosts entries] ***********",
                                "Tuesday 24 February 2026  00:23:02 +0000 (0:00:00.982)       0:00:10.312 ****** ",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=testbed-node-0 ssh-ed25519 AAAAC3NzaC1lZDI1NTE5AAAAIC9+a2YIH9Deu7Kp5MuasTBGpKoKvAwa8BiKo1b3H76+)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=testbed-node-0 ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABgQC1zWCyl2+3CBj3nqxQhXBIhbNhR0LfsVZ8HtUXLylfIQOHN5bMATCm8HbT8KZ3HbGbluD5kffIIFrqxlSvMyuKnSFBlIwg1dSEaJztZQm1QPmZEkPMD+C+8HudVc5itevm9fwKDVWPt0lZ8zKpFxzF6Rum/bfFKJQcmPxl+zIL46mr4PbBLibBlHbAUs1ZnkZXwGXtkghKmrl8lvFskCXPllSKZcLPeM6QImpZa99TwLNhC1FnkJhLWtW25iDh53zIkoCFizujMM38wHNGIiu7rKB/CMgnX8LsKYZoSYatqHivgegBJd79m3myxBkGOsQ0YmCGGeUPRpOzabncaSPmDRsWpTafchsA7JJX6xXc4Q8+v/Zx+CNOg2oPpPnMJegXxgBAMkXeiWCaB5sbHLFmCD8eBFrOxtBeS0RqaBSHylkQBJ4q6aNJ2jIK0i2XnDVTbTrfhwkeIY9Uz5z8DuKm9HJ3x6rPJofnDSdPLdXXT3XDnj6B1riH5sVGIbKRglk=)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=testbed-node-0 ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBA+CLgYVB+RtOt+86BTUxoGVVBxkcCRXmZpt5JEy6gYHueijrlVosdVEchOoU/pc/10Ze4k+BS7Mg3o6u6GUV8s=)\u001b[0m",
                                "",
                                "TASK [osism.commons.known_hosts : Write scanned known_hosts entries] ***********",
                                "Tuesday 24 February 2026  00:23:03 +0000 (0:00:01.043)       0:00:11.355 ****** ",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=testbed-node-1 ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBPjiHNE7pTy7o8Vd28ZoO+w4U30fFOydWvvQBQp9qayGswAO7W20tyqB2rWFbwwhAjAYAGZX2sFLF0X/ulaZkVA=)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=testbed-node-1 ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABgQCbVubwVzGRukWm151/SYPu5+ZUCffbxzUGqgLxjp6PNUEvt1yzTcQ1bVHQ4y09uUCZjwDql6LL3gOIfV6fpdzXAd/FEDQt2fSoHvRXxH8BafXLEVwctwrdcoLPcJkpgl+qMgysgFgL9bHa0Rd6jP2Npur0vk3Y5ju723LOiRroRByOYiHC4XNS7DxHu6u+L/D6hVqQz5jwzP0tDXCF/+LNJ6uiaTDOd+sywTVq7ITDItNLaXIZGpsNjubQy1lMqyVZeaI5Le1/wwJDWhsrBrIyFm3Kemhecn/PVJZmPPk2OFvWp48Uyo0+NubcquE50ezJVbAmhbFG5c6HUrpVHcWLNk4LzZ6F5bRpR6V42QvBvcgwnMtzBw8QRXf96EFtzv4K5EG4Y+1Ew1sWzjWXYD/ZlA6Cc9qu5D23ebFPZIiY8o8dhBlj4iXHxO3kABxCDEJS49apyfvvKwFQ5+FjMCy88w1eO5LOwT8BXpUXcXgNcNWHidtPZ4ufzsqLYoP9Fhk=)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=testbed-node-1 ssh-ed25519 AAAAC3NzaC1lZDI1NTE5AAAAIPlYo1pFim6htda1AoAJkI0dbcELCmM5QFZYRxrp2T8p)\u001b[0m",
                                "",
                                "TASK [osism.commons.known_hosts : Write scanned known_hosts entries] ***********",
                                "Tuesday 24 February 2026  00:23:04 +0000 (0:00:00.978)       0:00:12.334 ****** ",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=testbed-node-2 ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBNMB0mFLquOvLDLHWMx+FbOs6ia8nAHzsLjwVkHj6z+8y/yvQBEmH3hQRj/jml5GScCC1Ncbk/HPjJCUU52aLP4=)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=testbed-node-2 ssh-ed25519 AAAAC3NzaC1lZDI1NTE5AAAAID9TiTT29JB5SGhfDhxbtI4boFbWaY41+GMclNzBxg2p)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=testbed-node-2 ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABgQDCMRk074A67PfDSfiSZySlWWdBg82NzopzXCqwhSXtyjr/LVN4q3Z3VWst7PL1hjux2gaZvX6+h/RWI8ub/hOF8VMqPQ90NdJOy53UuMtsrIlovBWpDHMUa3v5FrrtAFY1Fve1POj+5kcPG69xUj/N6wLkiLab8+VqDnJLia7IsrYFnU4xShfL963gkMDvZvamLExqWws9ZZHfZc37vaYL9ntX3bcmjW76hTp+q/jMjXIxy5MRNZP+0YLq6Ig7BQb8VrrHSMx0FQgWX0UGso0BQQcaRjUl5T9yWYCJBMt3ciRZvs+Rordy878/xbYLADmgU4Vyy6RFkGxkoOQPiNMtZfMx0kLSqd8FYlysLgjvUjKFAv5uB3MAWD2Sp4CSIKI/Hbn7lzTEGIcgUopve3iif2R+oTGLje8I2g0Tqbxl53+oOWjZrPnYuCFarbJ2siPpQljjEVmsI1Y72JHFcQyRpJUp0rReRH1BpnxiPIFzr76k4hIgPNqbBi0o8zJri2k=)\u001b[0m",
                                "",
                                "TASK [osism.commons.known_hosts : Run ssh-keyscan for all hosts with ansible_host] ***",
                                "Tuesday 24 February 2026  00:23:05 +0000 (0:00:00.981)       0:00:13.315 ****** ",
                                "\u001b[0;32mok: [testbed-manager] => (item=testbed-node-3)\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager] => (item=testbed-node-4)\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager] => (item=testbed-node-5)\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager] => (item=testbed-manager)\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager] => (item=testbed-node-0)\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager] => (item=testbed-node-1)\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager] => (item=testbed-node-2)\u001b[0m",
                                "",
                                "TASK [osism.commons.known_hosts : Write scanned known_hosts entries for all hosts with ansible_host] ***",
                                "Tuesday 24 February 2026  00:23:10 +0000 (0:00:05.048)       0:00:18.363 ****** ",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/known_hosts/tasks/write-scanned.yml for testbed-manager => (item=Scanned entries of testbed-node-3)\u001b[0m",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/known_hosts/tasks/write-scanned.yml for testbed-manager => (item=Scanned entries of testbed-node-4)\u001b[0m",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/known_hosts/tasks/write-scanned.yml for testbed-manager => (item=Scanned entries of testbed-node-5)\u001b[0m",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/known_hosts/tasks/write-scanned.yml for testbed-manager => (item=Scanned entries of testbed-manager)\u001b[0m",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/known_hosts/tasks/write-scanned.yml for testbed-manager => (item=Scanned entries of testbed-node-0)\u001b[0m",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/known_hosts/tasks/write-scanned.yml for testbed-manager => (item=Scanned entries of testbed-node-1)\u001b[0m",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/known_hosts/tasks/write-scanned.yml for testbed-manager => (item=Scanned entries of testbed-node-2)\u001b[0m",
                                "",
                                "TASK [osism.commons.known_hosts : Write scanned known_hosts entries] ***********",
                                "Tuesday 24 February 2026  00:23:11 +0000 (0:00:00.171)       0:00:18.535 ****** ",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=192.168.16.13 ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBL2uknstYHYYZxd3ycHLBWlZpyvOW+d9YIVVFgEYz3Zxjh4zS4EdPoYhodFG1Qr3nQQ5Rvp4lzvZiJjxhFxTZ28=)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=192.168.16.13 ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABgQDOWRdk1N8OIeC4uCbXt2ifzDinxCcJxpFg+cy8/O9qdNiU6uF6MYSNDWdTmj5LsQfBOv9B2tGl+oFtMoV6HFtFDsUgRunw1ZTuofhokH0eYBBrukiCLi7dxc7SnmFZ2ITcQjZB2BhCHKwtvOxTBksVGswVmjoPGjlciYKDTaeLZD3zYiLjKGP1UKtIfjzrxuwVuHbzGMwnExq8l2JeYF84/mrQP2FRBfH/UkCVyuvZnHhQAHRoURNmhrCwaepBbx2ZaZNUEdmxJnkJMWJtvpk3C7BPgiGEPwLlhM79cqKzgtFZkZ+7/qZCZACjPF9aaTds5utFpV74TDHtAQrUKM5KAI7dFIf9jeNVWw6r0vgL44maMvPIYQD6vinwzlujJKG0cBCq678Up0uTRF07q5NpCBf0oGWOC8LoxdycSfI2Vs3ljaUt+b3T4R0BO5sNT2JCHRdcUNZwFAGwZ4RW4RCGD0j5w26RMy9pp+ofRgsAUEYDHRkZ/4WHfMxIUMjwmU0=)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=192.168.16.13 ssh-ed25519 AAAAC3NzaC1lZDI1NTE5AAAAIN2iQH254D6gxizx/kDK2ek50uPnRRnQlJj02mWpt5cH)\u001b[0m",
                                "",
                                "TASK [osism.commons.known_hosts : Write scanned known_hosts entries] ***********",
                                "Tuesday 24 February 2026  00:23:12 +0000 (0:00:01.013)       0:00:19.549 ****** ",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=192.168.16.14 ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBOly8FEun0/C8rfMAzTzUQzKUsbnR8IbBcxxSLcEEY/FPTQCBr4O94lHghhDYXhW3M0OoKHX8XqwcnbnJumkl4g=)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=192.168.16.14 ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABgQCQjWzk/nNVedMF8uWXwcthkgBn54kNzWDVJ5eAPCCqD++s091Bnx29/cjP7Q+lEdxp8BnEg6BnF1cH9i1RmNbAgQCZEjLF7WwCK0M9ptx6Ebp7M9j0UtRuFhtNBOIefbKOwTKJw21yOvzNX2xzvH8/WS1ODjJeH0ZbhWhXHgdKisLQjwJmaxGD+nYxDFAhWFjcfBE5kb+STYmdxujCZ6b8pARZS4XJUEPFZZgx7U0jylvG8nLUN7CgCxy9breJpHMKtjk3Mm7zSrH0Nhop1eiz4c/v8/Zag21EPO9jo1R3fvIOeAYIvQ3eQdBfRULWc89Ijgdvq8z6+OeFRx0wCP6Zzcw9w3nE1if7IkikF4MDaCXIIlf/KW77/6r1iWWol+aOBNR64BA9d9VjstePAQAHlQ/a/qONP5kD1NRXAEhAwuJzdxjvX7RjyuntHNE6n/kcDEfacZKEMThOs/WyIVRVkAgvjgj9wSkSYMqp1kkhKY5QsfOocegMHtkKtqb4FB8=)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=192.168.16.14 ssh-ed25519 AAAAC3NzaC1lZDI1NTE5AAAAINkXgg1KUois1zDu7aBhB83ATBEugpnY7XNMswDNHl3w)\u001b[0m",
                                "",
                                "TASK [osism.commons.known_hosts : Write scanned known_hosts entries] ***********",
                                "Tuesday 24 February 2026  00:23:13 +0000 (0:00:01.018)       0:00:20.568 ****** ",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=192.168.16.15 ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABgQCsN9mDKfSY623Rz4GeNVRKPyAQ2P/kyJ1M+ciS9jyXpN6jHd1oAzAqVipldMZnQkPqLtbfhsAgFdIVodbBCIlTURpF1N64E61lwTsipSvaqRW+gWwNykSp0Kd2tgjA8QDzk0CdLtTIR0Y3G2xepnzsMFF7ybQyNx80kovfbvbA/M5u8pekTSQ9hO27dVqafm/Rkg6Se7e7C3Va9Bunhj+O5UUJofrtdFJRpOMYUilHjqXgehDZpamdm9zIFDyxlXLqB/oJmcr3Zn9JFLgwMuZfSc+z8G/c5H1lJLYJr0Pd/0rsb826cqxKrFy//JDih1XP8Q/66TDP9wmvBJts2MrsmXe1d6zukaxINB436jntilYM+bzRL+8dH9eXexj/sUZZiBCyfGAlTJr+I4RARGU49FvPvKQtz1SQhQoUB1OR/6FooWx6yv1CA/Dw5PiZuSvECNDR++LE6ZvuLqAMvwVHm7zxgSx5q0VNtIqN+PTHN6yq9r2rnhErhLoOuQThUm8=)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=192.168.16.15 ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBAeSrLLHw6GzSjU89slqhT47VMqLM9hsem1yhDpwlj3R+Q4qo3kzcxiGcXhAd6bD1fjqCK3Wfu1+cKO9fhEy1g4=)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=192.168.16.15 ssh-ed25519 AAAAC3NzaC1lZDI1NTE5AAAAIP4XvFEG+/G6UU398M8h8QkghZwwa+RxCDcBl6tu+MbK)\u001b[0m",
                                "",
                                "TASK [osism.commons.known_hosts : Write scanned known_hosts entries] ***********",
                                "Tuesday 24 February 2026  00:23:14 +0000 (0:00:01.005)       0:00:21.573 ****** ",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=192.168.16.5 ssh-ed25519 AAAAC3NzaC1lZDI1NTE5AAAAIDfU67w4g0t5ruLoohFtsOVj0FEJKCFA6yTUXHoTeiU+)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=192.168.16.5 ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABgQCUeJXF5lVxZV05KJ56am57/G4Pj6M8Ir07kqn5lXoOZXaKtAM034AYvC6FfkuxmYUsc69IbF5cYYNzWSTi7uigANnV3Gtkq8OlDelwVdV3vQDp+71dO6t30HlzvuEvk4qjeYkciisK/RE5RqVZCkqkMQIQR7TunEGspNpQyMP0Jb6CHkRMFAhPzIDv+Hcun0aWylP6WK+44TYcReX6RIOILPhPUXYV9H0ZFoC2L/hup+kkY1S85c8WaqH/OcR/XnOLwODXyRpMm2DX9kwhU0QeIeO+VqU4Jp6aXAfjSLbFbgAcm7LMAA7qdTd3liWDXJO22pgua2aAolsOe8gcFWZtMlD0AwtOUBd5c2MFrrn85aoaINKiXu/fOJxsV67SX729tb+1aXygYdUsjDxJBNHKdkmaLGpWBVgkz++0yy6DS6viWLi2nDL/3p80JUPbLv57s8zpev1x9ZMHJxbWkCjGgqjZ7HZ2WAsNklDWNgvZ03i/KiiUoaCCDjIofbR2R0E=)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=192.168.16.5 ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBAAEM4MNA1yZvjsAJIz6pihEIhOsw75Cu+tJPTJEbZDNCdZ3BprKWAIEwVr1UFlQTAVPSVDKMXHBd3XzGkopW84=)\u001b[0m",
                                "",
                                "TASK [osism.commons.known_hosts : Write scanned known_hosts entries] ***********",
                                "Tuesday 24 February 2026  00:23:15 +0000 (0:00:01.013)       0:00:22.587 ****** ",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=192.168.16.10 ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBA+CLgYVB+RtOt+86BTUxoGVVBxkcCRXmZpt5JEy6gYHueijrlVosdVEchOoU/pc/10Ze4k+BS7Mg3o6u6GUV8s=)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=192.168.16.10 ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABgQC1zWCyl2+3CBj3nqxQhXBIhbNhR0LfsVZ8HtUXLylfIQOHN5bMATCm8HbT8KZ3HbGbluD5kffIIFrqxlSvMyuKnSFBlIwg1dSEaJztZQm1QPmZEkPMD+C+8HudVc5itevm9fwKDVWPt0lZ8zKpFxzF6Rum/bfFKJQcmPxl+zIL46mr4PbBLibBlHbAUs1ZnkZXwGXtkghKmrl8lvFskCXPllSKZcLPeM6QImpZa99TwLNhC1FnkJhLWtW25iDh53zIkoCFizujMM38wHNGIiu7rKB/CMgnX8LsKYZoSYatqHivgegBJd79m3myxBkGOsQ0YmCGGeUPRpOzabncaSPmDRsWpTafchsA7JJX6xXc4Q8+v/Zx+CNOg2oPpPnMJegXxgBAMkXeiWCaB5sbHLFmCD8eBFrOxtBeS0RqaBSHylkQBJ4q6aNJ2jIK0i2XnDVTbTrfhwkeIY9Uz5z8DuKm9HJ3x6rPJofnDSdPLdXXT3XDnj6B1riH5sVGIbKRglk=)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=192.168.16.10 ssh-ed25519 AAAAC3NzaC1lZDI1NTE5AAAAIC9+a2YIH9Deu7Kp5MuasTBGpKoKvAwa8BiKo1b3H76+)\u001b[0m",
                                "",
                                "TASK [osism.commons.known_hosts : Write scanned known_hosts entries] ***********",
                                "Tuesday 24 February 2026  00:23:16 +0000 (0:00:01.020)       0:00:23.607 ****** ",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=192.168.16.11 ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABgQCbVubwVzGRukWm151/SYPu5+ZUCffbxzUGqgLxjp6PNUEvt1yzTcQ1bVHQ4y09uUCZjwDql6LL3gOIfV6fpdzXAd/FEDQt2fSoHvRXxH8BafXLEVwctwrdcoLPcJkpgl+qMgysgFgL9bHa0Rd6jP2Npur0vk3Y5ju723LOiRroRByOYiHC4XNS7DxHu6u+L/D6hVqQz5jwzP0tDXCF/+LNJ6uiaTDOd+sywTVq7ITDItNLaXIZGpsNjubQy1lMqyVZeaI5Le1/wwJDWhsrBrIyFm3Kemhecn/PVJZmPPk2OFvWp48Uyo0+NubcquE50ezJVbAmhbFG5c6HUrpVHcWLNk4LzZ6F5bRpR6V42QvBvcgwnMtzBw8QRXf96EFtzv4K5EG4Y+1Ew1sWzjWXYD/ZlA6Cc9qu5D23ebFPZIiY8o8dhBlj4iXHxO3kABxCDEJS49apyfvvKwFQ5+FjMCy88w1eO5LOwT8BXpUXcXgNcNWHidtPZ4ufzsqLYoP9Fhk=)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=192.168.16.11 ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBPjiHNE7pTy7o8Vd28ZoO+w4U30fFOydWvvQBQp9qayGswAO7W20tyqB2rWFbwwhAjAYAGZX2sFLF0X/ulaZkVA=)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=192.168.16.11 ssh-ed25519 AAAAC3NzaC1lZDI1NTE5AAAAIPlYo1pFim6htda1AoAJkI0dbcELCmM5QFZYRxrp2T8p)\u001b[0m",
                                "",
                                "TASK [osism.commons.known_hosts : Write scanned known_hosts entries] ***********",
                                "Tuesday 24 February 2026  00:23:17 +0000 (0:00:00.988)       0:00:24.595 ****** ",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=192.168.16.12 ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBNMB0mFLquOvLDLHWMx+FbOs6ia8nAHzsLjwVkHj6z+8y/yvQBEmH3hQRj/jml5GScCC1Ncbk/HPjJCUU52aLP4=)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=192.168.16.12 ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABgQDCMRk074A67PfDSfiSZySlWWdBg82NzopzXCqwhSXtyjr/LVN4q3Z3VWst7PL1hjux2gaZvX6+h/RWI8ub/hOF8VMqPQ90NdJOy53UuMtsrIlovBWpDHMUa3v5FrrtAFY1Fve1POj+5kcPG69xUj/N6wLkiLab8+VqDnJLia7IsrYFnU4xShfL963gkMDvZvamLExqWws9ZZHfZc37vaYL9ntX3bcmjW76hTp+q/jMjXIxy5MRNZP+0YLq6Ig7BQb8VrrHSMx0FQgWX0UGso0BQQcaRjUl5T9yWYCJBMt3ciRZvs+Rordy878/xbYLADmgU4Vyy6RFkGxkoOQPiNMtZfMx0kLSqd8FYlysLgjvUjKFAv5uB3MAWD2Sp4CSIKI/Hbn7lzTEGIcgUopve3iif2R+oTGLje8I2g0Tqbxl53+oOWjZrPnYuCFarbJ2siPpQljjEVmsI1Y72JHFcQyRpJUp0rReRH1BpnxiPIFzr76k4hIgPNqbBi0o8zJri2k=)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=192.168.16.12 ssh-ed25519 AAAAC3NzaC1lZDI1NTE5AAAAID9TiTT29JB5SGhfDhxbtI4boFbWaY41+GMclNzBxg2p)\u001b[0m",
                                "",
                                "TASK [osism.commons.known_hosts : Write static known_hosts entries] ************",
                                "Tuesday 24 February 2026  00:23:18 +0000 (0:00:01.014)       0:00:25.609 ****** ",
                                "\u001b[0;36mskipping: [testbed-manager] => (item=testbed-node-3) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager] => (item=testbed-node-4) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager] => (item=testbed-node-5) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager] => (item=testbed-manager) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager] => (item=testbed-node-0) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager] => (item=testbed-node-1) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager] => (item=testbed-node-2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.commons.known_hosts : Write extra known_hosts entries] *************",
                                "Tuesday 24 February 2026  00:23:18 +0000 (0:00:00.161)       0:00:25.771 ****** ",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.commons.known_hosts : Delete known_hosts entries] ******************",
                                "Tuesday 24 February 2026  00:23:18 +0000 (0:00:00.043)       0:00:25.814 ****** ",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.commons.known_hosts : Set file permissions] ************************",
                                "Tuesday 24 February 2026  00:23:18 +0000 (0:00:00.050)       0:00:25.865 ****** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;33mtestbed-manager\u001b[0m            : \u001b[0;32mok=31  \u001b[0m \u001b[0;33mchanged=15  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=3   \u001b[0m rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Tuesday 24 February 2026  00:23:19 +0000 (0:00:00.700)       0:00:26.565 ****** ",
                                "=============================================================================== ",
                                "osism.commons.known_hosts : Run ssh-keyscan for all hosts with hostname --- 5.86s",
                                "osism.commons.known_hosts : Run ssh-keyscan for all hosts with ansible_host --- 5.05s",
                                "osism.commons.known_hosts : Write scanned known_hosts entries ----------- 1.12s",
                                "osism.commons.known_hosts : Write scanned known_hosts entries ----------- 1.04s",
                                "osism.commons.known_hosts : Write scanned known_hosts entries ----------- 1.02s",
                                "osism.commons.known_hosts : Write scanned known_hosts entries ----------- 1.02s",
                                "osism.commons.known_hosts : Write scanned known_hosts entries ----------- 1.02s",
                                "osism.commons.known_hosts : Write scanned known_hosts entries ----------- 1.01s",
                                "osism.commons.known_hosts : Write scanned known_hosts entries ----------- 1.01s",
                                "osism.commons.known_hosts : Write scanned known_hosts entries ----------- 1.01s",
                                "osism.commons.known_hosts : Write scanned known_hosts entries ----------- 1.01s",
                                "osism.commons.known_hosts : Write scanned known_hosts entries ----------- 1.00s",
                                "osism.commons.known_hosts : Write scanned known_hosts entries ----------- 0.99s",
                                "osism.commons.known_hosts : Write scanned known_hosts entries ----------- 0.98s",
                                "osism.commons.known_hosts : Write scanned known_hosts entries ----------- 0.98s",
                                "osism.commons.known_hosts : Write scanned known_hosts entries ----------- 0.98s",
                                "osism.commons.known_hosts : Set file permissions ------------------------ 0.70s",
                                "osism.commons.known_hosts : Write scanned known_hosts entries for all hosts with ansible_host --- 0.17s",
                                "osism.commons.known_hosts : Write scanned known_hosts entries for all hosts with hostname --- 0.16s",
                                "osism.commons.known_hosts : Write static known_hosts entries ------------ 0.16s",
                                "+ osism apply squid",
                                "\u001b[32m2026-02-24 00:23:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mPrepare task for execution of squid.\u001b[0m",
                                "\u001b[32m2026-02-24 00:23:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5bf1eadf-39a3-4b5b-9eb1-63c073d32e17 (squid) was prepared for execution.\u001b[0m",
                                "\u001b[32m2026-02-24 00:23:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task 5bf1eadf-39a3-4b5b-9eb1-63c073d32e17 (squid) has been started and output is visible here.\u001b[0m",
                                "",
                                "PLAY [Apply role squid] ********************************************************",
                                "",
                                "TASK [osism.services.squid : Include install tasks] ****************************",
                                "Tuesday 24 February 2026  00:23:35 +0000 (0:00:00.159)       0:00:00.159 ****** ",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/services/roles/squid/tasks/install-Debian-family.yml for testbed-manager\u001b[0m",
                                "",
                                "TASK [osism.services.squid : Install required packages] ************************",
                                "Tuesday 24 February 2026  00:23:35 +0000 (0:00:00.079)       0:00:00.238 ****** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.services.squid : Create required directories] **********************",
                                "Tuesday 24 February 2026  00:23:37 +0000 (0:00:01.349)       0:00:01.588 ****** ",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=/opt/squid/configuration)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=/opt/squid/configuration/conf.d)\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager] => (item=/opt/squid)\u001b[0m",
                                "",
                                "TASK [osism.services.squid : Copy squid configuration files] *******************",
                                "Tuesday 24 February 2026  00:23:38 +0000 (0:00:01.014)       0:00:02.602 ****** ",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=osism.conf)\u001b[0m",
                                "",
                                "TASK [osism.services.squid : Remove osism_allow_list.conf configuration file] ***",
                                "Tuesday 24 February 2026  00:23:39 +0000 (0:00:00.929)       0:00:03.532 ****** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.services.squid : Copy docker-compose.yml file] *********************",
                                "Tuesday 24 February 2026  00:23:39 +0000 (0:00:00.298)       0:00:03.830 ****** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.services.squid : Manage squid service] *****************************",
                                "Tuesday 24 February 2026  00:23:40 +0000 (0:00:00.847)       0:00:04.678 ****** ",
                                "\u001b[1;30mFAILED - RETRYING: [testbed-manager]: Manage squid service (10 retries left).\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "",
                                "RUNNING HANDLER [osism.services.squid : Restart squid service] *****************",
                                "Tuesday 24 February 2026  00:24:10 +0000 (0:00:30.698)       0:00:35.376 ****** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "RUNNING HANDLER [osism.services.squid : Wait for squid service to start] *******",
                                "Tuesday 24 February 2026  00:24:22 +0000 (0:00:11.737)       0:00:47.113 ****** ",
                                "Pausing for 60 seconds",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "RUNNING HANDLER [osism.services.squid : Register that squid service was restarted] ***",
                                "Tuesday 24 February 2026  00:25:22 +0000 (0:01:00.075)       0:01:47.189 ****** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "",
                                "RUNNING HANDLER [osism.services.squid : Wait for an healthy squid service] *****",
                                "Tuesday 24 February 2026  00:25:22 +0000 (0:00:00.067)       0:01:47.257 ****** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;33mtestbed-manager\u001b[0m            : \u001b[0;32mok=11  \u001b[0m \u001b[0;33mchanged=6   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Tuesday 24 February 2026  00:25:23 +0000 (0:00:00.608)       0:01:47.865 ****** ",
                                "=============================================================================== ",
                                "osism.services.squid : Wait for squid service to start ----------------- 60.08s",
                                "osism.services.squid : Manage squid service ---------------------------- 30.70s",
                                "osism.services.squid : Restart squid service --------------------------- 11.74s",
                                "osism.services.squid : Install required packages ------------------------ 1.35s",
                                "osism.services.squid : Create required directories ---------------------- 1.01s",
                                "osism.services.squid : Copy squid configuration files ------------------- 0.93s",
                                "osism.services.squid : Copy docker-compose.yml file --------------------- 0.85s",
                                "osism.services.squid : Wait for an healthy squid service ---------------- 0.61s",
                                "osism.services.squid : Remove osism_allow_list.conf configuration file --- 0.30s",
                                "osism.services.squid : Include install tasks ---------------------------- 0.08s",
                                "osism.services.squid : Register that squid service was restarted -------- 0.07s",
                                "+ [[ latest != \\l\\a\\t\\e\\s\\t ]]",
                                "+ /opt/configuration/scripts/set-kolla-namespace.sh kolla",
                                "+ set -e",
                                "+ NAMESPACE=kolla",
                                "+ sed -i 's#docker_namespace: .*#docker_namespace: kolla#g' /opt/configuration/inventory/group_vars/all/kolla.yml",
                                "++ semver latest 9.0.0",
                                "+ [[ -1 -lt 0 ]]",
                                "+ [[ latest != \\l\\a\\t\\e\\s\\t ]]",
                                "+ osism apply operator -u ubuntu -l testbed-nodes",
                                "\u001b[32m2026-02-24 00:25:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mPrepare task for execution of operator.\u001b[0m",
                                "\u001b[32m2026-02-24 00:25:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4ac53be9-301e-43c7-aa44-9dcc07469c96 (operator) was prepared for execution.\u001b[0m",
                                "\u001b[32m2026-02-24 00:25:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task 4ac53be9-301e-43c7-aa44-9dcc07469c96 (operator) has been started and output is visible here.\u001b[0m",
                                "",
                                "PLAY [Make ssh pipelining working] *********************************************",
                                "",
                                "TASK [Gathering Facts] *********************************************************",
                                "Tuesday 24 February 2026  00:25:39 +0000 (0:00:00.104)       0:00:00.104 ****** ",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Do not require tty for all users] ****************************************",
                                "Tuesday 24 February 2026  00:25:42 +0000 (0:00:03.366)       0:00:03.471 ****** ",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "",
                                "PLAY [Apply role operator] *****************************************************",
                                "",
                                "TASK [osism.commons.operator : Gather variables for each operating system] *****",
                                "Tuesday 24 February 2026  00:25:43 +0000 (0:00:00.751)       0:00:04.222 ****** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.commons.operator : Set operator_groups variable to default value] ***",
                                "Tuesday 24 February 2026  00:25:43 +0000 (0:00:00.155)       0:00:04.378 ****** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.commons.operator : Create operator group] **************************",
                                "Tuesday 24 February 2026  00:25:43 +0000 (0:00:00.127)       0:00:04.505 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [osism.commons.operator : Create user] ************************************",
                                "Tuesday 24 February 2026  00:25:44 +0000 (0:00:00.609)       0:00:05.114 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [osism.commons.operator : Add user to additional groups] ******************",
                                "Tuesday 24 February 2026  00:25:45 +0000 (0:00:00.801)       0:00:05.916 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=adm)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=adm)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=adm)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=adm)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=adm)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=adm)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=sudo)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=sudo)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=sudo)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=sudo)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=sudo)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=sudo)\u001b[0m",
                                "",
                                "TASK [osism.commons.operator : Copy user sudoers file] *************************",
                                "Tuesday 24 February 2026  00:25:46 +0000 (0:00:01.185)       0:00:07.101 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [osism.commons.operator : Set language variables in .bashrc configuration file] ***",
                                "Tuesday 24 February 2026  00:25:47 +0000 (0:00:01.219)       0:00:08.320 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=export LANGUAGE=C.UTF-8)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=export LANGUAGE=C.UTF-8)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=export LANGUAGE=C.UTF-8)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=export LANGUAGE=C.UTF-8)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=export LANGUAGE=C.UTF-8)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=export LANGUAGE=C.UTF-8)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=export LANG=C.UTF-8)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=export LANG=C.UTF-8)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=export LANG=C.UTF-8)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=export LANG=C.UTF-8)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=export LANG=C.UTF-8)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=export LANG=C.UTF-8)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=export LC_ALL=C.UTF-8)\u001b[0m",
                                "\u001b[1;35m[WARNING]: Module remote_tmp /root/.ansible/tmp did not exist and was created\u001b[0m",
                                "\u001b[1;35mwith a mode of 0700, this may cause issues when running as another user. To\u001b[0m",
                                "\u001b[1;35mavoid this, create the remote_tmp dir with the correct permissions manually\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=export LC_ALL=C.UTF-8)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=export LC_ALL=C.UTF-8)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=export LC_ALL=C.UTF-8)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=export LC_ALL=C.UTF-8)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=export LC_ALL=C.UTF-8)\u001b[0m",
                                "",
                                "TASK [osism.commons.operator : Set custom environment variables in .bashrc configuration file] ***",
                                "Tuesday 24 February 2026  00:25:48 +0000 (0:00:01.222)       0:00:09.543 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.commons.operator : Set custom PS1 prompt in .bashrc configuration file] ***",
                                "Tuesday 24 February 2026  00:25:49 +0000 (0:00:00.147)       0:00:09.690 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.commons.operator : Create .ssh directory] **************************",
                                "Tuesday 24 February 2026  00:25:49 +0000 (0:00:00.177)       0:00:09.868 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [osism.commons.operator : Check number of SSH authorized keys] ************",
                                "Tuesday 24 February 2026  00:25:49 +0000 (0:00:00.560)       0:00:10.428 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.commons.operator : Set ssh authorized keys] ************************",
                                "Tuesday 24 February 2026  00:25:50 +0000 (0:00:00.154)       0:00:10.583 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=None)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=None)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=None)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=None)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=None)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=None)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [osism.commons.operator : Delete ssh authorized keys] *********************",
                                "Tuesday 24 February 2026  00:25:50 +0000 (0:00:00.733)       0:00:11.317 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.commons.operator : Set authorized GitHub accounts] *****************",
                                "Tuesday 24 February 2026  00:25:50 +0000 (0:00:00.142)       0:00:11.459 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.commons.operator : Delete authorized GitHub accounts] **************",
                                "Tuesday 24 February 2026  00:25:51 +0000 (0:00:00.134)       0:00:11.594 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.commons.operator : Set password] ***********************************",
                                "Tuesday 24 February 2026  00:25:51 +0000 (0:00:00.137)       0:00:11.732 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.commons.operator : Unset & lock password] **************************",
                                "Tuesday 24 February 2026  00:25:51 +0000 (0:00:00.650)       0:00:12.382 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=12  \u001b[0m \u001b[0;33mchanged=8   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=7   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=12  \u001b[0m \u001b[0;33mchanged=8   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=7   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=12  \u001b[0m \u001b[0;33mchanged=8   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=7   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-3\u001b[0m             : \u001b[0;32mok=12  \u001b[0m \u001b[0;33mchanged=8   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=7   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-4\u001b[0m             : \u001b[0;32mok=12  \u001b[0m \u001b[0;33mchanged=8   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=7   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-5\u001b[0m             : \u001b[0;32mok=12  \u001b[0m \u001b[0;33mchanged=8   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=7   \u001b[0m rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Tuesday 24 February 2026  00:25:52 +0000 (0:00:00.205)       0:00:12.588 ****** ",
                                "=============================================================================== ",
                                "Gathering Facts --------------------------------------------------------- 3.37s",
                                "osism.commons.operator : Set language variables in .bashrc configuration file --- 1.22s",
                                "osism.commons.operator : Copy user sudoers file ------------------------- 1.22s",
                                "osism.commons.operator : Add user to additional groups ------------------ 1.19s",
                                "osism.commons.operator : Create user ------------------------------------ 0.80s",
                                "Do not require tty for all users ---------------------------------------- 0.75s",
                                "osism.commons.operator : Set ssh authorized keys ------------------------ 0.73s",
                                "osism.commons.operator : Set password ----------------------------------- 0.65s",
                                "osism.commons.operator : Create operator group -------------------------- 0.61s",
                                "osism.commons.operator : Create .ssh directory -------------------------- 0.56s",
                                "osism.commons.operator : Unset & lock password -------------------------- 0.21s",
                                "osism.commons.operator : Set custom PS1 prompt in .bashrc configuration file --- 0.18s",
                                "osism.commons.operator : Gather variables for each operating system ----- 0.16s",
                                "osism.commons.operator : Check number of SSH authorized keys ------------ 0.15s",
                                "osism.commons.operator : Set custom environment variables in .bashrc configuration file --- 0.15s",
                                "osism.commons.operator : Delete ssh authorized keys --------------------- 0.14s",
                                "osism.commons.operator : Delete authorized GitHub accounts -------------- 0.14s",
                                "osism.commons.operator : Set authorized GitHub accounts ----------------- 0.13s",
                                "osism.commons.operator : Set operator_groups variable to default value --- 0.13s",
                                "+ osism apply --environment custom facts",
                                "\u001b[32m2026-02-24 00:25:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTrying to run play facts in environment custom\u001b[0m",
                                "\u001b[32m2026-02-24 00:26:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mPrepare task for execution of facts.\u001b[0m",
                                "\u001b[32m2026-02-24 00:26:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b9478762-dd77-4d27-951c-36d6391ab37d (facts) was prepared for execution.\u001b[0m",
                                "\u001b[32m2026-02-24 00:26:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task b9478762-dd77-4d27-951c-36d6391ab37d (facts) has been started and output is visible here.\u001b[0m",
                                "",
                                "PLAY [Copy custom network devices fact] ****************************************",
                                "",
                                "TASK [Create custom facts directory] *******************************************",
                                "Tuesday 24 February 2026  00:26:08 +0000 (0:00:00.068)       0:00:00.068 ****** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [Copy fact file] **********************************************************",
                                "Tuesday 24 February 2026  00:26:09 +0000 (0:00:01.405)       0:00:01.474 ****** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "",
                                "PLAY [Copy custom ceph devices facts] ******************************************",
                                "",
                                "TASK [osism.commons.repository : Gather variables for each operating system] ***",
                                "Tuesday 24 February 2026  00:26:11 +0000 (0:00:01.219)       0:00:02.694 ****** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.commons.repository : Set repository_default fact to default value] ***",
                                "Tuesday 24 February 2026  00:26:11 +0000 (0:00:00.088)       0:00:02.782 ****** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.commons.repository : Set repositories to default] ******************",
                                "Tuesday 24 February 2026  00:26:11 +0000 (0:00:00.191)       0:00:02.974 ****** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.commons.repository : Include distribution specific repository tasks] ***",
                                "Tuesday 24 February 2026  00:26:11 +0000 (0:00:00.216)       0:00:03.191 ****** ",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/repository/tasks/Ubuntu.yml for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m",
                                "",
                                "TASK [osism.commons.repository : Create /etc/apt/sources.list.d directory] *****",
                                "Tuesday 24 February 2026  00:26:11 +0000 (0:00:00.137)       0:00:03.328 ****** ",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.commons.repository : Include tasks for Ubuntu < 24.04] *************",
                                "Tuesday 24 February 2026  00:26:12 +0000 (0:00:00.432)       0:00:03.761 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.commons.repository : Copy 99osism apt configuration] ***************",
                                "Tuesday 24 February 2026  00:26:12 +0000 (0:00:00.152)       0:00:03.914 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [osism.commons.repository : Remove sources.list file] *********************",
                                "Tuesday 24 February 2026  00:26:13 +0000 (0:00:01.077)       0:00:04.991 ****** ",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.commons.repository : Copy ubuntu.sources file] *********************",
                                "Tuesday 24 February 2026  00:26:13 +0000 (0:00:00.473)       0:00:05.465 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.commons.repository : Update package cache] *************************",
                                "Tuesday 24 February 2026  00:26:15 +0000 (0:00:01.062)       0:00:06.528 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Install required packages (RedHat)] **************************************",
                                "Tuesday 24 February 2026  00:26:32 +0000 (0:00:17.185)       0:00:23.713 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Install required packages (Debian)] **************************************",
                                "Tuesday 24 February 2026  00:26:32 +0000 (0:00:00.095)       0:00:23.809 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Create custom facts directory] *******************************************",
                                "Tuesday 24 February 2026  00:26:40 +0000 (0:00:08.327)       0:00:32.136 ****** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Copy fact files] *********************************************************",
                                "Tuesday 24 February 2026  00:26:41 +0000 (0:00:00.466)       0:00:32.602 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=testbed_ceph_devices)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=testbed_ceph_devices)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=testbed_ceph_devices)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=testbed_ceph_devices_all)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=testbed_ceph_devices_all)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=testbed_ceph_devices_all)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=testbed_ceph_osd_devices)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=testbed_ceph_osd_devices)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=testbed_ceph_osd_devices)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=testbed_ceph_osd_devices_all)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=testbed_ceph_osd_devices_all)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=testbed_ceph_osd_devices_all)\u001b[0m",
                                "",
                                "RUNNING HANDLER [osism.commons.repository : Force update of package cache] *****",
                                "Tuesday 24 February 2026  00:26:44 +0000 (0:00:03.556)       0:00:36.159 ****** ",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "",
                                "PLAY [Gather facts for all hosts] **********************************************",
                                "",
                                "TASK [Gathers facts about hosts] ***********************************************",
                                "Tuesday 24 February 2026  00:26:45 +0000 (0:00:01.345)       0:00:37.505 ****** ",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;32mtestbed-manager\u001b[0m            : \u001b[0;32mok=3   \u001b[0m changed=0    unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=3   \u001b[0m \u001b[0;33mchanged=2   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=3   \u001b[0m \u001b[0;33mchanged=2   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=3   \u001b[0m \u001b[0;33mchanged=2   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-3\u001b[0m             : \u001b[0;32mok=16  \u001b[0m \u001b[0;33mchanged=7   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-4\u001b[0m             : \u001b[0;32mok=16  \u001b[0m \u001b[0;33mchanged=7   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-5\u001b[0m             : \u001b[0;32mok=16  \u001b[0m \u001b[0;33mchanged=7   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Tuesday 24 February 2026  00:26:49 +0000 (0:00:03.769)       0:00:41.275 ****** ",
                                "=============================================================================== ",
                                "osism.commons.repository : Update package cache ------------------------ 17.19s",
                                "Install required packages (Debian) -------------------------------------- 8.33s",
                                "Gathers facts about hosts ----------------------------------------------- 3.77s",
                                "Copy fact files --------------------------------------------------------- 3.56s",
                                "Create custom facts directory ------------------------------------------- 1.41s",
                                "osism.commons.repository : Force update of package cache ---------------- 1.35s",
                                "Copy fact file ---------------------------------------------------------- 1.22s",
                                "osism.commons.repository : Copy 99osism apt configuration --------------- 1.08s",
                                "osism.commons.repository : Copy ubuntu.sources file --------------------- 1.06s",
                                "osism.commons.repository : Remove sources.list file --------------------- 0.47s",
                                "Create custom facts directory ------------------------------------------- 0.47s",
                                "osism.commons.repository : Create /etc/apt/sources.list.d directory ----- 0.43s",
                                "osism.commons.repository : Set repositories to default ------------------ 0.22s",
                                "osism.commons.repository : Set repository_default fact to default value --- 0.19s",
                                "osism.commons.repository : Include tasks for Ubuntu < 24.04 ------------- 0.15s",
                                "osism.commons.repository : Include distribution specific repository tasks --- 0.14s",
                                "Install required packages (RedHat) -------------------------------------- 0.10s",
                                "osism.commons.repository : Gather variables for each operating system --- 0.09s",
                                "+ osism apply bootstrap",
                                "\u001b[32m2026-02-24 00:27:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mPrepare task for execution of bootstrap.\u001b[0m",
                                "\u001b[32m2026-02-24 00:27:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 2991ae07-7a8f-4705-9073-594027f6153c (bootstrap) was prepared for execution.\u001b[0m",
                                "\u001b[32m2026-02-24 00:27:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task 2991ae07-7a8f-4705-9073-594027f6153c (bootstrap) has been started and output is visible here.\u001b[0m",
                                "",
                                "PLAY [Group hosts based on state bootstrap] ************************************",
                                "",
                                "TASK [Group hosts based on state bootstrap] ************************************",
                                "Tuesday 24 February 2026  00:27:06 +0000 (0:00:00.142)       0:00:00.142 ****** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "PLAY [Gather facts for all hosts] **********************************************",
                                "",
                                "TASK [Gathers facts about hosts] ***********************************************",
                                "Tuesday 24 February 2026  00:27:06 +0000 (0:00:00.249)       0:00:00.392 ****** ",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "",
                                "PLAY [Gather facts for all hosts (if using --limit)] ***************************",
                                "",
                                "TASK [Gathers facts about hosts] ***********************************************",
                                "Tuesday 24 February 2026  00:27:10 +0000 (0:00:03.716)       0:00:04.109 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-3) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-4) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=testbed-node-3) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-5) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=testbed-node-4) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=testbed-node-3) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-manager) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=testbed-node-5) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-0) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=testbed-node-4) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=testbed-manager) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=testbed-node-5) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-1) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=testbed-node-0) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager] => (item=testbed-node-3) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=testbed-node-1) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=testbed-manager) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=testbed-node-2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=testbed-node-0) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager] => (item=testbed-node-4) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-3) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=testbed-node-1) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-4) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager] => (item=testbed-node-5) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=testbed-node-2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-5) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=testbed-node-3) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-manager) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager] => (item=testbed-manager) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=testbed-node-4) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-0) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager] => (item=testbed-node-0) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=testbed-node-3) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-1) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=testbed-node-5) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager] => (item=testbed-node-1) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=testbed-node-4) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager] => (item=testbed-node-2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=testbed-node-5) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=testbed-manager) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=testbed-node-0) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=testbed-manager) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=testbed-node-1) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=testbed-node-0) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=testbed-node-2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=testbed-node-1) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=testbed-node-2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "PLAY [Apply bootstrap roles part 1] ********************************************",
                                "",
                                "TASK [osism.commons.hostname : Set hostname] ***********************************",
                                "Tuesday 24 February 2026  00:27:11 +0000 (0:00:00.473)       0:00:04.583 ****** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [osism.commons.hostname : Copy /etc/hostname] *****************************",
                                "Tuesday 24 February 2026  00:27:12 +0000 (0:00:01.188)       0:00:05.771 ****** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [osism.commons.hosts : Include type specific tasks] ***********************",
                                "Tuesday 24 February 2026  00:27:13 +0000 (0:00:01.228)       0:00:07.000 ****** ",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/hosts/tasks/type-template.yml for testbed-node-3, testbed-node-4, testbed-node-5, testbed-manager, testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [osism.commons.hosts : Copy /etc/hosts file] ******************************",
                                "Tuesday 24 February 2026  00:27:13 +0000 (0:00:00.244)       0:00:07.244 ****** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [osism.commons.proxy : Include distribution specific tasks] ***************",
                                "Tuesday 24 February 2026  00:27:15 +0000 (0:00:01.931)       0:00:09.176 ****** ",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/proxy/tasks/Debian-family.yml for testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [osism.commons.proxy : Configure proxy parameters for apt] ****************",
                                "Tuesday 24 February 2026  00:27:16 +0000 (0:00:00.315)       0:00:09.491 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [osism.commons.proxy : Set system wide settings in environment file] ******",
                                "Tuesday 24 February 2026  00:27:17 +0000 (0:00:01.104)       0:00:10.596 ****** ",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [osism.commons.proxy : Remove system wide settings in environment file] ***",
                                "Tuesday 24 February 2026  00:27:17 +0000 (0:00:00.553)       0:00:11.149 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.commons.resolvconf : Check minimum and maximum number of name servers] ***",
                                "Tuesday 24 February 2026  00:27:18 +0000 (0:00:00.489)       0:00:11.639 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [osism.commons.resolvconf : Include resolvconf tasks] *********************",
                                "Tuesday 24 February 2026  00:27:18 +0000 (0:00:00.226)       0:00:11.866 ****** ",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/resolvconf/tasks/configure-resolv.yml for testbed-node-3, testbed-node-4, testbed-node-5, testbed-manager, testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [osism.commons.resolvconf : Include distribution specific installation tasks] ***",
                                "Tuesday 24 February 2026  00:27:18 +0000 (0:00:00.288)       0:00:12.154 ****** ",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/resolvconf/tasks/install-Debian-family.yml for testbed-node-3, testbed-node-4, testbed-node-5, testbed-manager, testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [osism.commons.resolvconf : Remove packages configuring /etc/resolv.conf] ***",
                                "Tuesday 24 February 2026  00:27:19 +0000 (0:00:00.440)       0:00:12.595 ****** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [osism.commons.resolvconf : Install package systemd-resolved] *************",
                                "Tuesday 24 February 2026  00:27:20 +0000 (0:00:01.431)       0:00:14.027 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [osism.commons.resolvconf : Retrieve file status of /etc/resolv.conf] *****",
                                "Tuesday 24 February 2026  00:27:20 +0000 (0:00:00.239)       0:00:14.267 ****** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [osism.commons.resolvconf : Archive existing file /etc/resolv.conf] *******",
                                "Tuesday 24 February 2026  00:27:21 +0000 (0:00:00.535)       0:00:14.802 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [osism.commons.resolvconf : Link /run/systemd/resolve/stub-resolv.conf to /etc/resolv.conf] ***",
                                "Tuesday 24 February 2026  00:27:21 +0000 (0:00:00.219)       0:00:15.022 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [osism.commons.resolvconf : Copy configuration files] *********************",
                                "Tuesday 24 February 2026  00:27:22 +0000 (0:00:00.654)       0:00:15.676 ****** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [osism.commons.resolvconf : Start/enable systemd-resolved service] ********",
                                "Tuesday 24 February 2026  00:27:23 +0000 (0:00:01.139)       0:00:16.815 ****** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [osism.commons.resolvconf : Include distribution specific configuration tasks] ***",
                                "Tuesday 24 February 2026  00:27:24 +0000 (0:00:01.032)       0:00:17.848 ****** ",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/resolvconf/tasks/configure-Debian-family.yml for testbed-node-3, testbed-node-4, testbed-node-5, testbed-manager, testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [osism.commons.resolvconf : Restart systemd-resolved service] *************",
                                "Tuesday 24 February 2026  00:27:24 +0000 (0:00:00.347)       0:00:18.195 ****** ",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [osism.commons.repository : Gather variables for each operating system] ***",
                                "Tuesday 24 February 2026  00:27:26 +0000 (0:00:01.231)       0:00:19.427 ****** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [osism.commons.repository : Set repository_default fact to default value] ***",
                                "Tuesday 24 February 2026  00:27:26 +0000 (0:00:00.261)       0:00:19.689 ****** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [osism.commons.repository : Set repositories to default] ******************",
                                "Tuesday 24 February 2026  00:27:26 +0000 (0:00:00.245)       0:00:19.934 ****** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [osism.commons.repository : Include distribution specific repository tasks] ***",
                                "Tuesday 24 February 2026  00:27:26 +0000 (0:00:00.229)       0:00:20.164 ****** ",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/repository/tasks/Ubuntu.yml for testbed-node-3, testbed-node-4, testbed-node-5, testbed-manager, testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [osism.commons.repository : Create /etc/apt/sources.list.d directory] *****",
                                "Tuesday 24 February 2026  00:27:27 +0000 (0:00:00.314)       0:00:20.478 ****** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [osism.commons.repository : Include tasks for Ubuntu < 24.04] *************",
                                "Tuesday 24 February 2026  00:27:27 +0000 (0:00:00.528)       0:00:21.007 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [osism.commons.repository : Copy 99osism apt configuration] ***************",
                                "Tuesday 24 February 2026  00:27:27 +0000 (0:00:00.235)       0:00:21.242 ****** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [osism.commons.repository : Remove sources.list file] *********************",
                                "Tuesday 24 February 2026  00:27:29 +0000 (0:00:01.189)       0:00:22.431 ****** ",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [osism.commons.repository : Copy ubuntu.sources file] *********************",
                                "Tuesday 24 February 2026  00:27:29 +0000 (0:00:00.569)       0:00:23.000 ****** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [osism.commons.repository : Update package cache] *************************",
                                "Tuesday 24 February 2026  00:27:30 +0000 (0:00:01.184)       0:00:24.184 ****** ",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [osism.services.rsyslog : Gather variables for each operating system] *****",
                                "Tuesday 24 February 2026  00:27:47 +0000 (0:00:17.061)       0:00:41.246 ****** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [osism.services.rsyslog : Set rsyslog_user variable to default value] *****",
                                "Tuesday 24 February 2026  00:27:48 +0000 (0:00:00.236)       0:00:41.483 ****** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [osism.services.rsyslog : Set rsyslog_workdir variable to default value] ***",
                                "Tuesday 24 February 2026  00:27:48 +0000 (0:00:00.233)       0:00:41.716 ****** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [osism.services.rsyslog : Include distribution specific install tasks] ****",
                                "Tuesday 24 February 2026  00:27:48 +0000 (0:00:00.210)       0:00:41.927 ****** ",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/services/roles/rsyslog/tasks/install-Debian-family.yml for testbed-node-3, testbed-node-4, testbed-node-5, testbed-manager, testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [osism.services.rsyslog : Install rsyslog package] ************************",
                                "Tuesday 24 February 2026  00:27:48 +0000 (0:00:00.275)       0:00:42.202 ****** ",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [osism.services.rsyslog : Copy rsyslog.conf configuration file] ***********",
                                "Tuesday 24 February 2026  00:27:50 +0000 (0:00:01.470)       0:00:43.673 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [osism.services.rsyslog : Manage rsyslog service] *************************",
                                "Tuesday 24 February 2026  00:27:51 +0000 (0:00:01.084)       0:00:44.757 ****** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [osism.services.rsyslog : Include fluentd tasks] **************************",
                                "Tuesday 24 February 2026  00:27:52 +0000 (0:00:00.805)       0:00:45.563 ****** ",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/services/roles/rsyslog/tasks/fluentd.yml for testbed-node-3, testbed-node-4, testbed-node-5, testbed-manager, testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [osism.services.rsyslog : Forward syslog message to local fluentd daemon] ***",
                                "Tuesday 24 February 2026  00:27:52 +0000 (0:00:00.302)       0:00:45.866 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [osism.services.rsyslog : Include additional log server tasks] ************",
                                "Tuesday 24 February 2026  00:27:53 +0000 (0:00:00.937)       0:00:46.803 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [osism.services.rsyslog : Include logrotate tasks] ************************",
                                "Tuesday 24 February 2026  00:27:53 +0000 (0:00:00.251)       0:00:47.054 ****** ",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/services/roles/rsyslog/tasks/logrotate.yml for testbed-node-3, testbed-node-4, testbed-node-5, testbed-manager, testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [osism.services.rsyslog : Ensure logrotate package is installed] **********",
                                "Tuesday 24 February 2026  00:27:53 +0000 (0:00:00.332)       0:00:47.387 ****** ",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [osism.services.rsyslog : Configure logrotate for rsyslog] ****************",
                                "Tuesday 24 February 2026  00:27:55 +0000 (0:00:01.760)       0:00:49.147 ****** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [osism.commons.systohc : Install util-linux-extra package] ****************",
                                "Tuesday 24 February 2026  00:27:56 +0000 (0:00:01.235)       0:00:50.383 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.commons.systohc : Sync hardware clock] *****************************",
                                "Tuesday 24 February 2026  00:28:08 +0000 (0:00:11.127)       0:01:01.511 ****** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [osism.commons.configfs : Start sys-kernel-config mount] ******************",
                                "Tuesday 24 February 2026  00:28:09 +0000 (0:00:01.552)       0:01:03.064 ****** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [osism.commons.packages : Gather variables for each operating system] *****",
                                "Tuesday 24 February 2026  00:28:10 +0000 (0:00:00.798)       0:01:03.862 ****** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [osism.commons.packages : Set required_packages_distribution variable to default value] ***",
                                "Tuesday 24 February 2026  00:28:10 +0000 (0:00:00.217)       0:01:04.080 ****** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [osism.commons.packages : Include distribution specific package tasks] ****",
                                "Tuesday 24 February 2026  00:28:10 +0000 (0:00:00.210)       0:01:04.291 ****** ",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/packages/tasks/package-Debian-family.yml for testbed-node-3, testbed-node-4, testbed-node-5, testbed-manager, testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [osism.commons.packages : Install needrestart package] ********************",
                                "Tuesday 24 February 2026  00:28:11 +0000 (0:00:00.313)       0:01:04.604 ****** ",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [osism.commons.packages : Set needrestart mode] ***************************",
                                "Tuesday 24 February 2026  00:28:12 +0000 (0:00:01.389)       0:01:05.993 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [osism.commons.packages : Set apt_cache_valid_time variable to default value] ***",
                                "Tuesday 24 February 2026  00:28:13 +0000 (0:00:00.550)       0:01:06.544 ****** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [osism.commons.packages : Update package cache] ***************************",
                                "Tuesday 24 February 2026  00:28:13 +0000 (0:00:00.250)       0:01:06.794 ****** ",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [osism.commons.packages : Download upgrade packages] **********************",
                                "Tuesday 24 February 2026  00:28:14 +0000 (0:00:01.016)       0:01:07.810 ****** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [osism.commons.packages : Upgrade packages] *******************************",
                                "Tuesday 24 February 2026  00:28:16 +0000 (0:00:01.625)       0:01:09.436 ****** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [osism.commons.packages : Download required packages] *********************",
                                "Tuesday 24 February 2026  00:28:18 +0000 (0:00:02.763)       0:01:12.199 ****** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.commons.packages : Install required packages] **********************",
                                "Tuesday 24 February 2026  00:29:06 +0000 (0:00:48.155)       0:02:00.355 ****** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [osism.commons.packages : Remove useless packages from the cache] *********",
                                "Tuesday 24 February 2026  00:30:32 +0000 (0:01:25.767)       0:03:26.122 ****** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [osism.commons.packages : Remove dependencies that are no longer required] ***",
                                "Tuesday 24 February 2026  00:30:34 +0000 (0:00:02.126)       0:03:28.248 ****** ",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.commons.sysctl : Include sysctl tasks] *****************************",
                                "Tuesday 24 February 2026  00:30:45 +0000 (0:00:11.002)       0:03:39.251 ****** ",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/sysctl/tasks/sysctl.yml for testbed-node-3, testbed-node-4, testbed-node-5, testbed-manager, testbed-node-0, testbed-node-1, testbed-node-2 => (item={'key': 'elasticsearch', 'value': [{'name': 'vm.max_map_count', 'value': 262144}]})\u001b[0m",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/sysctl/tasks/sysctl.yml for testbed-node-3, testbed-node-4, testbed-node-5, testbed-manager, testbed-node-0, testbed-node-1, testbed-node-2 => (item={'key': 'rabbitmq', 'value': [{'name': 'net.ipv4.tcp_keepalive_time', 'value': 6}, {'name': 'net.ipv4.tcp_keepalive_intvl', 'value': 3}, {'name': 'net.ipv4.tcp_keepalive_probes', 'value': 3}, {'name': 'net.core.wmem_max', 'value': 16777216}, {'name': 'net.core.rmem_max', 'value': 16777216}, {'name': 'net.ipv4.tcp_fin_timeout', 'value': 20}, {'name': 'net.ipv4.tcp_tw_reuse', 'value': 1}, {'name': 'net.core.somaxconn', 'value': 4096}, {'name': 'net.ipv4.tcp_syncookies', 'value': 0}, {'name': 'net.ipv4.tcp_max_syn_backlog', 'value': 8192}]})\u001b[0m",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/sysctl/tasks/sysctl.yml for testbed-node-3, testbed-node-4, testbed-node-5, testbed-manager, testbed-node-0, testbed-node-1, testbed-node-2 => (item={'key': 'generic', 'value': [{'name': 'vm.swappiness', 'value': 1}]})\u001b[0m",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/sysctl/tasks/sysctl.yml for testbed-node-3, testbed-node-4, testbed-node-5, testbed-manager, testbed-node-0, testbed-node-1, testbed-node-2 => (item={'key': 'compute', 'value': [{'name': 'net.netfilter.nf_conntrack_max', 'value': 1048576}]})\u001b[0m",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/sysctl/tasks/sysctl.yml for testbed-node-3, testbed-node-4, testbed-node-5, testbed-manager, testbed-node-0, testbed-node-1, testbed-node-2 => (item={'key': 'network', 'value': [{'name': 'net.netfilter.nf_conntrack_max', 'value': 1048576}]})\u001b[0m",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/sysctl/tasks/sysctl.yml for testbed-node-3, testbed-node-4, testbed-node-5, testbed-manager, testbed-node-0, testbed-node-1, testbed-node-2 => (item={'key': 'k3s_node', 'value': [{'name': 'fs.inotify.max_user_instances', 'value': 1024}]})\u001b[0m",
                                "",
                                "TASK [osism.commons.sysctl : Set sysctl parameters on elasticsearch] ***********",
                                "Tuesday 24 February 2026  00:30:46 +0000 (0:00:00.381)       0:03:39.632 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'name': 'vm.max_map_count', 'value': 262144}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'name': 'vm.max_map_count', 'value': 262144}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'name': 'vm.max_map_count', 'value': 262144}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager] => (item={'name': 'vm.max_map_count', 'value': 262144}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'name': 'vm.max_map_count', 'value': 262144})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'name': 'vm.max_map_count', 'value': 262144})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'name': 'vm.max_map_count', 'value': 262144})\u001b[0m",
                                "",
                                "TASK [osism.commons.sysctl : Set sysctl parameters on rabbitmq] ****************",
                                "Tuesday 24 February 2026  00:30:47 +0000 (0:00:00.849)       0:03:40.482 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'name': 'net.ipv4.tcp_keepalive_time', 'value': 6}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'name': 'net.ipv4.tcp_keepalive_intvl', 'value': 3}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'name': 'net.ipv4.tcp_keepalive_probes', 'value': 3}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'name': 'net.core.wmem_max', 'value': 16777216}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'name': 'net.core.rmem_max', 'value': 16777216}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'name': 'net.ipv4.tcp_fin_timeout', 'value': 20}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'name': 'net.ipv4.tcp_tw_reuse', 'value': 1}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'name': 'net.core.somaxconn', 'value': 4096}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'name': 'net.ipv4.tcp_keepalive_time', 'value': 6}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'name': 'net.ipv4.tcp_syncookies', 'value': 0}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'name': 'net.ipv4.tcp_keepalive_intvl', 'value': 3}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'name': 'net.ipv4.tcp_max_syn_backlog', 'value': 8192}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'name': 'net.ipv4.tcp_keepalive_probes', 'value': 3}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'name': 'net.core.wmem_max', 'value': 16777216}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'name': 'net.core.rmem_max', 'value': 16777216}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'name': 'net.ipv4.tcp_fin_timeout', 'value': 20}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'name': 'net.ipv4.tcp_tw_reuse', 'value': 1}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'name': 'net.core.somaxconn', 'value': 4096}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'name': 'net.ipv4.tcp_syncookies', 'value': 0}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'name': 'net.ipv4.tcp_keepalive_time', 'value': 6}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'name': 'net.ipv4.tcp_max_syn_backlog', 'value': 8192}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'name': 'net.ipv4.tcp_keepalive_intvl', 'value': 3}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'name': 'net.ipv4.tcp_keepalive_probes', 'value': 3}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'name': 'net.core.wmem_max', 'value': 16777216}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'name': 'net.core.rmem_max', 'value': 16777216}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'name': 'net.ipv4.tcp_fin_timeout', 'value': 20}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'name': 'net.ipv4.tcp_tw_reuse', 'value': 1}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'name': 'net.core.somaxconn', 'value': 4096}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'name': 'net.ipv4.tcp_syncookies', 'value': 0}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'name': 'net.ipv4.tcp_max_syn_backlog', 'value': 8192}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager] => (item={'name': 'net.ipv4.tcp_keepalive_time', 'value': 6}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager] => (item={'name': 'net.ipv4.tcp_keepalive_intvl', 'value': 3}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager] => (item={'name': 'net.ipv4.tcp_keepalive_probes', 'value': 3}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager] => (item={'name': 'net.core.wmem_max', 'value': 16777216}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager] => (item={'name': 'net.core.rmem_max', 'value': 16777216}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager] => (item={'name': 'net.ipv4.tcp_fin_timeout', 'value': 20}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager] => (item={'name': 'net.ipv4.tcp_tw_reuse', 'value': 1}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager] => (item={'name': 'net.core.somaxconn', 'value': 4096}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager] => (item={'name': 'net.ipv4.tcp_syncookies', 'value': 0}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager] => (item={'name': 'net.ipv4.tcp_max_syn_backlog', 'value': 8192}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'name': 'net.ipv4.tcp_keepalive_time', 'value': 6})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'name': 'net.ipv4.tcp_keepalive_time', 'value': 6})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'name': 'net.ipv4.tcp_keepalive_time', 'value': 6})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'name': 'net.ipv4.tcp_keepalive_intvl', 'value': 3})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'name': 'net.ipv4.tcp_keepalive_intvl', 'value': 3})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'name': 'net.ipv4.tcp_keepalive_intvl', 'value': 3})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'name': 'net.ipv4.tcp_keepalive_probes', 'value': 3})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'name': 'net.ipv4.tcp_keepalive_probes', 'value': 3})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'name': 'net.ipv4.tcp_keepalive_probes', 'value': 3})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'name': 'net.core.wmem_max', 'value': 16777216})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'name': 'net.core.wmem_max', 'value': 16777216})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'name': 'net.core.wmem_max', 'value': 16777216})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'name': 'net.core.rmem_max', 'value': 16777216})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'name': 'net.core.rmem_max', 'value': 16777216})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'name': 'net.core.rmem_max', 'value': 16777216})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'name': 'net.ipv4.tcp_fin_timeout', 'value': 20})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'name': 'net.ipv4.tcp_fin_timeout', 'value': 20})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'name': 'net.ipv4.tcp_tw_reuse', 'value': 1})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'name': 'net.ipv4.tcp_tw_reuse', 'value': 1})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'name': 'net.core.somaxconn', 'value': 4096})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'name': 'net.core.somaxconn', 'value': 4096})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'name': 'net.ipv4.tcp_fin_timeout', 'value': 20})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'name': 'net.ipv4.tcp_syncookies', 'value': 0})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'name': 'net.ipv4.tcp_syncookies', 'value': 0})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'name': 'net.ipv4.tcp_tw_reuse', 'value': 1})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'name': 'net.ipv4.tcp_max_syn_backlog', 'value': 8192})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'name': 'net.ipv4.tcp_max_syn_backlog', 'value': 8192})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'name': 'net.core.somaxconn', 'value': 4096})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'name': 'net.ipv4.tcp_syncookies', 'value': 0})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'name': 'net.ipv4.tcp_max_syn_backlog', 'value': 8192})\u001b[0m",
                                "",
                                "TASK [osism.commons.sysctl : Set sysctl parameters on generic] *****************",
                                "Tuesday 24 February 2026  00:30:51 +0000 (0:00:04.767)       0:03:45.249 ****** ",
                                "\u001b[0;33mchanged: [testbed-manager] => (item={'name': 'vm.swappiness', 'value': 1})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'name': 'vm.swappiness', 'value': 1})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'name': 'vm.swappiness', 'value': 1})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'name': 'vm.swappiness', 'value': 1})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'name': 'vm.swappiness', 'value': 1})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'name': 'vm.swappiness', 'value': 1})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'name': 'vm.swappiness', 'value': 1})\u001b[0m",
                                "",
                                "TASK [osism.commons.sysctl : Set sysctl parameters on compute] *****************",
                                "Tuesday 24 February 2026  00:30:52 +0000 (0:00:00.563)       0:03:45.812 ****** ",
                                "\u001b[0;36mskipping: [testbed-manager] => (item={'name': 'net.netfilter.nf_conntrack_max', 'value': 1048576}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'name': 'net.netfilter.nf_conntrack_max', 'value': 1048576}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'name': 'net.netfilter.nf_conntrack_max', 'value': 1048576}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'name': 'net.netfilter.nf_conntrack_max', 'value': 1048576}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'name': 'net.netfilter.nf_conntrack_max', 'value': 1048576})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'name': 'net.netfilter.nf_conntrack_max', 'value': 1048576})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'name': 'net.netfilter.nf_conntrack_max', 'value': 1048576})\u001b[0m",
                                "",
                                "TASK [osism.commons.sysctl : Set sysctl parameters on network] *****************",
                                "Tuesday 24 February 2026  00:30:52 +0000 (0:00:00.457)       0:03:46.269 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'name': 'net.netfilter.nf_conntrack_max', 'value': 1048576}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'name': 'net.netfilter.nf_conntrack_max', 'value': 1048576}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'name': 'net.netfilter.nf_conntrack_max', 'value': 1048576}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager] => (item={'name': 'net.netfilter.nf_conntrack_max', 'value': 1048576}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'name': 'net.netfilter.nf_conntrack_max', 'value': 1048576})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'name': 'net.netfilter.nf_conntrack_max', 'value': 1048576})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'name': 'net.netfilter.nf_conntrack_max', 'value': 1048576})\u001b[0m",
                                "",
                                "TASK [osism.commons.sysctl : Set sysctl parameters on k3s_node] ****************",
                                "Tuesday 24 February 2026  00:30:53 +0000 (0:00:00.610)       0:03:46.880 ****** ",
                                "\u001b[0;36mskipping: [testbed-manager] => (item={'name': 'fs.inotify.max_user_instances', 'value': 1024}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'name': 'fs.inotify.max_user_instances', 'value': 1024}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'name': 'fs.inotify.max_user_instances', 'value': 1024}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'name': 'fs.inotify.max_user_instances', 'value': 1024}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'name': 'fs.inotify.max_user_instances', 'value': 1024})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'name': 'fs.inotify.max_user_instances', 'value': 1024})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'name': 'fs.inotify.max_user_instances', 'value': 1024})\u001b[0m",
                                "",
                                "TASK [osism.commons.limits : Include limits tasks] *****************************",
                                "Tuesday 24 February 2026  00:30:53 +0000 (0:00:00.533)       0:03:47.413 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [osism.commons.services : Populate service facts] *************************",
                                "Tuesday 24 February 2026  00:30:54 +0000 (0:00:00.270)       0:03:47.684 ****** ",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [osism.commons.services : Check services] *********************************",
                                "Tuesday 24 February 2026  00:30:59 +0000 (0:00:05.383)       0:03:53.067 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=nscd) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=nscd) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=nscd) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager] => (item=nscd) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=nscd) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=nscd) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=nscd) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [osism.commons.services : Start/enable required services] *****************",
                                "Tuesday 24 February 2026  00:30:59 +0000 (0:00:00.282)       0:03:53.350 ****** ",
                                "\u001b[0;32mok: [testbed-node-4] => (item=cron)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item=cron)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item=cron)\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager] => (item=cron)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0] => (item=cron)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item=cron)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item=cron)\u001b[0m",
                                "",
                                "TASK [osism.commons.motd : Include distribution specific configure tasks] ******",
                                "Tuesday 24 February 2026  00:31:01 +0000 (0:00:01.078)       0:03:54.429 ****** ",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/motd/tasks/configure-Debian-family.yml for testbed-node-3, testbed-node-4, testbed-node-5, testbed-manager, testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [osism.commons.motd : Remove update-motd package] *************************",
                                "Tuesday 24 February 2026  00:31:01 +0000 (0:00:00.405)       0:03:54.834 ****** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [osism.commons.motd : Check if /etc/default/motd-news exists] *************",
                                "Tuesday 24 February 2026  00:31:02 +0000 (0:00:01.437)       0:03:56.272 ****** ",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [osism.commons.motd : Disable the dynamic motd-news service] **************",
                                "Tuesday 24 February 2026  00:31:03 +0000 (0:00:00.628)       0:03:56.901 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [osism.commons.motd : Get all configuration files in /etc/pam.d] **********",
                                "Tuesday 24 February 2026  00:31:04 +0000 (0:00:00.613)       0:03:57.514 ****** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [osism.commons.motd : Remove pam_motd.so rule] ****************************",
                                "Tuesday 24 February 2026  00:31:04 +0000 (0:00:00.569)       0:03:58.084 ****** ",
                                "\u001b[0;33mchanged: [testbed-manager] => (item={'path': '/etc/pam.d/sshd', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 2133, 'inode': 567, 'dev': 2049, 'nlink': 1, 'atime': 1771891429.738088, 'mtime': 1740432309.0, 'ctime': 1743685035.2598536, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'path': '/etc/pam.d/sshd', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 2133, 'inode': 567, 'dev': 2049, 'nlink': 1, 'atime': 1771891470.7879937, 'mtime': 1740432309.0, 'ctime': 1743685035.2598536, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'path': '/etc/pam.d/sshd', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 2133, 'inode': 567, 'dev': 2049, 'nlink': 1, 'atime': 1771891463.073259, 'mtime': 1740432309.0, 'ctime': 1743685035.2598536, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'path': '/etc/pam.d/sshd', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 2133, 'inode': 567, 'dev': 2049, 'nlink': 1, 'atime': 1771891450.9726114, 'mtime': 1740432309.0, 'ctime': 1743685035.2598536, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'path': '/etc/pam.d/sshd', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 2133, 'inode': 567, 'dev': 2049, 'nlink': 1, 'atime': 1771891457.1204278, 'mtime': 1740432309.0, 'ctime': 1743685035.2598536, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'path': '/etc/pam.d/sshd', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 2133, 'inode': 567, 'dev': 2049, 'nlink': 1, 'atime': 1771891464.438678, 'mtime': 1740432309.0, 'ctime': 1743685035.2598536, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'path': '/etc/pam.d/sshd', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 2133, 'inode': 567, 'dev': 2049, 'nlink': 1, 'atime': 1771891461.3313303, 'mtime': 1740432309.0, 'ctime': 1743685035.2598536, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item={'path': '/etc/pam.d/login', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 4118, 'inode': 554, 'dev': 2049, 'nlink': 1, 'atime': 1743684808.8363404, 'mtime': 1712646062.0, 'ctime': 1743685035.2588537, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'path': '/etc/pam.d/login', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 4118, 'inode': 554, 'dev': 2049, 'nlink': 1, 'atime': 1743684808.8363404, 'mtime': 1712646062.0, 'ctime': 1743685035.2588537, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'path': '/etc/pam.d/login', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 4118, 'inode': 554, 'dev': 2049, 'nlink': 1, 'atime': 1743684808.8363404, 'mtime': 1712646062.0, 'ctime': 1743685035.2588537, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'path': '/etc/pam.d/login', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 4118, 'inode': 554, 'dev': 2049, 'nlink': 1, 'atime': 1743684808.8363404, 'mtime': 1712646062.0, 'ctime': 1743685035.2588537, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'path': '/etc/pam.d/login', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 4118, 'inode': 554, 'dev': 2049, 'nlink': 1, 'atime': 1743684808.8363404, 'mtime': 1712646062.0, 'ctime': 1743685035.2588537, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'path': '/etc/pam.d/login', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 4118, 'inode': 554, 'dev': 2049, 'nlink': 1, 'atime': 1743684808.8363404, 'mtime': 1712646062.0, 'ctime': 1743685035.2588537, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'path': '/etc/pam.d/login', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 4118, 'inode': 554, 'dev': 2049, 'nlink': 1, 'atime': 1743684808.8363404, 'mtime': 1712646062.0, 'ctime': 1743685035.2588537, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})\u001b[0m",
                                "",
                                "TASK [osism.commons.motd : Copy motd file] *************************************",
                                "Tuesday 24 February 2026  00:31:05 +0000 (0:00:00.941)       0:03:59.026 ****** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [osism.commons.motd : Copy issue file] ************************************",
                                "Tuesday 24 February 2026  00:31:06 +0000 (0:00:01.148)       0:04:00.175 ****** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [osism.commons.motd : Copy issue.net file] ********************************",
                                "Tuesday 24 February 2026  00:31:07 +0000 (0:00:01.120)       0:04:01.295 ****** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [osism.commons.motd : Configure SSH to print the motd] ********************",
                                "Tuesday 24 February 2026  00:31:09 +0000 (0:00:01.210)       0:04:02.506 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [osism.commons.motd : Configure SSH to not print the motd] ****************",
                                "Tuesday 24 February 2026  00:31:09 +0000 (0:00:00.215)       0:04:02.721 ****** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [osism.services.rng : Include distribution specific install tasks] ********",
                                "Tuesday 24 February 2026  00:31:09 +0000 (0:00:00.649)       0:04:03.370 ****** ",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/services/roles/rng/tasks/install-Debian-family.yml for testbed-node-3, testbed-node-4, testbed-node-5, testbed-manager, testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [osism.services.rng : Install rng package] ********************************",
                                "Tuesday 24 February 2026  00:31:10 +0000 (0:00:00.337)       0:04:03.707 ****** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [osism.services.rng : Remove haveged package] *****************************",
                                "Tuesday 24 February 2026  00:31:18 +0000 (0:00:08.681)       0:04:12.389 ****** ",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [osism.services.rng : Manage rng service] *********************************",
                                "Tuesday 24 February 2026  00:31:20 +0000 (0:00:01.231)       0:04:13.620 ****** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [osism.commons.cleanup : Gather variables for each operating system] ******",
                                "Tuesday 24 February 2026  00:31:21 +0000 (0:00:01.107)       0:04:14.727 ****** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [osism.commons.cleanup : Set cleanup_packages_distribution variable to default value] ***",
                                "Tuesday 24 February 2026  00:31:21 +0000 (0:00:00.284)       0:04:15.011 ****** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [osism.commons.cleanup : Set cleanup_services_distribution variable to default value] ***",
                                "Tuesday 24 February 2026  00:31:21 +0000 (0:00:00.289)       0:04:15.301 ****** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [osism.commons.cleanup : Populate service facts] **************************",
                                "Tuesday 24 February 2026  00:31:22 +0000 (0:00:00.272)       0:04:15.574 ****** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [osism.commons.cleanup : Include distribution specific timer tasks] *******",
                                "Tuesday 24 February 2026  00:31:28 +0000 (0:00:06.083)       0:04:21.657 ****** ",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/cleanup/tasks/timers-Debian-family.yml for testbed-node-3, testbed-node-4, testbed-node-5, testbed-manager, testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [osism.commons.cleanup : Disable apt-daily timers] ************************",
                                "Tuesday 24 February 2026  00:31:28 +0000 (0:00:00.406)       0:04:22.064 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=apt-daily-upgrade) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=apt-daily) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=apt-daily-upgrade) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=apt-daily) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=apt-daily-upgrade) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=apt-daily) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager] => (item=apt-daily-upgrade) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager] => (item=apt-daily) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=apt-daily-upgrade) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=apt-daily) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=apt-daily-upgrade) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=apt-daily) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=apt-daily-upgrade) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=apt-daily) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [osism.commons.cleanup : Include service tasks] ***************************",
                                "Tuesday 24 February 2026  00:31:28 +0000 (0:00:00.326)       0:04:22.390 ****** ",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/cleanup/tasks/services-Debian-family.yml for testbed-node-3, testbed-node-4, testbed-node-5, testbed-manager, testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [osism.commons.cleanup : Cleanup services] ********************************",
                                "Tuesday 24 February 2026  00:31:29 +0000 (0:00:00.391)       0:04:22.781 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=ModemManager.service) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=ModemManager.service) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=ModemManager.service) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager] => (item=ModemManager.service) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=ModemManager.service) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=ModemManager.service) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=ModemManager.service) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [osism.commons.cleanup : Include packages tasks] **************************",
                                "Tuesday 24 February 2026  00:31:29 +0000 (0:00:00.322)       0:04:23.104 ****** ",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/cleanup/tasks/packages-Debian-family.yml for testbed-node-3, testbed-node-4, testbed-node-5, testbed-manager, testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [osism.commons.cleanup : Cleanup installed packages] **********************",
                                "Tuesday 24 February 2026  00:31:30 +0000 (0:00:00.377)       0:04:23.481 ****** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [osism.commons.cleanup : Remove cloudinit package] ************************",
                                "Tuesday 24 February 2026  00:32:06 +0000 (0:00:36.193)       0:04:59.675 ****** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [osism.commons.cleanup : Uninstall unattended-upgrades package] ***********",
                                "Tuesday 24 February 2026  00:32:15 +0000 (0:00:09.248)       0:05:08.924 ****** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [osism.commons.cleanup : Remove useless packages from the cache] **********",
                                "Tuesday 24 February 2026  00:32:24 +0000 (0:00:08.711)       0:05:17.636 ****** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [osism.commons.cleanup : Remove dependencies that are no longer required] ***",
                                "Tuesday 24 February 2026  00:32:26 +0000 (0:00:02.094)       0:05:19.730 ****** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [osism.commons.cleanup : Include cloudinit tasks] *************************",
                                "Tuesday 24 February 2026  00:32:33 +0000 (0:00:06.841)       0:05:26.572 ****** ",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/cleanup/tasks/cloudinit.yml for testbed-node-3, testbed-node-4, testbed-node-5, testbed-manager, testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [osism.commons.cleanup : Remove cloud-init configuration directory] *******",
                                "Tuesday 24 February 2026  00:32:33 +0000 (0:00:00.367)       0:05:26.939 ****** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [osism.commons.timezone : Install tzdata package] *************************",
                                "Tuesday 24 February 2026  00:32:34 +0000 (0:00:00.644)       0:05:27.584 ****** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [osism.commons.timezone : Set timezone to UTC] ****************************",
                                "Tuesday 24 February 2026  00:32:36 +0000 (0:00:02.125)       0:05:29.709 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [osism.commons.timezone : Create /etc/adjtime file] ***********************",
                                "Tuesday 24 February 2026  00:32:37 +0000 (0:00:00.766)       0:05:30.476 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [osism.commons.timezone : Ensure UTC in /etc/adjtime] *********************",
                                "Tuesday 24 February 2026  00:32:37 +0000 (0:00:00.257)       0:05:30.734 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [osism.services.docker : Gather variables for each operating system] ******",
                                "Tuesday 24 February 2026  00:32:37 +0000 (0:00:00.374)       0:05:31.109 ****** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [osism.services.docker : Set docker_version variable to default value] ****",
                                "Tuesday 24 February 2026  00:32:37 +0000 (0:00:00.279)       0:05:31.388 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [osism.services.docker : Set docker_cli_version variable to default value] ***",
                                "Tuesday 24 February 2026  00:32:38 +0000 (0:00:00.293)       0:05:31.682 ****** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [osism.services.docker : Print used docker version] ***********************",
                                "Tuesday 24 February 2026  00:32:38 +0000 (0:00:00.283)       0:05:31.966 ****** ",
                                "\u001b[0;32mok: [testbed-node-3] => \u001b[0m",
                                "\u001b[0;32m  docker_version: 5:27.5.1\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => \u001b[0m",
                                "\u001b[0;32m  docker_version: 5:27.5.1\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => \u001b[0m",
                                "\u001b[0;32m  docker_version: 5:27.5.1\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager] => \u001b[0m",
                                "\u001b[0;32m  docker_version: 5:27.5.1\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0] => \u001b[0m",
                                "\u001b[0;32m  docker_version: 5:27.5.1\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => \u001b[0m",
                                "\u001b[0;32m  docker_version: 5:27.5.1\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => \u001b[0m",
                                "\u001b[0;32m  docker_version: 5:27.5.1\u001b[0m",
                                "",
                                "TASK [osism.services.docker : Print used docker cli version] *******************",
                                "Tuesday 24 February 2026  00:32:38 +0000 (0:00:00.258)       0:05:32.224 ****** ",
                                "\u001b[0;32mok: [testbed-node-3] => \u001b[0m",
                                "\u001b[0;32m  docker_cli_version: 5:27.5.1\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => \u001b[0m",
                                "\u001b[0;32m  docker_cli_version: 5:27.5.1\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => \u001b[0m",
                                "\u001b[0;32m  docker_cli_version: 5:27.5.1\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager] => \u001b[0m",
                                "\u001b[0;32m  docker_cli_version: 5:27.5.1\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0] => \u001b[0m",
                                "\u001b[0;32m  docker_cli_version: 5:27.5.1\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => \u001b[0m",
                                "\u001b[0;32m  docker_cli_version: 5:27.5.1\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => \u001b[0m",
                                "\u001b[0;32m  docker_cli_version: 5:27.5.1\u001b[0m",
                                "",
                                "TASK [osism.services.docker : Include block storage tasks] *********************",
                                "Tuesday 24 February 2026  00:32:39 +0000 (0:00:00.297)       0:05:32.521 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [osism.services.docker : Include zram storage tasks] **********************",
                                "Tuesday 24 February 2026  00:32:39 +0000 (0:00:00.257)       0:05:32.778 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [osism.services.docker : Include docker install tasks] ********************",
                                "Tuesday 24 February 2026  00:32:39 +0000 (0:00:00.410)       0:05:33.189 ****** ",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/services/roles/docker/tasks/install-docker-Debian-family.yml for testbed-node-3, testbed-node-4, testbed-node-5, testbed-manager, testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [osism.services.docker : Remove old architecture-dependent repository] ****",
                                "Tuesday 24 February 2026  00:32:40 +0000 (0:00:00.410)       0:05:33.600 ****** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [osism.services.docker : Gather package facts] ****************************",
                                "Tuesday 24 February 2026  00:32:41 +0000 (0:00:00.828)       0:05:34.428 ****** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [osism.services.docker : Check whether packages are installed that should not be installed] ***",
                                "Tuesday 24 February 2026  00:32:43 +0000 (0:00:02.913)       0:05:37.342 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=containerd) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=docker.io) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=docker-engine) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=containerd) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=docker.io) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=docker-engine) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=containerd) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=docker.io) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=docker-engine) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager] => (item=containerd) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager] => (item=docker.io) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager] => (item=docker-engine) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=containerd) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=docker.io) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=docker-engine) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=containerd) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=docker.io) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=docker-engine) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=containerd) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=docker.io) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=docker-engine) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [osism.services.docker : Install apt-transport-https package] *************",
                                "Tuesday 24 February 2026  00:32:44 +0000 (0:00:00.643)       0:05:37.986 ****** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [osism.services.docker : Add repository gpg key] **************************",
                                "Tuesday 24 February 2026  00:32:52 +0000 (0:00:07.739)       0:05:45.725 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [osism.services.docker : Add repository] **********************************",
                                "Tuesday 24 February 2026  00:32:53 +0000 (0:00:01.122)       0:05:46.848 ****** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [osism.services.docker : Update package cache] ****************************",
                                "Tuesday 24 February 2026  00:33:02 +0000 (0:00:09.260)       0:05:56.108 ****** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [osism.services.docker : Pin docker package version] **********************",
                                "Tuesday 24 February 2026  00:33:06 +0000 (0:00:03.809)       0:05:59.918 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [osism.services.docker : Pin docker-cli package version] ******************",
                                "Tuesday 24 February 2026  00:33:08 +0000 (0:00:01.532)       0:06:01.451 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [osism.services.docker : Unlock containerd package] ***********************",
                                "Tuesday 24 February 2026  00:33:09 +0000 (0:00:01.315)       0:06:02.767 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.services.docker : Install containerd package] **********************",
                                "Tuesday 24 February 2026  00:33:10 +0000 (0:00:00.794)       0:06:03.561 ****** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [osism.services.docker : Lock containerd package] *************************",
                                "Tuesday 24 February 2026  00:33:20 +0000 (0:00:10.380)       0:06:13.942 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [osism.services.docker : Install docker-cli package] **********************",
                                "Tuesday 24 February 2026  00:33:21 +0000 (0:00:00.951)       0:06:14.894 ****** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [osism.services.docker : Install docker package] **************************",
                                "Tuesday 24 February 2026  00:33:32 +0000 (0:00:10.585)       0:06:25.479 ****** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [osism.services.docker : Unblock installation of python docker packages] ***",
                                "Tuesday 24 February 2026  00:33:44 +0000 (0:00:12.126)       0:06:37.606 ****** ",
                                "\u001b[0;32mok: [testbed-node-3] => (item=python3-docker)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item=python3-docker)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item=python3-docker)\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager] => (item=python3-docker)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0] => (item=python3-docker)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item=python-docker)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item=python3-docker)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item=python-docker)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item=python3-docker)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item=python-docker)\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager] => (item=python-docker)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0] => (item=python-docker)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item=python-docker)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item=python-docker)\u001b[0m",
                                "",
                                "TASK [osism.services.docker : Install python3 docker package] ******************",
                                "Tuesday 24 February 2026  00:33:45 +0000 (0:00:01.194)       0:06:38.800 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [osism.services.docker : Install python3 docker package from Debian Sid] ***",
                                "Tuesday 24 February 2026  00:33:45 +0000 (0:00:00.527)       0:06:39.328 ****** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [osism.services.docker : Remove python docker packages (install python bindings from pip)] ***",
                                "Tuesday 24 February 2026  00:33:50 +0000 (0:00:04.196)       0:06:43.525 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [osism.services.docker : Block installation of python docker packages (install python bindings from pip)] ***",
                                "Tuesday 24 February 2026  00:33:50 +0000 (0:00:00.685)       0:06:44.211 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=python3-docker) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=python-docker) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=python3-docker) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=python-docker) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=python3-docker) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=python-docker) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager] => (item=python3-docker) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager] => (item=python-docker) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=python3-docker) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=python-docker) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=python3-docker) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=python-docker) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=python3-docker) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=python-docker) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [osism.services.docker : Install python3-pip package (install python bindings from pip)] ***",
                                "Tuesday 24 February 2026  00:33:51 +0000 (0:00:00.544)       0:06:44.755 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [osism.services.docker : Install docker packages (install python bindings from pip)] ***",
                                "Tuesday 24 February 2026  00:33:51 +0000 (0:00:00.529)       0:06:45.284 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [osism.services.docker : Install packages required by docker login] *******",
                                "Tuesday 24 February 2026  00:33:52 +0000 (0:00:00.497)       0:06:45.781 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [osism.services.docker : Ensure that some packages are not installed] *****",
                                "Tuesday 24 February 2026  00:33:52 +0000 (0:00:00.536)       0:06:46.318 ****** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [osism.services.docker : Include config tasks] ****************************",
                                "Tuesday 24 February 2026  00:33:55 +0000 (0:00:02.246)       0:06:48.565 ****** ",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/services/roles/docker/tasks/config.yml for testbed-node-3, testbed-node-4, testbed-node-5, testbed-manager, testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [osism.services.docker : Create plugins directory] ************************",
                                "Tuesday 24 February 2026  00:33:56 +0000 (0:00:00.948)       0:06:49.513 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [osism.services.docker : Create systemd overlay directory] ****************",
                                "Tuesday 24 February 2026  00:33:57 +0000 (0:00:00.978)       0:06:50.492 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [osism.services.docker : Copy systemd overlay file] ***********************",
                                "Tuesday 24 February 2026  00:33:58 +0000 (0:00:01.060)       0:06:51.553 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [osism.services.docker : Reload systemd daemon if systemd overlay file is changed] ***",
                                "Tuesday 24 February 2026  00:33:59 +0000 (0:00:01.451)       0:06:53.004 ****** ",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [osism.services.docker : Copy limits configuration file] ******************",
                                "Tuesday 24 February 2026  00:34:01 +0000 (0:00:01.481)       0:06:54.486 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [osism.services.docker : Copy daemon.json configuration file] *************",
                                "Tuesday 24 February 2026  00:34:02 +0000 (0:00:01.426)       0:06:55.913 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [osism.services.docker : Include service tasks] ***************************",
                                "Tuesday 24 February 2026  00:34:03 +0000 (0:00:01.437)       0:06:57.350 ****** ",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/services/roles/docker/tasks/service.yml for testbed-node-3, testbed-node-4, testbed-node-5, testbed-manager, testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [osism.services.docker : Reload systemd daemon] ***************************",
                                "Tuesday 24 February 2026  00:34:05 +0000 (0:00:01.132)       0:06:58.483 ****** ",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [osism.services.docker : Manage service] **********************************",
                                "Tuesday 24 February 2026  00:34:06 +0000 (0:00:01.410)       0:06:59.894 ****** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [osism.services.docker : Manage docker socket service] ********************",
                                "Tuesday 24 February 2026  00:34:07 +0000 (0:00:01.238)       0:07:01.133 ****** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [osism.services.docker : Manage containerd service] ***********************",
                                "Tuesday 24 February 2026  00:34:08 +0000 (0:00:01.188)       0:07:02.321 ****** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [osism.services.docker : Include bootstrap tasks] *************************",
                                "Tuesday 24 February 2026  00:34:10 +0000 (0:00:01.555)       0:07:03.877 ****** ",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/services/roles/docker/tasks/bootstrap.yml for testbed-node-3, testbed-node-4, testbed-node-5, testbed-manager, testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [osism.services.docker : Flush handlers] **********************************",
                                "Tuesday 24 February 2026  00:34:11 +0000 (0:00:00.950)       0:07:04.828 ****** ",
                                "",
                                "TASK [osism.services.docker : Flush handlers] **********************************",
                                "Tuesday 24 February 2026  00:34:11 +0000 (0:00:00.040)       0:07:04.868 ****** ",
                                "",
                                "TASK [osism.services.docker : Flush handlers] **********************************",
                                "Tuesday 24 February 2026  00:34:11 +0000 (0:00:00.039)       0:07:04.908 ****** ",
                                "",
                                "TASK [osism.services.docker : Flush handlers] **********************************",
                                "Tuesday 24 February 2026  00:34:11 +0000 (0:00:00.052)       0:07:04.960 ****** ",
                                "",
                                "TASK [osism.services.docker : Flush handlers] **********************************",
                                "Tuesday 24 February 2026  00:34:11 +0000 (0:00:00.040)       0:07:05.001 ****** ",
                                "",
                                "TASK [osism.services.docker : Flush handlers] **********************************",
                                "Tuesday 24 February 2026  00:34:11 +0000 (0:00:00.038)       0:07:05.040 ****** ",
                                "",
                                "TASK [osism.services.docker : Flush handlers] **********************************",
                                "Tuesday 24 February 2026  00:34:11 +0000 (0:00:00.049)       0:07:05.089 ****** ",
                                "",
                                "RUNNING HANDLER [osism.commons.repository : Force update of package cache] *****",
                                "Tuesday 24 February 2026  00:34:11 +0000 (0:00:00.039)       0:07:05.129 ****** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "",
                                "RUNNING HANDLER [osism.services.rsyslog : Restart rsyslog service] *************",
                                "Tuesday 24 February 2026  00:34:13 +0000 (0:00:01.325)       0:07:06.454 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "",
                                "RUNNING HANDLER [osism.services.rsyslog : Restart logrotate service] ***********",
                                "Tuesday 24 February 2026  00:34:14 +0000 (0:00:01.566)       0:07:08.021 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [osism.services.docker : Restart docker service] ***************",
                                "Tuesday 24 February 2026  00:34:15 +0000 (0:00:01.182)       0:07:09.203 ****** ",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "",
                                "RUNNING HANDLER [osism.services.docker : Wait after docker service restart] ****",
                                "Tuesday 24 February 2026  00:34:18 +0000 (0:00:02.420)       0:07:11.624 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [osism.services.docker : Add user to docker group] ************************",
                                "Tuesday 24 February 2026  00:34:18 +0000 (0:00:00.095)       0:07:11.720 ****** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [osism.services.docker : Log into private registry and force re-authorization] ***",
                                "Tuesday 24 February 2026  00:34:19 +0000 (0:00:00.994)       0:07:12.715 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [osism.services.docker : Include facts tasks] *****************************",
                                "Tuesday 24 February 2026  00:34:20 +0000 (0:00:00.809)       0:07:13.524 ****** ",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/services/roles/docker/tasks/facts.yml for testbed-node-3, testbed-node-4, testbed-node-5, testbed-manager, testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [osism.services.docker : Create facts directory] **************************",
                                "Tuesday 24 February 2026  00:34:20 +0000 (0:00:00.858)       0:07:14.383 ****** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [osism.services.docker : Copy docker fact files] **************************",
                                "Tuesday 24 February 2026  00:34:21 +0000 (0:00:00.804)       0:07:15.188 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=docker_containers)\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager] => (item=docker_containers)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=docker_containers)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=docker_containers)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=docker_containers)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=docker_containers)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=docker_containers)\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager] => (item=docker_images)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=docker_images)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=docker_images)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=docker_images)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=docker_images)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=docker_images)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=docker_images)\u001b[0m",
                                "",
                                "TASK [osism.commons.docker_compose : This install type is not supported] *******",
                                "Tuesday 24 February 2026  00:34:24 +0000 (0:00:02.694)       0:07:17.882 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [osism.commons.docker_compose : Include distribution specific install tasks] ***",
                                "Tuesday 24 February 2026  00:34:24 +0000 (0:00:00.514)       0:07:18.397 ****** ",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/docker_compose/tasks/install-Debian-family.yml for testbed-node-3, testbed-node-4, testbed-node-5, testbed-manager, testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [osism.commons.docker_compose : Remove docker-compose apt preferences file] ***",
                                "Tuesday 24 February 2026  00:34:25 +0000 (0:00:00.822)       0:07:19.219 ****** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [osism.commons.docker_compose : Get checksum of docker-compose file] ******",
                                "Tuesday 24 February 2026  00:34:26 +0000 (0:00:00.888)       0:07:20.108 ****** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [osism.commons.docker_compose : Remove docker-compose binary] *************",
                                "Tuesday 24 February 2026  00:34:27 +0000 (0:00:01.100)       0:07:21.208 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [osism.commons.docker_compose : Uninstall docker-compose package] *********",
                                "Tuesday 24 February 2026  00:34:28 +0000 (0:00:00.524)       0:07:21.733 ****** ",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [osism.commons.docker_compose : Copy docker-compose script] ***************",
                                "Tuesday 24 February 2026  00:34:29 +0000 (0:00:01.522)       0:07:23.256 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [osism.commons.docker_compose : Install docker-compose-plugin package] ****",
                                "Tuesday 24 February 2026  00:34:30 +0000 (0:00:00.510)       0:07:23.766 ****** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [osism.commons.docker_compose : Copy osism.target systemd file] ***********",
                                "Tuesday 24 February 2026  00:34:38 +0000 (0:00:08.545)       0:07:32.312 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [osism.commons.docker_compose : Enable osism.target] **********************",
                                "Tuesday 24 February 2026  00:34:40 +0000 (0:00:01.374)       0:07:33.686 ****** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [osism.commons.docker_compose : Copy docker-compose systemd unit file] ****",
                                "Tuesday 24 February 2026  00:34:41 +0000 (0:00:01.718)       0:07:35.405 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [osism.commons.facts : Create custom facts directory] *********************",
                                "Tuesday 24 February 2026  00:34:43 +0000 (0:00:01.696)       0:07:37.101 ****** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [osism.commons.facts : Copy fact files] ***********************************",
                                "Tuesday 24 February 2026  00:34:44 +0000 (0:00:01.089)       0:07:38.191 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [osism.services.chrony : Check minimum and maximum number of servers] *****",
                                "Tuesday 24 February 2026  00:34:45 +0000 (0:00:00.817)       0:07:39.008 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [osism.services.chrony : Gather variables for each operating system] ******",
                                "Tuesday 24 February 2026  00:34:46 +0000 (0:00:00.514)       0:07:39.523 ****** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [osism.services.chrony : Set chrony_conf_file variable to default value] ***",
                                "Tuesday 24 February 2026  00:34:46 +0000 (0:00:00.557)       0:07:40.080 ****** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [osism.services.chrony : Set chrony_key_file variable to default value] ***",
                                "Tuesday 24 February 2026  00:34:47 +0000 (0:00:00.716)       0:07:40.796 ****** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [osism.services.chrony : Populate service facts] **************************",
                                "Tuesday 24 February 2026  00:34:47 +0000 (0:00:00.520)       0:07:41.317 ****** ",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [osism.services.chrony : Manage timesyncd service] ************************",
                                "Tuesday 24 February 2026  00:34:53 +0000 (0:00:05.588)       0:07:46.906 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [osism.services.chrony : Include distribution specific install tasks] *****",
                                "Tuesday 24 February 2026  00:34:54 +0000 (0:00:00.612)       0:07:47.518 ****** ",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/services/roles/chrony/tasks/install-Debian-family.yml for testbed-node-3, testbed-node-4, testbed-node-5, testbed-manager, testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [osism.services.chrony : Install package] *********************************",
                                "Tuesday 24 February 2026  00:34:55 +0000 (0:00:01.241)       0:07:48.760 ****** ",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [osism.services.chrony : Manage chrony service] ***************************",
                                "Tuesday 24 February 2026  00:34:57 +0000 (0:00:02.207)       0:07:50.967 ****** ",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [osism.services.chrony : Check if configuration file exists] **************",
                                "Tuesday 24 February 2026  00:34:58 +0000 (0:00:01.225)       0:07:52.192 ****** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [osism.services.chrony : Copy configuration file] *************************",
                                "Tuesday 24 February 2026  00:34:59 +0000 (0:00:00.913)       0:07:53.106 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=/usr/share/ansible/collections/ansible_collections/osism/services/roles/chrony/templates/chrony.conf.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=/usr/share/ansible/collections/ansible_collections/osism/services/roles/chrony/templates/chrony.conf.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=/usr/share/ansible/collections/ansible_collections/osism/services/roles/chrony/templates/chrony.conf.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=/usr/share/ansible/collections/ansible_collections/osism/services/roles/chrony/templates/chrony.conf.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=/usr/share/ansible/collections/ansible_collections/osism/services/roles/chrony/templates/chrony.conf.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=/usr/share/ansible/collections/ansible_collections/osism/services/roles/chrony/templates/chrony.conf.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=/usr/share/ansible/collections/ansible_collections/osism/services/roles/chrony/templates/chrony.conf.j2)\u001b[0m",
                                "",
                                "TASK [osism.services.lldpd : Include distribution specific install tasks] ******",
                                "Tuesday 24 February 2026  00:35:01 +0000 (0:00:01.978)       0:07:55.084 ****** ",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/services/roles/lldpd/tasks/install-Debian-family.yml for testbed-node-3, testbed-node-4, testbed-node-5, testbed-manager, testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [osism.services.lldpd : Install lldpd package] ****************************",
                                "Tuesday 24 February 2026  00:35:02 +0000 (0:00:00.857)       0:07:55.942 ****** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [osism.services.lldpd : Manage lldpd service] *****************************",
                                "Tuesday 24 February 2026  00:35:12 +0000 (0:00:10.392)       0:08:06.335 ****** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [osism.commons.docker_compose : Reload systemd daemon] *********",
                                "Tuesday 24 February 2026  00:35:14 +0000 (0:00:01.864)       0:08:08.199 ****** ",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "",
                                "RUNNING HANDLER [osism.services.chrony : Restart chrony service] ***************",
                                "Tuesday 24 February 2026  00:35:16 +0000 (0:00:01.297)       0:08:09.496 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "PLAY [Apply bootstrap role part 2] *********************************************",
                                "",
                                "TASK [Include hardening role] **************************************************",
                                "Tuesday 24 February 2026  00:35:17 +0000 (0:00:01.360)       0:08:10.857 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "PLAY [Apply bootstrap roles part 3] ********************************************",
                                "",
                                "TASK [osism.services.journald : Copy configuration file] ***********************",
                                "Tuesday 24 February 2026  00:35:17 +0000 (0:00:00.491)       0:08:11.349 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [osism.services.journald : Manage journald service] ***********************",
                                "Tuesday 24 February 2026  00:35:19 +0000 (0:00:01.335)       0:08:12.684 ****** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [Include auditd role] *****************************************************",
                                "Tuesday 24 February 2026  00:35:20 +0000 (0:00:01.375)       0:08:14.060 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [Include smartd role] *****************************************************",
                                "Tuesday 24 February 2026  00:35:21 +0000 (0:00:00.586)       0:08:14.646 ****** ",
                                "\u001b[0;36mincluded: osism.services.smartd for testbed-node-3, testbed-node-4, testbed-node-5, testbed-manager, testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [osism.services.smartd : Include distribution specific install tasks] *****",
                                "Tuesday 24 February 2026  00:35:22 +0000 (0:00:00.832)       0:08:15.479 ****** ",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/services/roles/smartd/tasks/install-Debian-family.yml for testbed-node-3, testbed-node-4, testbed-node-5, testbed-manager, testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [osism.services.smartd : Install smartmontools package] *******************",
                                "Tuesday 24 February 2026  00:35:22 +0000 (0:00:00.779)       0:08:16.258 ****** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [osism.services.smartd : Create /var/log/smartd directory] ****************",
                                "Tuesday 24 February 2026  00:35:32 +0000 (0:00:09.728)       0:08:25.986 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [osism.services.smartd : Copy smartmontools configuration file] ***********",
                                "Tuesday 24 February 2026  00:35:33 +0000 (0:00:00.874)       0:08:26.861 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [osism.services.smartd : Manage smartd service] ***************************",
                                "Tuesday 24 February 2026  00:35:34 +0000 (0:00:01.387)       0:08:28.248 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [osism.services.journald : Restart journald service] ***********",
                                "Tuesday 24 February 2026  00:35:36 +0000 (0:00:02.020)       0:08:30.269 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [osism.services.smartd : Restart smartd service] ***************",
                                "Tuesday 24 February 2026  00:35:38 +0000 (0:00:01.227)       0:08:31.496 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "",
                                "PLAY [Set state bootstrap] *****************************************************",
                                "",
                                "TASK [Set osism.bootstrap.status fact] *****************************************",
                                "Tuesday 24 February 2026  00:35:39 +0000 (0:00:01.125)       0:08:32.621 ****** ",
                                "\u001b[0;36mincluded: osism.commons.state for testbed-node-3, testbed-node-4, testbed-node-5, testbed-manager, testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [osism.commons.state : Create custom facts directory] *********************",
                                "Tuesday 24 February 2026  00:35:40 +0000 (0:00:00.992)       0:08:33.614 ****** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [osism.commons.state : Write state into file] *****************************",
                                "Tuesday 24 February 2026  00:35:41 +0000 (0:00:00.885)       0:08:34.499 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [Set osism.bootstrap.timestamp fact] **************************************",
                                "Tuesday 24 February 2026  00:35:42 +0000 (0:00:01.117)       0:08:35.617 ****** ",
                                "\u001b[0;36mincluded: osism.commons.state for testbed-node-3, testbed-node-4, testbed-node-5, testbed-manager, testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [osism.commons.state : Create custom facts directory] *********************",
                                "Tuesday 24 February 2026  00:35:43 +0000 (0:00:01.032)       0:08:36.649 ****** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [osism.commons.state : Write state into file] *****************************",
                                "Tuesday 24 February 2026  00:35:44 +0000 (0:00:00.849)       0:08:37.498 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;33mtestbed-manager\u001b[0m            : \u001b[0;32mok=168 \u001b[0m \u001b[0;33mchanged=40  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=42  \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=177 \u001b[0m \u001b[0;33mchanged=69  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=36  \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=177 \u001b[0m \u001b[0;33mchanged=69  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=36  \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=177 \u001b[0m \u001b[0;33mchanged=69  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=36  \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-3\u001b[0m             : \u001b[0;32mok=175 \u001b[0m \u001b[0;33mchanged=65  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=38  \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-4\u001b[0m             : \u001b[0;32mok=175 \u001b[0m \u001b[0;33mchanged=65  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=37  \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-5\u001b[0m             : \u001b[0;32mok=175 \u001b[0m \u001b[0;33mchanged=65  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=37  \u001b[0m rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Tuesday 24 February 2026  00:35:45 +0000 (0:00:01.108)       0:08:38.607 ****** ",
                                "=============================================================================== ",
                                "osism.commons.packages : Install required packages --------------------- 85.77s",
                                "osism.commons.packages : Download required packages -------------------- 48.16s",
                                "osism.commons.cleanup : Cleanup installed packages --------------------- 36.19s",
                                "osism.commons.repository : Update package cache ------------------------ 17.06s",
                                "osism.services.docker : Install docker package ------------------------- 12.13s",
                                "osism.commons.systohc : Install util-linux-extra package --------------- 11.13s",
                                "osism.commons.packages : Remove dependencies that are no longer required -- 11.00s",
                                "osism.services.docker : Install docker-cli package --------------------- 10.59s",
                                "osism.services.lldpd : Install lldpd package --------------------------- 10.39s",
                                "osism.services.docker : Install containerd package --------------------- 10.38s",
                                "osism.services.smartd : Install smartmontools package ------------------- 9.73s",
                                "osism.services.docker : Add repository ---------------------------------- 9.26s",
                                "osism.commons.cleanup : Remove cloudinit package ------------------------ 9.25s",
                                "osism.commons.cleanup : Uninstall unattended-upgrades package ----------- 8.71s",
                                "osism.services.rng : Install rng package -------------------------------- 8.68s",
                                "osism.commons.docker_compose : Install docker-compose-plugin package ---- 8.55s",
                                "osism.services.docker : Install apt-transport-https package ------------- 7.74s",
                                "osism.commons.cleanup : Remove dependencies that are no longer required --- 6.84s",
                                "osism.commons.cleanup : Populate service facts -------------------------- 6.08s",
                                "osism.services.chrony : Populate service facts -------------------------- 5.59s",
                                "+ osism apply fail2ban",
                                "\u001b[32m2026-02-24 00:35:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mPrepare task for execution of fail2ban.\u001b[0m",
                                "\u001b[32m2026-02-24 00:35:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c6792215-2f8a-4696-9f5d-fb0127fc1bb6 (fail2ban) was prepared for execution.\u001b[0m",
                                "\u001b[32m2026-02-24 00:35:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task c6792215-2f8a-4696-9f5d-fb0127fc1bb6 (fail2ban) has been started and output is visible here.\u001b[0m",
                                "",
                                "PLAY [Apply role fail2ban] *****************************************************",
                                "",
                                "TASK [osism.services.fail2ban : Include distribution specific install tasks] ***",
                                "Tuesday 24 February 2026  00:36:02 +0000 (0:00:00.239)       0:00:00.239 ****** ",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/services/roles/fail2ban/tasks/install-Debian-family.yml for testbed-manager, testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m",
                                "",
                                "TASK [osism.services.fail2ban : Install fail2ban package] **********************",
                                "Tuesday 24 February 2026  00:36:03 +0000 (0:00:01.180)       0:00:01.419 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [osism.services.fail2ban : Copy configuration files] **********************",
                                "Tuesday 24 February 2026  00:36:15 +0000 (0:00:11.893)       0:00:13.313 ****** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.services.fail2ban : Manage fail2ban service] ***********************",
                                "Tuesday 24 February 2026  00:36:17 +0000 (0:00:01.509)       0:00:14.822 ****** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.services.fail2ban : Reload fail2ban configuration] *****************",
                                "Tuesday 24 February 2026  00:36:18 +0000 (0:00:01.538)       0:00:16.361 ****** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;33mtestbed-manager\u001b[0m            : \u001b[0;32mok=5   \u001b[0m \u001b[0;33mchanged=3   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=5   \u001b[0m \u001b[0;33mchanged=3   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=5   \u001b[0m \u001b[0;33mchanged=3   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=5   \u001b[0m \u001b[0;33mchanged=3   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-3\u001b[0m             : \u001b[0;32mok=5   \u001b[0m \u001b[0;33mchanged=3   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-4\u001b[0m             : \u001b[0;32mok=5   \u001b[0m \u001b[0;33mchanged=3   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-5\u001b[0m             : \u001b[0;32mok=5   \u001b[0m \u001b[0;33mchanged=3   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Tuesday 24 February 2026  00:36:20 +0000 (0:00:01.717)       0:00:18.079 ****** ",
                                "=============================================================================== ",
                                "osism.services.fail2ban : Install fail2ban package --------------------- 11.89s",
                                "osism.services.fail2ban : Reload fail2ban configuration ----------------- 1.72s",
                                "osism.services.fail2ban : Manage fail2ban service ----------------------- 1.54s",
                                "osism.services.fail2ban : Copy configuration files ---------------------- 1.51s",
                                "osism.services.fail2ban : Include distribution specific install tasks --- 1.18s",
                                "+ [[ -e /etc/redhat-release ]]",
                                "+ osism apply network",
                                "\u001b[32m2026-02-24 00:36:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mPrepare task for execution of network.\u001b[0m",
                                "\u001b[32m2026-02-24 00:36:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ee38c4cf-b985-4420-8dba-d9d7697ff09e (network) was prepared for execution.\u001b[0m",
                                "\u001b[32m2026-02-24 00:36:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task ee38c4cf-b985-4420-8dba-d9d7697ff09e (network) has been started and output is visible here.\u001b[0m",
                                "",
                                "PLAY [Apply role network] ******************************************************",
                                "",
                                "TASK [osism.commons.network : Gather variables for each operating system] ******",
                                "Tuesday 24 February 2026  00:36:37 +0000 (0:00:00.257)       0:00:00.257 ****** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.commons.network : Include type specific tasks] *********************",
                                "Tuesday 24 February 2026  00:36:38 +0000 (0:00:00.693)       0:00:00.951 ****** ",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/network/tasks/netplan-Debian-family.yml for testbed-manager, testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m",
                                "",
                                "TASK [osism.commons.network : Install required packages] ***********************",
                                "Tuesday 24 February 2026  00:36:39 +0000 (0:00:01.208)       0:00:02.159 ****** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [osism.commons.network : Remove ifupdown package] *************************",
                                "Tuesday 24 February 2026  00:36:41 +0000 (0:00:02.126)       0:00:04.286 ****** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.commons.network : Create required directories] *********************",
                                "Tuesday 24 February 2026  00:36:43 +0000 (0:00:01.850)       0:00:06.136 ****** ",
                                "\u001b[0;32mok: [testbed-manager] => (item=/etc/netplan)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item=/etc/netplan)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0] => (item=/etc/netplan)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item=/etc/netplan)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item=/etc/netplan)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item=/etc/netplan)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item=/etc/netplan)\u001b[0m",
                                "",
                                "TASK [osism.commons.network : Prepare netplan configuration template] **********",
                                "Tuesday 24 February 2026  00:36:44 +0000 (0:00:01.015)       0:00:07.152 ****** ",
                                "\u001b[0;32mok: [testbed-node-1 -> localhost]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2 -> localhost]\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager -> localhost]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5 -> localhost]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0 -> localhost]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3 -> localhost]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4 -> localhost]\u001b[0m",
                                "",
                                "TASK [osism.commons.network : Copy netplan configuration] **********************",
                                "Tuesday 24 February 2026  00:36:48 +0000 (0:00:03.559)       0:00:10.712 ****** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.commons.network : Remove netplan configuration template] ***********",
                                "Tuesday 24 February 2026  00:36:50 +0000 (0:00:01.654)       0:00:12.366 ****** ",
                                "\u001b[0;32mok: [testbed-node-0 -> localhost]\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager -> localhost]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2 -> localhost]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1 -> localhost]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4 -> localhost]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3 -> localhost]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5 -> localhost]\u001b[0m",
                                "",
                                "TASK [osism.commons.network : Check if path for interface file exists] *********",
                                "Tuesday 24 February 2026  00:36:52 +0000 (0:00:01.949)       0:00:14.316 ****** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.commons.network : Copy interfaces file] ****************************",
                                "Tuesday 24 February 2026  00:36:53 +0000 (0:00:01.137)       0:00:15.454 ****** ",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.commons.network : Install package networkd-dispatcher] *************",
                                "Tuesday 24 February 2026  00:36:53 +0000 (0:00:00.714)       0:00:16.168 ****** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.commons.network : Copy dispatcher scripts] *************************",
                                "Tuesday 24 February 2026  00:36:56 +0000 (0:00:02.427)       0:00:18.596 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item={'dest': 'routable.d/iptables.sh', 'src': '/opt/configuration/network/iptables.sh'})\u001b[0m",
                                "",
                                "TASK [osism.commons.network : Manage service networkd-dispatcher] **************",
                                "Tuesday 24 February 2026  00:36:57 +0000 (0:00:00.912)       0:00:19.508 ****** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.commons.network : Include cleanup tasks] ***************************",
                                "Tuesday 24 February 2026  00:36:58 +0000 (0:00:01.693)       0:00:21.201 ****** ",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/network/tasks/cleanup-netplan.yml for testbed-manager, testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m",
                                "",
                                "TASK [osism.commons.network : List existing configuration files] ***************",
                                "Tuesday 24 February 2026  00:37:00 +0000 (0:00:01.280)       0:00:22.482 ****** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.commons.network : Set network_configured_files fact] ***************",
                                "Tuesday 24 February 2026  00:37:01 +0000 (0:00:01.169)       0:00:23.651 ****** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.commons.network : Remove unused configuration files] ***************",
                                "Tuesday 24 February 2026  00:37:02 +0000 (0:00:00.706)       0:00:24.357 ****** ",
                                "\u001b[0;36mskipping: [testbed-manager] => (item=/etc/netplan/01-osism.yaml) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=/etc/netplan/01-osism.yaml) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=/etc/netplan/01-osism.yaml) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=/etc/netplan/01-osism.yaml) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=/etc/netplan/50-cloud-init.yaml)\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=/etc/netplan/01-osism.yaml) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=/etc/netplan/01-osism.yaml) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=/etc/netplan/50-cloud-init.yaml)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=/etc/netplan/50-cloud-init.yaml)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=/etc/netplan/50-cloud-init.yaml)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=/etc/netplan/50-cloud-init.yaml)\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=/etc/netplan/01-osism.yaml) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=/etc/netplan/50-cloud-init.yaml)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=/etc/netplan/50-cloud-init.yaml)\u001b[0m",
                                "",
                                "TASK [osism.commons.network : Include dummy interfaces] ************************",
                                "Tuesday 24 February 2026  00:37:03 +0000 (0:00:01.222)       0:00:25.579 ****** ",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.commons.network : Include vxlan interfaces] ************************",
                                "Tuesday 24 February 2026  00:37:03 +0000 (0:00:00.603)       0:00:26.183 ****** ",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/network/tasks/vxlan-interfaces.yml for testbed-node-4, testbed-manager, testbed-node-1, testbed-node-0, testbed-node-2, testbed-node-3, testbed-node-5\u001b[0m",
                                "",
                                "TASK [osism.commons.network : Create systemd networkd netdev files] ************",
                                "Tuesday 24 February 2026  00:37:08 +0000 (0:00:04.481)       0:00:30.664 ****** ",
                                "\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'vxlan0', 'value': {'addresses': ['192.168.112.5/20'], 'dests': ['192.168.16.10', '192.168.16.11', '192.168.16.12', '192.168.16.13', '192.168.16.14', '192.168.16.15'], 'local_ip': '192.168.16.5', 'mtu': 1350, 'vni': 42}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'vxlan0', 'value': {'addresses': [], 'dests': ['192.168.16.11', '192.168.16.12', '192.168.16.13', '192.168.16.14', '192.168.16.15', '192.168.16.5'], 'local_ip': '192.168.16.10', 'mtu': 1350, 'vni': 42}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'vxlan0', 'value': {'addresses': [], 'dests': ['192.168.16.10', '192.168.16.11', '192.168.16.13', '192.168.16.14', '192.168.16.15', '192.168.16.5'], 'local_ip': '192.168.16.12', 'mtu': 1350, 'vni': 42}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'vxlan1', 'value': {'addresses': ['192.168.128.5/20'], 'dests': ['192.168.16.10', '192.168.16.11', '192.168.16.12', '192.168.16.13', '192.168.16.14', '192.168.16.15'], 'local_ip': '192.168.16.5', 'mtu': 1350, 'vni': 23}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'vxlan0', 'value': {'addresses': [], 'dests': ['192.168.16.10', '192.168.16.12', '192.168.16.13', '192.168.16.14', '192.168.16.15', '192.168.16.5'], 'local_ip': '192.168.16.11', 'mtu': 1350, 'vni': 42}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'vxlan0', 'value': {'addresses': [], 'dests': ['192.168.16.10', '192.168.16.11', '192.168.16.12', '192.168.16.13', '192.168.16.14', '192.168.16.5'], 'local_ip': '192.168.16.15', 'mtu': 1350, 'vni': 42}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'vxlan0', 'value': {'addresses': [], 'dests': ['192.168.16.10', '192.168.16.11', '192.168.16.12', '192.168.16.14', '192.168.16.15', '192.168.16.5'], 'local_ip': '192.168.16.13', 'mtu': 1350, 'vni': 42}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'vxlan1', 'value': {'addresses': ['192.168.128.10/20'], 'dests': ['192.168.16.11', '192.168.16.12', '192.168.16.13', '192.168.16.14', '192.168.16.15', '192.168.16.5'], 'local_ip': '192.168.16.10', 'mtu': 1350, 'vni': 23}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'vxlan0', 'value': {'addresses': [], 'dests': ['192.168.16.10', '192.168.16.11', '192.168.16.12', '192.168.16.13', '192.168.16.15', '192.168.16.5'], 'local_ip': '192.168.16.14', 'mtu': 1350, 'vni': 42}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'vxlan1', 'value': {'addresses': ['192.168.128.12/20'], 'dests': ['192.168.16.10', '192.168.16.11', '192.168.16.13', '192.168.16.14', '192.168.16.15', '192.168.16.5'], 'local_ip': '192.168.16.12', 'mtu': 1350, 'vni': 23}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'vxlan1', 'value': {'addresses': ['192.168.128.11/20'], 'dests': ['192.168.16.10', '192.168.16.12', '192.168.16.13', '192.168.16.14', '192.168.16.15', '192.168.16.5'], 'local_ip': '192.168.16.11', 'mtu': 1350, 'vni': 23}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'vxlan1', 'value': {'addresses': ['192.168.128.15/20'], 'dests': ['192.168.16.10', '192.168.16.11', '192.168.16.12', '192.168.16.13', '192.168.16.14', '192.168.16.5'], 'local_ip': '192.168.16.15', 'mtu': 1350, 'vni': 23}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'vxlan1', 'value': {'addresses': ['192.168.128.13/20'], 'dests': ['192.168.16.10', '192.168.16.11', '192.168.16.12', '192.168.16.14', '192.168.16.15', '192.168.16.5'], 'local_ip': '192.168.16.13', 'mtu': 1350, 'vni': 23}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'vxlan1', 'value': {'addresses': ['192.168.128.14/20'], 'dests': ['192.168.16.10', '192.168.16.11', '192.168.16.12', '192.168.16.13', '192.168.16.15', '192.168.16.5'], 'local_ip': '192.168.16.14', 'mtu': 1350, 'vni': 23}})\u001b[0m",
                                "",
                                "TASK [osism.commons.network : Create systemd networkd network files] ***********",
                                "Tuesday 24 February 2026  00:37:13 +0000 (0:00:05.206)       0:00:35.871 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'vxlan0', 'value': {'addresses': [], 'dests': ['192.168.16.10', '192.168.16.11', '192.168.16.13', '192.168.16.14', '192.168.16.15', '192.168.16.5'], 'local_ip': '192.168.16.12', 'mtu': 1350, 'vni': 42}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'vxlan0', 'value': {'addresses': ['192.168.112.5/20'], 'dests': ['192.168.16.10', '192.168.16.11', '192.168.16.12', '192.168.16.13', '192.168.16.14', '192.168.16.15'], 'local_ip': '192.168.16.5', 'mtu': 1350, 'vni': 42}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'vxlan0', 'value': {'addresses': [], 'dests': ['192.168.16.11', '192.168.16.12', '192.168.16.13', '192.168.16.14', '192.168.16.15', '192.168.16.5'], 'local_ip': '192.168.16.10', 'mtu': 1350, 'vni': 42}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'vxlan0', 'value': {'addresses': [], 'dests': ['192.168.16.10', '192.168.16.12', '192.168.16.13', '192.168.16.14', '192.168.16.15', '192.168.16.5'], 'local_ip': '192.168.16.11', 'mtu': 1350, 'vni': 42}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'vxlan0', 'value': {'addresses': [], 'dests': ['192.168.16.10', '192.168.16.11', '192.168.16.12', '192.168.16.13', '192.168.16.14', '192.168.16.5'], 'local_ip': '192.168.16.15', 'mtu': 1350, 'vni': 42}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'vxlan0', 'value': {'addresses': [], 'dests': ['192.168.16.10', '192.168.16.11', '192.168.16.12', '192.168.16.14', '192.168.16.15', '192.168.16.5'], 'local_ip': '192.168.16.13', 'mtu': 1350, 'vni': 42}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'vxlan1', 'value': {'addresses': ['192.168.128.5/20'], 'dests': ['192.168.16.10', '192.168.16.11', '192.168.16.12', '192.168.16.13', '192.168.16.14', '192.168.16.15'], 'local_ip': '192.168.16.5', 'mtu': 1350, 'vni': 23}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'vxlan1', 'value': {'addresses': ['192.168.128.12/20'], 'dests': ['192.168.16.10', '192.168.16.11', '192.168.16.13', '192.168.16.14', '192.168.16.15', '192.168.16.5'], 'local_ip': '192.168.16.12', 'mtu': 1350, 'vni': 23}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'vxlan0', 'value': {'addresses': [], 'dests': ['192.168.16.10', '192.168.16.11', '192.168.16.12', '192.168.16.13', '192.168.16.15', '192.168.16.5'], 'local_ip': '192.168.16.14', 'mtu': 1350, 'vni': 42}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'vxlan1', 'value': {'addresses': ['192.168.128.10/20'], 'dests': ['192.168.16.11', '192.168.16.12', '192.168.16.13', '192.168.16.14', '192.168.16.15', '192.168.16.5'], 'local_ip': '192.168.16.10', 'mtu': 1350, 'vni': 23}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'vxlan1', 'value': {'addresses': ['192.168.128.11/20'], 'dests': ['192.168.16.10', '192.168.16.12', '192.168.16.13', '192.168.16.14', '192.168.16.15', '192.168.16.5'], 'local_ip': '192.168.16.11', 'mtu': 1350, 'vni': 23}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'vxlan1', 'value': {'addresses': ['192.168.128.15/20'], 'dests': ['192.168.16.10', '192.168.16.11', '192.168.16.12', '192.168.16.13', '192.168.16.14', '192.168.16.5'], 'local_ip': '192.168.16.15', 'mtu': 1350, 'vni': 23}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'vxlan1', 'value': {'addresses': ['192.168.128.13/20'], 'dests': ['192.168.16.10', '192.168.16.11', '192.168.16.12', '192.168.16.14', '192.168.16.15', '192.168.16.5'], 'local_ip': '192.168.16.13', 'mtu': 1350, 'vni': 23}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'vxlan1', 'value': {'addresses': ['192.168.128.14/20'], 'dests': ['192.168.16.10', '192.168.16.11', '192.168.16.12', '192.168.16.13', '192.168.16.15', '192.168.16.5'], 'local_ip': '192.168.16.14', 'mtu': 1350, 'vni': 23}})\u001b[0m",
                                "",
                                "TASK [osism.commons.network : Include networkd cleanup tasks] ******************",
                                "Tuesday 24 February 2026  00:37:18 +0000 (0:00:05.271)       0:00:41.142 ****** ",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/network/tasks/cleanup-networkd.yml for testbed-manager, testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m",
                                "",
                                "TASK [osism.commons.network : List existing configuration files] ***************",
                                "Tuesday 24 February 2026  00:37:20 +0000 (0:00:01.223)       0:00:42.366 ****** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.commons.network : Remove unused configuration files] ***************",
                                "Tuesday 24 February 2026  00:37:20 +0000 (0:00:00.877)       0:00:43.243 ****** ",
                                "\u001b[0;36mskipping: [testbed-manager] => (item=/etc/systemd/network/30-vxlan1.network) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager] => (item=/etc/systemd/network/30-vxlan0.network) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager] => (item=/etc/systemd/network/30-vxlan1.netdev) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager] => (item=/etc/systemd/network/30-vxlan0.netdev) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=/etc/systemd/network/30-vxlan1.network) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=/etc/systemd/network/30-vxlan0.network) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=/etc/systemd/network/30-vxlan1.netdev) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=/etc/systemd/network/30-vxlan0.netdev) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=/etc/systemd/network/30-vxlan1.network) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=/etc/systemd/network/30-vxlan0.network) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=/etc/systemd/network/30-vxlan1.netdev) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=/etc/systemd/network/30-vxlan0.netdev) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=/etc/systemd/network/30-vxlan1.network) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=/etc/systemd/network/30-vxlan0.network) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=/etc/systemd/network/30-vxlan1.netdev) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=/etc/systemd/network/30-vxlan0.netdev) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=/etc/systemd/network/30-vxlan1.network) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=/etc/systemd/network/30-vxlan0.network) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=/etc/systemd/network/30-vxlan1.netdev) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=/etc/systemd/network/30-vxlan0.netdev) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=/etc/systemd/network/30-vxlan1.network) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=/etc/systemd/network/30-vxlan0.network) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=/etc/systemd/network/30-vxlan1.netdev) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=/etc/systemd/network/30-vxlan0.netdev) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=/etc/systemd/network/30-vxlan1.network) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=/etc/systemd/network/30-vxlan0.network) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=/etc/systemd/network/30-vxlan1.netdev) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=/etc/systemd/network/30-vxlan0.netdev) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.commons.network : Include network extra init] **********************",
                                "Tuesday 24 February 2026  00:37:21 +0000 (0:00:00.969)       0:00:44.212 ****** ",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/commons/roles/network/tasks/network-extra-init.yml for testbed-manager, testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m",
                                "",
                                "TASK [osism.commons.network : Deploy network-extra-init script] ****************",
                                "Tuesday 24 February 2026  00:37:23 +0000 (0:00:01.312)       0:00:45.525 ****** ",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.commons.network : Deploy network-extra-init systemd service] *******",
                                "Tuesday 24 February 2026  00:37:23 +0000 (0:00:00.727)       0:00:46.253 ****** ",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.commons.network : Enable and start network-extra-init service] *****",
                                "Tuesday 24 February 2026  00:37:24 +0000 (0:00:00.912)       0:00:47.165 ****** ",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.commons.network : Disable and stop network-extra-init service] *****",
                                "Tuesday 24 February 2026  00:37:25 +0000 (0:00:00.665)       0:00:47.831 ****** ",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.commons.network : Remove network-extra-init systemd service] *******",
                                "Tuesday 24 February 2026  00:37:27 +0000 (0:00:01.812)       0:00:49.643 ****** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.commons.network : Remove network-extra-init script] ****************",
                                "Tuesday 24 February 2026  00:37:28 +0000 (0:00:01.000)       0:00:50.643 ****** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "RUNNING HANDLER [osism.commons.network : Reload systemd-networkd] **************",
                                "Tuesday 24 February 2026  00:37:30 +0000 (0:00:02.417)       0:00:53.061 ****** ",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "RUNNING HANDLER [osism.commons.network : Netplan configuration changed] ********",
                                "Tuesday 24 February 2026  00:37:31 +0000 (0:00:01.084)       0:00:54.145 ****** ",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;33mtestbed-manager\u001b[0m            : \u001b[0;32mok=25  \u001b[0m \u001b[0;33mchanged=5   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=8   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=24  \u001b[0m \u001b[0;33mchanged=5   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=9   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=24  \u001b[0m \u001b[0;33mchanged=5   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=9   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=24  \u001b[0m \u001b[0;33mchanged=5   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=9   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-3\u001b[0m             : \u001b[0;32mok=24  \u001b[0m \u001b[0;33mchanged=5   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=9   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-4\u001b[0m             : \u001b[0;32mok=24  \u001b[0m \u001b[0;33mchanged=5   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=9   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-5\u001b[0m             : \u001b[0;32mok=24  \u001b[0m \u001b[0;33mchanged=5   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=9   \u001b[0m rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Tuesday 24 February 2026  00:37:32 +0000 (0:00:00.570)       0:00:54.716 ****** ",
                                "=============================================================================== ",
                                "osism.commons.network : Create systemd networkd network files ----------- 5.27s",
                                "osism.commons.network : Create systemd networkd netdev files ------------ 5.21s",
                                "osism.commons.network : Include vxlan interfaces ------------------------ 4.48s",
                                "osism.commons.network : Prepare netplan configuration template ---------- 3.56s",
                                "osism.commons.network : Install package networkd-dispatcher ------------- 2.43s",
                                "osism.commons.network : Remove network-extra-init script ---------------- 2.42s",
                                "osism.commons.network : Install required packages ----------------------- 2.13s",
                                "osism.commons.network : Remove netplan configuration template ----------- 1.95s",
                                "osism.commons.network : Remove ifupdown package ------------------------- 1.85s",
                                "osism.commons.network : Disable and stop network-extra-init service ----- 1.81s",
                                "osism.commons.network : Manage service networkd-dispatcher -------------- 1.69s",
                                "osism.commons.network : Copy netplan configuration ---------------------- 1.65s",
                                "osism.commons.network : Include network extra init ---------------------- 1.31s",
                                "osism.commons.network : Include cleanup tasks --------------------------- 1.28s",
                                "osism.commons.network : Include networkd cleanup tasks ------------------ 1.22s",
                                "osism.commons.network : Remove unused configuration files --------------- 1.22s",
                                "osism.commons.network : Include type specific tasks --------------------- 1.21s",
                                "osism.commons.network : List existing configuration files --------------- 1.17s",
                                "osism.commons.network : Check if path for interface file exists --------- 1.14s",
                                "osism.commons.network : Reload systemd-networkd ------------------------- 1.08s",
                                "+ osism apply wireguard",
                                "\u001b[32m2026-02-24 00:37:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mPrepare task for execution of wireguard.\u001b[0m",
                                "\u001b[32m2026-02-24 00:37:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1a3ae285-433a-47d0-8887-3b85c5dc6177 (wireguard) was prepared for execution.\u001b[0m",
                                "\u001b[32m2026-02-24 00:37:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task 1a3ae285-433a-47d0-8887-3b85c5dc6177 (wireguard) has been started and output is visible here.\u001b[0m",
                                "",
                                "PLAY [Apply role wireguard] ****************************************************",
                                "",
                                "TASK [osism.services.wireguard : Install iptables package] *********************",
                                "Tuesday 24 February 2026  00:37:49 +0000 (0:00:00.243)       0:00:00.243 ****** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.services.wireguard : Install wireguard package] ********************",
                                "Tuesday 24 February 2026  00:37:51 +0000 (0:00:01.628)       0:00:01.872 ****** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.services.wireguard : Create public and private key - server] *******",
                                "Tuesday 24 February 2026  00:37:58 +0000 (0:00:07.002)       0:00:08.875 ****** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.services.wireguard : Create preshared key] *************************",
                                "Tuesday 24 February 2026  00:37:58 +0000 (0:00:00.549)       0:00:09.424 ****** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.services.wireguard : Get preshared key] ****************************",
                                "Tuesday 24 February 2026  00:37:59 +0000 (0:00:00.461)       0:00:09.885 ****** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.services.wireguard : Get public key - server] **********************",
                                "Tuesday 24 February 2026  00:38:00 +0000 (0:00:00.720)       0:00:10.606 ****** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.services.wireguard : Get private key - server] *********************",
                                "Tuesday 24 February 2026  00:38:00 +0000 (0:00:00.424)       0:00:11.031 ****** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.services.wireguard : Copy wg0.conf configuration file] *************",
                                "Tuesday 24 February 2026  00:38:01 +0000 (0:00:00.508)       0:00:11.539 ****** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.services.wireguard : Copy client configuration files] **************",
                                "Tuesday 24 February 2026  00:38:02 +0000 (0:00:01.341)       0:00:12.881 ****** ",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=None)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.services.wireguard : Manage wg-quick@wg0.service service] **********",
                                "Tuesday 24 February 2026  00:38:03 +0000 (0:00:01.027)       0:00:13.908 ****** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "RUNNING HANDLER [osism.services.wireguard : Restart wg0 service] ***************",
                                "Tuesday 24 February 2026  00:38:05 +0000 (0:00:01.749)       0:00:15.658 ****** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;33mtestbed-manager\u001b[0m            : \u001b[0;32mok=11  \u001b[0m \u001b[0;33mchanged=7   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Tuesday 24 February 2026  00:38:06 +0000 (0:00:01.055)       0:00:16.714 ****** ",
                                "=============================================================================== ",
                                "osism.services.wireguard : Install wireguard package -------------------- 7.00s",
                                "osism.services.wireguard : Manage wg-quick@wg0.service service ---------- 1.75s",
                                "osism.services.wireguard : Install iptables package --------------------- 1.63s",
                                "osism.services.wireguard : Copy wg0.conf configuration file ------------- 1.34s",
                                "osism.services.wireguard : Restart wg0 service -------------------------- 1.06s",
                                "osism.services.wireguard : Copy client configuration files -------------- 1.03s",
                                "osism.services.wireguard : Get preshared key ---------------------------- 0.72s",
                                "osism.services.wireguard : Create public and private key - server ------- 0.55s",
                                "osism.services.wireguard : Get private key - server --------------------- 0.51s",
                                "osism.services.wireguard : Create preshared key ------------------------- 0.46s",
                                "osism.services.wireguard : Get public key - server ---------------------- 0.42s",
                                "+ sh -c /opt/configuration/scripts/prepare-wireguard-configuration.sh",
                                "  % Total    % Received % Xferd  Average Speed   Time    Time     Time  Current",
                                "                                 Dload  Upload   Total   Spent    Left  Speed",
                                "",
                                "  0     0    0     0    0     0      0      0 --:--:-- --:--:-- --:--:--     0",
                                "100    15  100    15    0     0    195      0 --:--:-- --:--:-- --:--:--   197",
                                "+ osism apply --environment custom workarounds",
                                "\u001b[32m2026-02-24 00:38:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTrying to run play workarounds in environment custom\u001b[0m",
                                "\u001b[32m2026-02-24 00:38:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mPrepare task for execution of workarounds.\u001b[0m",
                                "\u001b[32m2026-02-24 00:38:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 8e00bff8-7c24-4924-8aee-b1d560d35af1 (workarounds) was prepared for execution.\u001b[0m",
                                "\u001b[32m2026-02-24 00:38:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task 8e00bff8-7c24-4924-8aee-b1d560d35af1 (workarounds) has been started and output is visible here.\u001b[0m",
                                "",
                                "PLAY [Group hosts based on configuration] **************************************",
                                "",
                                "TASK [Group hosts based on virtualization_role] ********************************",
                                "Tuesday 24 February 2026  00:38:23 +0000 (0:00:00.132)       0:00:00.132 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=virtualization_role_guest)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=virtualization_role_guest)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=virtualization_role_guest)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=virtualization_role_guest)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=virtualization_role_guest)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=virtualization_role_guest)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=virtualization_role_guest)\u001b[0m",
                                "",
                                "PLAY [Apply netplan configuration on the manager node] *************************",
                                "",
                                "TASK [Apply netplan configuration] *********************************************",
                                "Tuesday 24 February 2026  00:38:24 +0000 (0:00:00.851)       0:00:00.984 ****** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "",
                                "PLAY [Apply netplan configuration on all other nodes] **************************",
                                "",
                                "TASK [Apply netplan configuration] *********************************************",
                                "Tuesday 24 February 2026  00:38:27 +0000 (0:00:02.538)       0:00:03.522 ****** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "PLAY [Add custom CA certificates to non-manager nodes] *************************",
                                "",
                                "TASK [Copy custom CA certificates] *********************************************",
                                "Tuesday 24 February 2026  00:38:29 +0000 (0:00:01.863)       0:00:05.386 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=/opt/configuration/environments/kolla/certificates/ca/testbed.crt)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=/opt/configuration/environments/kolla/certificates/ca/testbed.crt)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=/opt/configuration/environments/kolla/certificates/ca/testbed.crt)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=/opt/configuration/environments/kolla/certificates/ca/testbed.crt)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=/opt/configuration/environments/kolla/certificates/ca/testbed.crt)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=/opt/configuration/environments/kolla/certificates/ca/testbed.crt)\u001b[0m",
                                "",
                                "TASK [Run update-ca-certificates] **********************************************",
                                "Tuesday 24 February 2026  00:38:30 +0000 (0:00:01.581)       0:00:06.967 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Run update-ca-trust] *****************************************************",
                                "Tuesday 24 February 2026  00:38:34 +0000 (0:00:03.867)       0:00:10.834 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "PLAY [Add a workaround service] ************************************************",
                                "",
                                "TASK [Copy workarounds.sh scripts] *********************************************",
                                "Tuesday 24 February 2026  00:38:35 +0000 (0:00:00.711)       0:00:11.546 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Copy workarounds systemd unit file] **************************************",
                                "Tuesday 24 February 2026  00:38:36 +0000 (0:00:01.525)       0:00:13.071 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [Reload systemd daemon] ***************************************************",
                                "Tuesday 24 February 2026  00:38:38 +0000 (0:00:01.604)       0:00:14.676 ****** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [Enable workarounds.service (Debian)] *************************************",
                                "Tuesday 24 February 2026  00:38:39 +0000 (0:00:01.550)       0:00:16.226 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [Enable and start workarounds.service (RedHat)] ***************************",
                                "Tuesday 24 February 2026  00:38:41 +0000 (0:00:01.706)       0:00:17.933 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "",
                                "PLAY [On Ubuntu 24.04 install python3-docker from Debian Sid] ******************",
                                "",
                                "TASK [Install python3-docker] **************************************************",
                                "Tuesday 24 February 2026  00:38:42 +0000 (0:00:00.608)       0:00:18.541 ****** ",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;33mtestbed-manager\u001b[0m            : \u001b[0;32mok=7   \u001b[0m \u001b[0;33mchanged=4   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=1   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=9   \u001b[0m \u001b[0;33mchanged=6   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=9   \u001b[0m \u001b[0;33mchanged=6   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=9   \u001b[0m \u001b[0;33mchanged=6   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-3\u001b[0m             : \u001b[0;32mok=9   \u001b[0m \u001b[0;33mchanged=6   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-4\u001b[0m             : \u001b[0;32mok=9   \u001b[0m \u001b[0;33mchanged=6   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-5\u001b[0m             : \u001b[0;32mok=9   \u001b[0m \u001b[0;33mchanged=6   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Tuesday 24 February 2026  00:38:45 +0000 (0:00:02.839)       0:00:21.381 ****** ",
                                "=============================================================================== ",
                                "Run update-ca-certificates ---------------------------------------------- 3.87s",
                                "Install python3-docker -------------------------------------------------- 2.84s",
                                "Apply netplan configuration --------------------------------------------- 2.54s",
                                "Apply netplan configuration --------------------------------------------- 1.86s",
                                "Enable workarounds.service (Debian) ------------------------------------- 1.71s",
                                "Copy workarounds systemd unit file -------------------------------------- 1.60s",
                                "Copy custom CA certificates --------------------------------------------- 1.58s",
                                "Reload systemd daemon --------------------------------------------------- 1.55s",
                                "Copy workarounds.sh scripts --------------------------------------------- 1.53s",
                                "Group hosts based on virtualization_role -------------------------------- 0.85s",
                                "Run update-ca-trust ----------------------------------------------------- 0.71s",
                                "Enable and start workarounds.service (RedHat) --------------------------- 0.61s",
                                "+ osism apply reboot -l testbed-nodes -e ireallymeanit=yes",
                                "\u001b[32m2026-02-24 00:38:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mPrepare task for execution of reboot.\u001b[0m",
                                "\u001b[32m2026-02-24 00:38:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 11252132-c821-496b-81ce-f5f47a776033 (reboot) was prepared for execution.\u001b[0m",
                                "\u001b[32m2026-02-24 00:38:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task 11252132-c821-496b-81ce-f5f47a776033 (reboot) has been started and output is visible here.\u001b[0m",
                                "",
                                "PLAY [Reboot systems] **********************************************************",
                                "",
                                "TASK [Exit playbook, if user did not mean to reboot systems] *******************",
                                "Tuesday 24 February 2026  00:39:01 +0000 (0:00:00.149)       0:00:00.149 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [Reboot system - do not wait for the reboot to complete] ******************",
                                "Tuesday 24 February 2026  00:39:01 +0000 (0:00:00.086)       0:00:00.235 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [Reboot system - wait for the reboot to complete] *************************",
                                "Tuesday 24 February 2026  00:39:02 +0000 (0:00:00.923)       0:00:01.159 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "PLAY [Reboot systems] **********************************************************",
                                "",
                                "TASK [Exit playbook, if user did not mean to reboot systems] *******************",
                                "Tuesday 24 February 2026  00:39:02 +0000 (0:00:00.090)       0:00:01.249 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [Reboot system - do not wait for the reboot to complete] ******************",
                                "Tuesday 24 February 2026  00:39:03 +0000 (0:00:00.081)       0:00:01.331 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [Reboot system - wait for the reboot to complete] *************************",
                                "Tuesday 24 February 2026  00:39:03 +0000 (0:00:00.656)       0:00:01.987 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "PLAY [Reboot systems] **********************************************************",
                                "",
                                "TASK [Exit playbook, if user did not mean to reboot systems] *******************",
                                "Tuesday 24 February 2026  00:39:03 +0000 (0:00:00.087)       0:00:02.075 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [Reboot system - do not wait for the reboot to complete] ******************",
                                "Tuesday 24 February 2026  00:39:03 +0000 (0:00:00.125)       0:00:02.201 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [Reboot system - wait for the reboot to complete] *************************",
                                "Tuesday 24 February 2026  00:39:04 +0000 (0:00:00.652)       0:00:02.853 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "PLAY [Reboot systems] **********************************************************",
                                "",
                                "TASK [Exit playbook, if user did not mean to reboot systems] *******************",
                                "Tuesday 24 February 2026  00:39:04 +0000 (0:00:00.119)       0:00:02.972 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Reboot system - do not wait for the reboot to complete] ******************",
                                "Tuesday 24 February 2026  00:39:04 +0000 (0:00:00.088)       0:00:03.061 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Reboot system - wait for the reboot to complete] *************************",
                                "Tuesday 24 February 2026  00:39:05 +0000 (0:00:00.694)       0:00:03.755 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "PLAY [Reboot systems] **********************************************************",
                                "",
                                "TASK [Exit playbook, if user did not mean to reboot systems] *******************",
                                "Tuesday 24 February 2026  00:39:05 +0000 (0:00:00.101)       0:00:03.857 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Reboot system - do not wait for the reboot to complete] ******************",
                                "Tuesday 24 February 2026  00:39:05 +0000 (0:00:00.096)       0:00:03.954 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Reboot system - wait for the reboot to complete] *************************",
                                "Tuesday 24 February 2026  00:39:06 +0000 (0:00:00.686)       0:00:04.640 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "PLAY [Reboot systems] **********************************************************",
                                "",
                                "TASK [Exit playbook, if user did not mean to reboot systems] *******************",
                                "Tuesday 24 February 2026  00:39:06 +0000 (0:00:00.110)       0:00:04.751 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Reboot system - do not wait for the reboot to complete] ******************",
                                "Tuesday 24 February 2026  00:39:06 +0000 (0:00:00.086)       0:00:04.838 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Reboot system - wait for the reboot to complete] *************************",
                                "Tuesday 24 February 2026  00:39:07 +0000 (0:00:00.720)       0:00:05.559 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=1   \u001b[0m \u001b[0;33mchanged=1   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=1   \u001b[0m \u001b[0;33mchanged=1   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=1   \u001b[0m \u001b[0;33mchanged=1   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-3\u001b[0m             : \u001b[0;32mok=1   \u001b[0m \u001b[0;33mchanged=1   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-4\u001b[0m             : \u001b[0;32mok=1   \u001b[0m \u001b[0;33mchanged=1   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-5\u001b[0m             : \u001b[0;32mok=1   \u001b[0m \u001b[0;33mchanged=1   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Tuesday 24 February 2026  00:39:07 +0000 (0:00:00.047)       0:00:05.606 ****** ",
                                "=============================================================================== ",
                                "Reboot system - do not wait for the reboot to complete ------------------ 4.33s",
                                "Exit playbook, if user did not mean to reboot systems ------------------- 0.57s",
                                "Reboot system - wait for the reboot to complete ------------------------- 0.56s",
                                "+ osism apply wait-for-connection -l testbed-nodes -e ireallymeanit=yes",
                                "\u001b[32m2026-02-24 00:39:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mPrepare task for execution of wait-for-connection.\u001b[0m",
                                "\u001b[32m2026-02-24 00:39:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4909c938-92d1-403b-8766-57378ac3da4b (wait-for-connection) was prepared for execution.\u001b[0m",
                                "\u001b[32m2026-02-24 00:39:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task 4909c938-92d1-403b-8766-57378ac3da4b (wait-for-connection) has been started and output is visible here.\u001b[0m",
                                "",
                                "PLAY [Wait until remote systems are reachable] *********************************",
                                "",
                                "TASK [Wait until remote system is reachable] ***********************************",
                                "Tuesday 24 February 2026  00:39:25 +0000 (0:00:00.215)       0:00:00.215 ****** ",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;32mtestbed-node-0\u001b[0m             : \u001b[0;32mok=1   \u001b[0m changed=0    unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "\u001b[0;32mtestbed-node-1\u001b[0m             : \u001b[0;32mok=1   \u001b[0m changed=0    unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "\u001b[0;32mtestbed-node-2\u001b[0m             : \u001b[0;32mok=1   \u001b[0m changed=0    unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "\u001b[0;32mtestbed-node-3\u001b[0m             : \u001b[0;32mok=1   \u001b[0m changed=0    unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "\u001b[0;32mtestbed-node-4\u001b[0m             : \u001b[0;32mok=1   \u001b[0m changed=0    unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "\u001b[0;32mtestbed-node-5\u001b[0m             : \u001b[0;32mok=1   \u001b[0m changed=0    unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Tuesday 24 February 2026  00:39:36 +0000 (0:00:11.472)       0:00:11.688 ****** ",
                                "=============================================================================== ",
                                "Wait until remote system is reachable ---------------------------------- 11.47s",
                                "+ osism apply hddtemp",
                                "\u001b[32m2026-02-24 00:39:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mPrepare task for execution of hddtemp.\u001b[0m",
                                "\u001b[32m2026-02-24 00:39:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3aa2872d-3529-407f-83f1-db28cd516501 (hddtemp) was prepared for execution.\u001b[0m",
                                "\u001b[32m2026-02-24 00:39:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task 3aa2872d-3529-407f-83f1-db28cd516501 (hddtemp) has been started and output is visible here.\u001b[0m",
                                "",
                                "PLAY [Apply role hddtemp] ******************************************************",
                                "",
                                "TASK [osism.services.hddtemp : Gather variables for each operating system] *****",
                                "Tuesday 24 February 2026  00:39:53 +0000 (0:00:00.273)       0:00:00.273 ****** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.services.hddtemp : Include distribution specific install tasks] ****",
                                "Tuesday 24 February 2026  00:39:54 +0000 (0:00:00.795)       0:00:01.069 ****** ",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/services/roles/hddtemp/tasks/install-Debian-family.yml for testbed-manager, testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m",
                                "",
                                "TASK [osism.services.hddtemp : Remove hddtemp package] *************************",
                                "Tuesday 24 February 2026  00:39:55 +0000 (0:00:01.217)       0:00:02.286 ****** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [osism.services.hddtemp : Enable Kernel Module drivetemp] *****************",
                                "Tuesday 24 February 2026  00:39:57 +0000 (0:00:02.025)       0:00:04.311 ****** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.services.hddtemp : Check if drivetemp module is available] *********",
                                "Tuesday 24 February 2026  00:39:59 +0000 (0:00:01.285)       0:00:05.597 ****** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.services.hddtemp : Load Kernel Module drivetemp] *******************",
                                "Tuesday 24 February 2026  00:40:00 +0000 (0:00:01.214)       0:00:06.812 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.services.hddtemp : Install lm-sensors] *****************************",
                                "Tuesday 24 February 2026  00:40:01 +0000 (0:00:00.869)       0:00:07.681 ****** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [osism.services.hddtemp : Include distribution specific service tasks] ****",
                                "Tuesday 24 February 2026  00:40:16 +0000 (0:00:14.768)       0:00:22.449 ****** ",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/services/roles/hddtemp/tasks/service-Debian-family.yml for testbed-manager, testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m",
                                "",
                                "TASK [osism.services.hddtemp : Manage lm-sensors service] **********************",
                                "Tuesday 24 February 2026  00:40:17 +0000 (0:00:01.273)       0:00:23.723 ****** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;33mtestbed-manager\u001b[0m            : \u001b[0;32mok=9   \u001b[0m \u001b[0;33mchanged=4   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=8   \u001b[0m \u001b[0;33mchanged=3   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=1   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=8   \u001b[0m \u001b[0;33mchanged=3   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=1   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=8   \u001b[0m \u001b[0;33mchanged=3   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=1   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-3\u001b[0m             : \u001b[0;32mok=8   \u001b[0m \u001b[0;33mchanged=3   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=1   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-4\u001b[0m             : \u001b[0;32mok=8   \u001b[0m \u001b[0;33mchanged=3   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=1   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-5\u001b[0m             : \u001b[0;32mok=8   \u001b[0m \u001b[0;33mchanged=3   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=1   \u001b[0m rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Tuesday 24 February 2026  00:40:19 +0000 (0:00:01.942)       0:00:25.666 ****** ",
                                "=============================================================================== ",
                                "osism.services.hddtemp : Install lm-sensors ---------------------------- 14.77s",
                                "osism.services.hddtemp : Remove hddtemp package ------------------------- 2.03s",
                                "osism.services.hddtemp : Manage lm-sensors service ---------------------- 1.94s",
                                "osism.services.hddtemp : Enable Kernel Module drivetemp ----------------- 1.29s",
                                "osism.services.hddtemp : Include distribution specific service tasks ---- 1.27s",
                                "osism.services.hddtemp : Include distribution specific install tasks ---- 1.22s",
                                "osism.services.hddtemp : Check if drivetemp module is available --------- 1.21s",
                                "osism.services.hddtemp : Load Kernel Module drivetemp ------------------- 0.87s",
                                "osism.services.hddtemp : Gather variables for each operating system ----- 0.80s",
                                "++ semver latest 7.1.1",
                                "+ [[ -1 -ge 0 ]]",
                                "+ [[ latest == \\l\\a\\t\\e\\s\\t ]]",
                                "+ sudo systemctl restart manager.service",
                                "+ [[ ceph-ansible == \\c\\e\\p\\h\\-\\a\\n\\s\\i\\b\\l\\e ]]",
                                "+ wait_for_container_healthy 60 ceph-ansible",
                                "+ local max_attempts=60",
                                "+ local name=ceph-ansible",
                                "+ local attempt_num=1",
                                "++ /usr/bin/docker inspect -f '{{.State.Health.Status}}' ceph-ansible",
                                "+ [[ unhealthy == \\h\\e\\a\\l\\t\\h\\y ]]",
                                "+ ((  attempt_num++ == max_attempts  ))",
                                "+ sleep 5",
                                "++ /usr/bin/docker inspect -f '{{.State.Health.Status}}' ceph-ansible",
                                "+ [[ unhealthy == \\h\\e\\a\\l\\t\\h\\y ]]",
                                "+ ((  attempt_num++ == max_attempts  ))",
                                "+ sleep 5",
                                "++ /usr/bin/docker inspect -f '{{.State.Health.Status}}' ceph-ansible",
                                "+ [[ unhealthy == \\h\\e\\a\\l\\t\\h\\y ]]",
                                "+ ((  attempt_num++ == max_attempts  ))",
                                "+ sleep 5",
                                "++ /usr/bin/docker inspect -f '{{.State.Health.Status}}' ceph-ansible",
                                "+ [[ unhealthy == \\h\\e\\a\\l\\t\\h\\y ]]",
                                "+ ((  attempt_num++ == max_attempts  ))",
                                "+ sleep 5",
                                "++ /usr/bin/docker inspect -f '{{.State.Health.Status}}' ceph-ansible",
                                "+ [[ unhealthy == \\h\\e\\a\\l\\t\\h\\y ]]",
                                "+ ((  attempt_num++ == max_attempts  ))",
                                "+ sleep 5",
                                "++ /usr/bin/docker inspect -f '{{.State.Health.Status}}' ceph-ansible",
                                "+ [[ unhealthy == \\h\\e\\a\\l\\t\\h\\y ]]",
                                "+ ((  attempt_num++ == max_attempts  ))",
                                "+ sleep 5",
                                "++ /usr/bin/docker inspect -f '{{.State.Health.Status}}' ceph-ansible",
                                "+ [[ unhealthy == \\h\\e\\a\\l\\t\\h\\y ]]",
                                "+ ((  attempt_num++ == max_attempts  ))",
                                "+ sleep 5",
                                "++ /usr/bin/docker inspect -f '{{.State.Health.Status}}' ceph-ansible",
                                "+ [[ starting == \\h\\e\\a\\l\\t\\h\\y ]]",
                                "+ ((  attempt_num++ == max_attempts  ))",
                                "+ sleep 5",
                                "++ /usr/bin/docker inspect -f '{{.State.Health.Status}}' ceph-ansible",
                                "+ [[ starting == \\h\\e\\a\\l\\t\\h\\y ]]",
                                "+ ((  attempt_num++ == max_attempts  ))",
                                "+ sleep 5",
                                "++ /usr/bin/docker inspect -f '{{.State.Health.Status}}' ceph-ansible",
                                "+ [[ starting == \\h\\e\\a\\l\\t\\h\\y ]]",
                                "+ ((  attempt_num++ == max_attempts  ))",
                                "+ sleep 5",
                                "++ /usr/bin/docker inspect -f '{{.State.Health.Status}}' ceph-ansible",
                                "+ [[ starting == \\h\\e\\a\\l\\t\\h\\y ]]",
                                "+ ((  attempt_num++ == max_attempts  ))",
                                "+ sleep 5",
                                "++ /usr/bin/docker inspect -f '{{.State.Health.Status}}' ceph-ansible",
                                "+ [[ starting == \\h\\e\\a\\l\\t\\h\\y ]]",
                                "+ ((  attempt_num++ == max_attempts  ))",
                                "+ sleep 5",
                                "++ /usr/bin/docker inspect -f '{{.State.Health.Status}}' ceph-ansible",
                                "+ [[ starting == \\h\\e\\a\\l\\t\\h\\y ]]",
                                "+ ((  attempt_num++ == max_attempts  ))",
                                "+ sleep 5",
                                "++ /usr/bin/docker inspect -f '{{.State.Health.Status}}' ceph-ansible",
                                "+ [[ healthy == \\h\\e\\a\\l\\t\\h\\y ]]",
                                "+ wait_for_container_healthy 60 kolla-ansible",
                                "+ local max_attempts=60",
                                "+ local name=kolla-ansible",
                                "+ local attempt_num=1",
                                "++ /usr/bin/docker inspect -f '{{.State.Health.Status}}' kolla-ansible",
                                "+ [[ healthy == \\h\\e\\a\\l\\t\\h\\y ]]",
                                "+ wait_for_container_healthy 60 osism-ansible",
                                "+ local max_attempts=60",
                                "+ local name=osism-ansible",
                                "+ local attempt_num=1",
                                "++ /usr/bin/docker inspect -f '{{.State.Health.Status}}' osism-ansible",
                                "+ [[ healthy == \\h\\e\\a\\l\\t\\h\\y ]]",
                                "+ [[ true == \\t\\r\\u\\e ]]",
                                "+ sh -c /opt/configuration/scripts/disable-ara.sh",
                                "ARA in ceph-ansible already disabled.",
                                "ARA in kolla-ansible already disabled.",
                                "ARA in osism-ansible already disabled.",
                                "ARA in osism-kubernetes already disabled.",
                                "+ osism apply gather-facts",
                                "\u001b[32m2026-02-24 00:41:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mPrepare task for execution of gather-facts.\u001b[0m",
                                "\u001b[32m2026-02-24 00:41:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 09f896ba-61bc-4bff-955b-0f3f7aeaadc4 (gather-facts) was prepared for execution.\u001b[0m",
                                "\u001b[32m2026-02-24 00:41:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task 09f896ba-61bc-4bff-955b-0f3f7aeaadc4 (gather-facts) has been started and output is visible here.\u001b[0m",
                                "",
                                "PLAY [Gather facts for all hosts] **********************************************",
                                "",
                                "TASK [Gathers facts about hosts] ***********************************************",
                                "Tuesday 24 February 2026  00:41:56 +0000 (0:00:00.197)       0:00:00.197 ****** ",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "PLAY [Gather facts for all hosts if using --limit] *****************************",
                                "",
                                "TASK [Gather facts for all hosts] **********************************************",
                                "Tuesday 24 February 2026  00:42:04 +0000 (0:00:08.546)       0:00:08.743 ****** ",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;32mtestbed-manager\u001b[0m            : \u001b[0;32mok=1   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=1   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;32mtestbed-node-0\u001b[0m             : \u001b[0;32mok=1   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=1   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;32mtestbed-node-1\u001b[0m             : \u001b[0;32mok=1   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=1   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;32mtestbed-node-2\u001b[0m             : \u001b[0;32mok=1   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=1   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;32mtestbed-node-3\u001b[0m             : \u001b[0;32mok=1   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=1   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;32mtestbed-node-4\u001b[0m             : \u001b[0;32mok=1   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=1   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;32mtestbed-node-5\u001b[0m             : \u001b[0;32mok=1   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=1   \u001b[0m rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Tuesday 24 February 2026  00:42:05 +0000 (0:00:00.444)       0:00:09.187 ****** ",
                                "=============================================================================== ",
                                "Gathers facts about hosts ----------------------------------------------- 8.55s",
                                "Gather facts for all hosts ---------------------------------------------- 0.44s",
                                "+ sudo ln -sf /opt/configuration/scripts/deploy/001-helpers.sh /usr/local/bin/deploy-helper",
                                "+ sudo ln -sf /opt/configuration/scripts/deploy/100-ceph-with-ansible.sh /usr/local/bin/deploy-ceph-with-ansible",
                                "+ sudo ln -sf /opt/configuration/scripts/deploy/100-ceph-with-rook.sh /usr/local/bin/deploy-ceph-with-rook",
                                "+ sudo ln -sf /opt/configuration/scripts/deploy/200-infrastructure.sh /usr/local/bin/deploy-infrastructure",
                                "+ sudo ln -sf /opt/configuration/scripts/deploy/300-openstack.sh /usr/local/bin/deploy-openstack",
                                "+ sudo ln -sf /opt/configuration/scripts/deploy/320-openstack-minimal.sh /usr/local/bin/deploy-openstack-minimal",
                                "+ sudo ln -sf /opt/configuration/scripts/deploy/400-monitoring.sh /usr/local/bin/deploy-monitoring",
                                "+ sudo ln -sf /opt/configuration/scripts/deploy/500-kubernetes.sh /usr/local/bin/deploy-kubernetes",
                                "+ sudo ln -sf /opt/configuration/scripts/deploy/510-clusterapi.sh /usr/local/bin/deploy-kubernetes-clusterapi",
                                "+ sudo ln -sf /opt/configuration/scripts/upgrade-manager.sh /usr/local/bin/upgrade-manager",
                                "+ sudo ln -sf /opt/configuration/scripts/upgrade/100-ceph-with-ansible.sh /usr/local/bin/upgrade-ceph-with-ansible",
                                "+ sudo ln -sf /opt/configuration/scripts/upgrade/100-ceph-with-rook.sh /usr/local/bin/upgrade-ceph-with-rook",
                                "+ sudo ln -sf /opt/configuration/scripts/upgrade/200-infrastructure.sh /usr/local/bin/upgrade-infrastructure",
                                "+ sudo ln -sf /opt/configuration/scripts/upgrade/300-openstack.sh /usr/local/bin/upgrade-openstack",
                                "+ sudo ln -sf /opt/configuration/scripts/upgrade/320-openstack-minimal.sh /usr/local/bin/upgrade-openstack-minimal",
                                "+ sudo ln -sf /opt/configuration/scripts/upgrade/400-monitoring.sh /usr/local/bin/upgrade-monitoring",
                                "+ sudo ln -sf /opt/configuration/scripts/upgrade/500-kubernetes.sh /usr/local/bin/upgrade-kubernetes",
                                "+ sudo ln -sf /opt/configuration/scripts/upgrade/510-clusterapi.sh /usr/local/bin/upgrade-kubernetes-clusterapi",
                                "+ sudo ln -sf /opt/configuration/scripts/bootstrap/300-openstack.sh /usr/local/bin/bootstrap-openstack",
                                "+ sudo ln -sf /opt/configuration/scripts/bootstrap/301-openstack-octavia-amhpora-image.sh /usr/local/bin/bootstrap-octavia",
                                "+ sudo ln -sf /opt/configuration/scripts/bootstrap/302-openstack-k8s-clusterapi-images.sh /usr/local/bin/bootstrap-clusterapi",
                                "+ sudo ln -sf /opt/configuration/scripts/disable-local-registry.sh /usr/local/bin/disable-local-registry",
                                "+ sudo ln -sf /opt/configuration/scripts/pull-images.sh /usr/local/bin/pull-images",
                                "+ [[ false == \\t\\r\\u\\e ]]"
                            ],
                            "zuul_log_id": "0242ac1f-6606-1b2b-a63a-00000000001a-1-orchestrator"
                        }
                    },
                    "task": {
                        "duration": {
                            "end": "2026-02-24T00:42:06.191138Z",
                            "start": "2026-02-24T00:17:17.109578Z"
                        },
                        "id": "0242ac1f-6606-1b2b-a63a-00000000001a",
                        "name": "Deploy manager + bootstrap nodes"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "ansible.builtin.command",
                            "changed": false,
                            "false_condition": "not _nutshell | bool",
                            "skip_reason": "Conditional result was False",
                            "skipped": true
                        }
                    },
                    "task": {
                        "duration": {
                            "end": "2026-02-24T00:42:06.801759Z",
                            "start": "2026-02-24T00:42:06.292090Z"
                        },
                        "id": "0242ac1f-6606-1b2b-a63a-00000000001b",
                        "name": "Deploy services"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "ansible.builtin.command",
                            "changed": true,
                            "cmd": [
                                "ssh",
                                "-i",
                                "/home/zuul-testbed02/src/github.com/osism/testbed/terraform/.id_rsa.ci",
                                "dragon@81.163.192.195",
                                "/opt/configuration/scripts/deploy-in-a-nutshell.sh"
                            ],
                            "delta": "0:36:40.101308",
                            "end": "2026-02-24 01:18:47.241058",
                            "invocation": {
                                "module_args": {
                                    "_raw_params": "ssh -i /home/zuul-testbed02/src/github.com/osism/testbed/terraform/.id_rsa.ci dragon@81.163.192.195 /opt/configuration/scripts/deploy-in-a-nutshell.sh",
                                    "_uses_shell": false,
                                    "argv": null,
                                    "chdir": null,
                                    "creates": null,
                                    "executable": null,
                                    "expand_argument_vars": true,
                                    "removes": null,
                                    "stdin": null,
                                    "stdin_add_newline": true,
                                    "strip_empty_ends": true,
                                    "zuul_ansible_split_streams": false,
                                    "zuul_log_id": "0242ac1f-6606-1b2b-a63a-00000000001c-1-orchestrator",
                                    "zuul_output_max_bytes": 1073741824
                                }
                            },
                            "msg": "",
                            "rc": 0,
                            "start": "2026-02-24 00:42:07.139750",
                            "stderr": "",
                            "stderr_lines": [],
                            "stdout": "+ set -e\n+ source /opt/configuration/scripts/include.sh\n++ export INTERACTIVE=false\n\n# PULL IMAGES\n\n++ INTERACTIVE=false\n++ export OSISM_APPLY_RETRY=1\n++ OSISM_APPLY_RETRY=1\n+ source /opt/manager-vars.sh\n++ export NUMBER_OF_NODES=6\n++ NUMBER_OF_NODES=6\n++ export CEPH_VERSION=reef\n++ CEPH_VERSION=reef\n++ export CONFIGURATION_VERSION=main\n++ CONFIGURATION_VERSION=main\n++ export MANAGER_VERSION=latest\n++ MANAGER_VERSION=latest\n++ export OPENSTACK_VERSION=2024.2\n++ OPENSTACK_VERSION=2024.2\n++ export ARA=false\n++ ARA=false\n++ export DEPLOY_MODE=manager\n++ DEPLOY_MODE=manager\n++ export TEMPEST=true\n++ TEMPEST=true\n++ export IS_ZUUL=true\n++ IS_ZUUL=true\n++ export MANAGER_PUBLIC_IP_ADDRESS=81.163.192.195\n++ MANAGER_PUBLIC_IP_ADDRESS=81.163.192.195\n++ export EXTERNAL_API=false\n++ EXTERNAL_API=false\n++ export IMAGE_USER=ubuntu\n++ IMAGE_USER=ubuntu\n++ export IMAGE_NODE_USER=ubuntu\n++ IMAGE_NODE_USER=ubuntu\n++ export CEPH_STACK=ceph-ansible\n++ CEPH_STACK=ceph-ansible\n+ echo\n+ echo '# PULL IMAGES'\n+ echo\n++ semver latest 7.0.0\n+ [[ -1 -ge 0 ]]\n+ [[ latest == \\l\\a\\t\\e\\s\\t ]]\n+ osism apply --no-wait -r 2 -e custom pull-images\n\u001b[32m2026-02-24 00:42:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTrying to run play pull-images in environment custom\u001b[0m\n\u001b[32m2026-02-24 00:42:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mPrepare task for execution of pull-images.\u001b[0m\n\u001b[32m2026-02-24 00:42:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 0b17a71b-4591-4989-ae1c-e5307c45d989 (pull-images) was prepared for execution.\u001b[0m\n\u001b[32m2026-02-24 00:42:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 0b17a71b-4591-4989-ae1c-e5307c45d989 is running in background. No more output. Check ARA for logs.\u001b[0m\n\u001b[32m2026-02-24 00:42:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTrying to run play wipe-partitions in environment custom\u001b[0m\n\u001b[32m2026-02-24 00:42:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mPrepare task for execution of wipe-partitions.\u001b[0m\n\u001b[32m2026-02-24 00:42:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 94d06f06-2939-4095-b22c-bd35654caca4 (wipe-partitions) was prepared for execution.\u001b[0m\n\u001b[32m2026-02-24 00:42:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task 94d06f06-2939-4095-b22c-bd35654caca4 (wipe-partitions) has been started and output is visible here.\u001b[0m\n\nPLAY [Wipe partitions] *********************************************************\n\nTASK [Find all logical devices owned by UID 167] *******************************\nTuesday 24 February 2026  00:42:35 +0000 (0:00:00.120)       0:00:00.120 ****** \n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nTASK [Remove all rook related logical devices] *********************************\nTuesday 24 February 2026  00:42:35 +0000 (0:00:00.598)       0:00:00.719 ****** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Find all logical devices with prefix ceph] *******************************\nTuesday 24 February 2026  00:42:36 +0000 (0:00:00.343)       0:00:01.063 ****** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\nTASK [Remove all ceph related logical devices] *********************************\nTuesday 24 February 2026  00:42:36 +0000 (0:00:00.557)       0:00:01.620 ****** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Check device availability] ***********************************************\nTuesday 24 February 2026  00:42:37 +0000 (0:00:00.250)       0:00:01.871 ****** \n\u001b[0;33mchanged: [testbed-node-3] => (item=/dev/sdb)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=/dev/sdb)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=/dev/sdb)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item=/dev/sdc)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=/dev/sdc)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=/dev/sdc)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item=/dev/sdd)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=/dev/sdd)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=/dev/sdd)\u001b[0m\n\nTASK [Wipe partitions with wipefs] *********************************************\nTuesday 24 February 2026  00:42:38 +0000 (0:00:01.290)       0:00:03.162 ****** \n\u001b[0;32mok: [testbed-node-3] => (item=/dev/sdb)\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item=/dev/sdb)\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item=/dev/sdb)\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item=/dev/sdc)\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item=/dev/sdc)\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item=/dev/sdc)\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item=/dev/sdd)\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item=/dev/sdd)\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item=/dev/sdd)\u001b[0m\n\nTASK [Overwrite first 32M with zeros] ******************************************\nTuesday 24 February 2026  00:42:39 +0000 (0:00:01.499)       0:00:04.661 ****** \n\u001b[0;33mchanged: [testbed-node-4] => (item=/dev/sdb)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item=/dev/sdb)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=/dev/sdb)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=/dev/sdc)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item=/dev/sdc)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=/dev/sdc)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=/dev/sdd)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item=/dev/sdd)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=/dev/sdd)\u001b[0m\n\nTASK [Reload udev rules] *******************************************************\nTuesday 24 February 2026  00:42:41 +0000 (0:00:02.088)       0:00:06.750 ****** \n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nTASK [Request device events from the kernel] ***********************************\nTuesday 24 February 2026  00:42:42 +0000 (0:00:00.643)       0:00:07.393 ****** \n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;33mtestbed-node-3\u001b[0m             : \u001b[0;32mok=7   \u001b[0m \u001b[0;33mchanged=5   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-4\u001b[0m             : \u001b[0;32mok=7   \u001b[0m \u001b[0;33mchanged=5   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-5\u001b[0m             : \u001b[0;32mok=7   \u001b[0m \u001b[0;33mchanged=5   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nTuesday 24 February 2026  00:42:43 +0000 (0:00:00.693)       0:00:08.087 ****** \n=============================================================================== \nOverwrite first 32M with zeros ------------------------------------------ 2.09s\nWipe partitions with wipefs --------------------------------------------- 1.50s\nCheck device availability ----------------------------------------------- 1.29s\nRequest device events from the kernel ----------------------------------- 0.69s\nReload udev rules ------------------------------------------------------- 0.64s\nFind all logical devices owned by UID 167 ------------------------------- 0.60s\nFind all logical devices with prefix ceph ------------------------------- 0.56s\nRemove all rook related logical devices --------------------------------- 0.34s\nRemove all ceph related logical devices --------------------------------- 0.25s\n\u001b[32m2026-02-24 00:42:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mPrepare task for execution of facts.\u001b[0m\n\u001b[32m2026-02-24 00:42:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3ce96cb6-2139-4792-bb10-c5cbb803795e (facts) was prepared for execution.\u001b[0m\n\u001b[32m2026-02-24 00:42:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task 3ce96cb6-2139-4792-bb10-c5cbb803795e (facts) has been started and output is visible here.\u001b[0m\n\nPLAY [Apply role facts] ********************************************************\n\nTASK [osism.commons.facts : Create custom facts directory] *********************\nTuesday 24 February 2026  00:43:00 +0000 (0:00:00.263)       0:00:00.263 ****** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [osism.commons.facts : Copy fact files] ***********************************\nTuesday 24 February 2026  00:43:01 +0000 (0:00:01.079)       0:00:01.342 ****** \n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nPLAY [Gather facts for all hosts] **********************************************\n\nTASK [Gathers facts about hosts] ***********************************************\nTuesday 24 February 2026  00:43:03 +0000 (0:00:01.231)       0:00:02.574 ****** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nPLAY [Gather facts for all hosts if using --limit] *****************************\n\nTASK [Gather facts for all hosts] **********************************************\nTuesday 24 February 2026  00:43:08 +0000 (0:00:04.968)       0:00:07.542 ****** \n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;32mtestbed-manager\u001b[0m            : \u001b[0;32mok=2   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   \n\u001b[0;32mtestbed-node-0\u001b[0m             : \u001b[0;32mok=2   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   \n\u001b[0;32mtestbed-node-1\u001b[0m             : \u001b[0;32mok=2   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   \n\u001b[0;32mtestbed-node-2\u001b[0m             : \u001b[0;32mok=2   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   \n\u001b[0;32mtestbed-node-3\u001b[0m             : \u001b[0;32mok=2   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   \n\u001b[0;32mtestbed-node-4\u001b[0m             : \u001b[0;32mok=2   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   \n\u001b[0;32mtestbed-node-5\u001b[0m             : \u001b[0;32mok=2   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nTuesday 24 February 2026  00:43:08 +0000 (0:00:00.521)       0:00:08.064 ****** \n=============================================================================== \nGathers facts about hosts ----------------------------------------------- 4.97s\nosism.commons.facts : Copy fact files ----------------------------------- 1.23s\nosism.commons.facts : Create custom facts directory --------------------- 1.08s\nGather facts for all hosts ---------------------------------------------- 0.52s\n\u001b[32m2026-02-24 00:43:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mPrepare task for execution of ceph-configure-lvm-volumes.\u001b[0m\n\u001b[32m2026-02-24 00:43:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 567e2c8a-8f76-4540-a3d2-6eaea431f2ce (ceph-configure-lvm-volumes) was prepared for execution.\u001b[0m\n\u001b[32m2026-02-24 00:43:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task 567e2c8a-8f76-4540-a3d2-6eaea431f2ce (ceph-configure-lvm-volumes) has been started and output is visible here.\u001b[0m\n\u001b[1;35m[WARNING]: Collection community.general does not support Ansible version\u001b[0m\n\u001b[1;35m2.16.14\u001b[0m\n\nPLAY [Ceph configure LVM] ******************************************************\n\nTASK [Get extra vars for Ceph configuration] ***********************************\nTuesday 24 February 2026  00:43:15 +0000 (0:00:00.240)       0:00:00.240 ****** \n\u001b[0;32mok: [testbed-node-3 -> testbed-manager(192.168.16.5)]\u001b[0m\n\nTASK [Get initial list of available block devices] *****************************\nTuesday 24 February 2026  00:43:15 +0000 (0:00:00.208)       0:00:00.448 ****** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nTuesday 24 February 2026  00:43:15 +0000 (0:00:00.197)       0:00:00.646 ****** \n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-3 => (item=loop0)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-3 => (item=loop1)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-3 => (item=loop2)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-3 => (item=loop3)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-3 => (item=loop4)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-3 => (item=loop5)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-3 => (item=loop6)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-3 => (item=loop7)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-3 => (item=sda)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-3 => (item=sdb)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-3 => (item=sdc)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-3 => (item=sdd)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-3 => (item=sr0)\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nTuesday 24 February 2026  00:43:16 +0000 (0:00:00.403)       0:00:01.050 ****** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nTuesday 24 February 2026  00:43:16 +0000 (0:00:00.191)       0:00:01.241 ****** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nTuesday 24 February 2026  00:43:16 +0000 (0:00:00.156)       0:00:01.397 ****** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nTuesday 24 February 2026  00:43:16 +0000 (0:00:00.185)       0:00:01.583 ****** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nTuesday 24 February 2026  00:43:17 +0000 (0:00:00.180)       0:00:01.764 ****** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nTuesday 24 February 2026  00:43:17 +0000 (0:00:00.186)       0:00:01.950 ****** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nTuesday 24 February 2026  00:43:17 +0000 (0:00:00.194)       0:00:02.145 ****** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nTuesday 24 February 2026  00:43:17 +0000 (0:00:00.176)       0:00:02.321 ****** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nTuesday 24 February 2026  00:43:17 +0000 (0:00:00.242)       0:00:02.564 ****** \n\u001b[0;32mok: [testbed-node-3] => (item=scsi-0QEMU_QEMU_HARDDISK_f90a4e80-1f36-4d14-b7da-de7b20e9fbca)\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item=scsi-SQEMU_QEMU_HARDDISK_f90a4e80-1f36-4d14-b7da-de7b20e9fbca)\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nTuesday 24 February 2026  00:43:18 +0000 (0:00:00.371)       0:00:02.936 ****** \n\u001b[0;32mok: [testbed-node-3] => (item=scsi-0QEMU_QEMU_HARDDISK_9c6b6763-9704-41ec-9b6d-97eb80a5145c)\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item=scsi-SQEMU_QEMU_HARDDISK_9c6b6763-9704-41ec-9b6d-97eb80a5145c)\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nTuesday 24 February 2026  00:43:18 +0000 (0:00:00.576)       0:00:03.512 ****** \n\u001b[0;32mok: [testbed-node-3] => (item=scsi-0QEMU_QEMU_HARDDISK_6644f191-69e9-4f3a-baff-cffe4152595b)\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item=scsi-SQEMU_QEMU_HARDDISK_6644f191-69e9-4f3a-baff-cffe4152595b)\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nTuesday 24 February 2026  00:43:19 +0000 (0:00:00.630)       0:00:04.143 ****** \n\u001b[0;32mok: [testbed-node-3] => (item=scsi-0QEMU_QEMU_HARDDISK_4534650b-23a0-46c8-8bf1-44520caee637)\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item=scsi-SQEMU_QEMU_HARDDISK_4534650b-23a0-46c8-8bf1-44520caee637)\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nTuesday 24 February 2026  00:43:20 +0000 (0:00:00.899)       0:00:05.042 ****** \n\u001b[0;32mok: [testbed-node-3] => (item=ata-QEMU_DVD-ROM_QM00001)\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nTuesday 24 February 2026  00:43:20 +0000 (0:00:00.330)       0:00:05.372 ****** \n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-3 => (item=loop0)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-3 => (item=loop1)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-3 => (item=loop2)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-3 => (item=loop3)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-3 => (item=loop4)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-3 => (item=loop5)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-3 => (item=loop6)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-3 => (item=loop7)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-3 => (item=sda)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-3 => (item=sdb)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-3 => (item=sdc)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-3 => (item=sdd)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-3 => (item=sr0)\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nTuesday 24 February 2026  00:43:20 +0000 (0:00:00.377)       0:00:05.750 ****** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nTuesday 24 February 2026  00:43:21 +0000 (0:00:00.237)       0:00:05.988 ****** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nTuesday 24 February 2026  00:43:21 +0000 (0:00:00.206)       0:00:06.195 ****** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nTuesday 24 February 2026  00:43:21 +0000 (0:00:00.214)       0:00:06.409 ****** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nTuesday 24 February 2026  00:43:21 +0000 (0:00:00.197)       0:00:06.606 ****** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nTuesday 24 February 2026  00:43:22 +0000 (0:00:00.209)       0:00:06.816 ****** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nTuesday 24 February 2026  00:43:22 +0000 (0:00:00.239)       0:00:07.056 ****** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nTuesday 24 February 2026  00:43:22 +0000 (0:00:00.193)       0:00:07.249 ****** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nTuesday 24 February 2026  00:43:22 +0000 (0:00:00.196)       0:00:07.446 ****** \n\u001b[0;32mok: [testbed-node-3] => (item=sda1)\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item=sda14)\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item=sda15)\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item=sda16)\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nTuesday 24 February 2026  00:43:23 +0000 (0:00:01.088)       0:00:08.534 ****** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nTuesday 24 February 2026  00:43:23 +0000 (0:00:00.209)       0:00:08.744 ****** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nTuesday 24 February 2026  00:43:24 +0000 (0:00:00.229)       0:00:08.974 ****** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nTuesday 24 February 2026  00:43:24 +0000 (0:00:00.205)       0:00:09.180 ****** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Set UUIDs for OSD VGs/LVs] ***********************************************\nTuesday 24 February 2026  00:43:24 +0000 (0:00:00.203)       0:00:09.384 ****** \n\u001b[0;32mok: [testbed-node-3] => (item={'key': 'sdb', 'value': None})\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item={'key': 'sdc', 'value': None})\u001b[0m\n\nTASK [Generate WAL VG names] ***************************************************\nTuesday 24 February 2026  00:43:24 +0000 (0:00:00.222)       0:00:09.606 ****** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Generate DB VG names] ****************************************************\nTuesday 24 February 2026  00:43:24 +0000 (0:00:00.144)       0:00:09.751 ****** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Generate shared DB/WAL VG names] *****************************************\nTuesday 24 February 2026  00:43:25 +0000 (0:00:00.156)       0:00:09.907 ****** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Define lvm_volumes structures] *******************************************\nTuesday 24 February 2026  00:43:25 +0000 (0:00:00.160)       0:00:10.068 ****** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\nTASK [Generate lvm_volumes structure (block only)] *****************************\nTuesday 24 February 2026  00:43:25 +0000 (0:00:00.125)       0:00:10.193 ****** \n\u001b[0;32mok: [testbed-node-3] => (item={'key': 'sdb', 'value': {'osd_lvm_uuid': '9b92f24c-dac1-5bff-a390-34d9ff9e0f4f'}})\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item={'key': 'sdc', 'value': {'osd_lvm_uuid': '7f6a8fe5-6adc-52b4-b256-edcdbfd9bf4e'}})\u001b[0m\n\nTASK [Generate lvm_volumes structure (block + db)] *****************************\nTuesday 24 February 2026  00:43:25 +0000 (0:00:00.162)       0:00:10.356 ****** \n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'sdb', 'value': {'osd_lvm_uuid': '9b92f24c-dac1-5bff-a390-34d9ff9e0f4f'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'sdc', 'value': {'osd_lvm_uuid': '7f6a8fe5-6adc-52b4-b256-edcdbfd9bf4e'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Generate lvm_volumes structure (block + wal)] ****************************\nTuesday 24 February 2026  00:43:25 +0000 (0:00:00.157)       0:00:10.513 ****** \n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'sdb', 'value': {'osd_lvm_uuid': '9b92f24c-dac1-5bff-a390-34d9ff9e0f4f'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'sdc', 'value': {'osd_lvm_uuid': '7f6a8fe5-6adc-52b4-b256-edcdbfd9bf4e'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Generate lvm_volumes structure (block + db + wal)] ***********************\nTuesday 24 February 2026  00:43:26 +0000 (0:00:00.386)       0:00:10.899 ****** \n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'sdb', 'value': {'osd_lvm_uuid': '9b92f24c-dac1-5bff-a390-34d9ff9e0f4f'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'sdc', 'value': {'osd_lvm_uuid': '7f6a8fe5-6adc-52b4-b256-edcdbfd9bf4e'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Compile lvm_volumes] *****************************************************\nTuesday 24 February 2026  00:43:26 +0000 (0:00:00.151)       0:00:11.051 ****** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\nTASK [Set OSD devices config data] *********************************************\nTuesday 24 February 2026  00:43:26 +0000 (0:00:00.142)       0:00:11.194 ****** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\nTASK [Set DB devices config data] **********************************************\nTuesday 24 February 2026  00:43:26 +0000 (0:00:00.119)       0:00:11.313 ****** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Set WAL devices config data] *********************************************\nTuesday 24 February 2026  00:43:26 +0000 (0:00:00.113)       0:00:11.427 ****** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Set DB+WAL devices config data] ******************************************\nTuesday 24 February 2026  00:43:26 +0000 (0:00:00.141)       0:00:11.569 ****** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Print ceph_osd_devices] **************************************************\nTuesday 24 February 2026  00:43:26 +0000 (0:00:00.134)       0:00:11.703 ****** \n\u001b[0;32mok: [testbed-node-3] => {\u001b[0m\n\u001b[0;32m    \"ceph_osd_devices\": {\u001b[0m\n\u001b[0;32m        \"sdb\": {\u001b[0m\n\u001b[0;32m            \"osd_lvm_uuid\": \"9b92f24c-dac1-5bff-a390-34d9ff9e0f4f\"\u001b[0m\n\u001b[0;32m        },\u001b[0m\n\u001b[0;32m        \"sdc\": {\u001b[0m\n\u001b[0;32m            \"osd_lvm_uuid\": \"7f6a8fe5-6adc-52b4-b256-edcdbfd9bf4e\"\u001b[0m\n\u001b[0;32m        }\u001b[0m\n\u001b[0;32m    }\u001b[0m\n\u001b[0;32m}\u001b[0m\n\nTASK [Print WAL devices] *******************************************************\nTuesday 24 February 2026  00:43:27 +0000 (0:00:00.153)       0:00:11.857 ****** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Print DB devices] ********************************************************\nTuesday 24 February 2026  00:43:27 +0000 (0:00:00.138)       0:00:11.995 ****** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Print shared DB/WAL devices] *********************************************\nTuesday 24 February 2026  00:43:27 +0000 (0:00:00.130)       0:00:12.126 ****** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Print configuration data] ************************************************\nTuesday 24 February 2026  00:43:27 +0000 (0:00:00.144)       0:00:12.270 ****** \n\u001b[0;33mchanged: [testbed-node-3] => {\u001b[0m\n\u001b[0;33m    \"_ceph_configure_lvm_config_data\": {\u001b[0m\n\u001b[0;33m        \"ceph_osd_devices\": {\u001b[0m\n\u001b[0;33m            \"sdb\": {\u001b[0m\n\u001b[0;33m                \"osd_lvm_uuid\": \"9b92f24c-dac1-5bff-a390-34d9ff9e0f4f\"\u001b[0m\n\u001b[0;33m            },\u001b[0m\n\u001b[0;33m            \"sdc\": {\u001b[0m\n\u001b[0;33m                \"osd_lvm_uuid\": \"7f6a8fe5-6adc-52b4-b256-edcdbfd9bf4e\"\u001b[0m\n\u001b[0;33m            }\u001b[0m\n\u001b[0;33m        },\u001b[0m\n\u001b[0;33m        \"lvm_volumes\": [\u001b[0m\n\u001b[0;33m            {\u001b[0m\n\u001b[0;33m                \"data\": \"osd-block-9b92f24c-dac1-5bff-a390-34d9ff9e0f4f\",\u001b[0m\n\u001b[0;33m                \"data_vg\": \"ceph-9b92f24c-dac1-5bff-a390-34d9ff9e0f4f\"\u001b[0m\n\u001b[0;33m            },\u001b[0m\n\u001b[0;33m            {\u001b[0m\n\u001b[0;33m                \"data\": \"osd-block-7f6a8fe5-6adc-52b4-b256-edcdbfd9bf4e\",\u001b[0m\n\u001b[0;33m                \"data_vg\": \"ceph-7f6a8fe5-6adc-52b4-b256-edcdbfd9bf4e\"\u001b[0m\n\u001b[0;33m            }\u001b[0m\n\u001b[0;33m        ]\u001b[0m\n\u001b[0;33m    }\u001b[0m\n\u001b[0;33m}\u001b[0m\n\nRUNNING HANDLER [Write configuration file] *************************************\nTuesday 24 February 2026  00:43:27 +0000 (0:00:00.416)       0:00:12.687 ****** \n\u001b[0;33mchanged: [testbed-node-3 -> testbed-manager(192.168.16.5)]\u001b[0m\n\nPLAY [Ceph configure LVM] ******************************************************\n\nTASK [Get extra vars for Ceph configuration] ***********************************\nTuesday 24 February 2026  00:43:29 +0000 (0:00:01.887)       0:00:14.574 ****** \n\u001b[0;32mok: [testbed-node-4 -> testbed-manager(192.168.16.5)]\u001b[0m\n\nTASK [Get initial list of available block devices] *****************************\nTuesday 24 February 2026  00:43:30 +0000 (0:00:00.276)       0:00:14.851 ****** \n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nTuesday 24 February 2026  00:43:30 +0000 (0:00:00.248)       0:00:15.100 ****** \n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-4 => (item=loop0)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-4 => (item=loop1)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-4 => (item=loop2)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-4 => (item=loop3)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-4 => (item=loop4)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-4 => (item=loop5)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-4 => (item=loop6)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-4 => (item=loop7)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-4 => (item=sda)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-4 => (item=sdb)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-4 => (item=sdc)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-4 => (item=sdd)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-4 => (item=sr0)\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nTuesday 24 February 2026  00:43:30 +0000 (0:00:00.395)       0:00:15.496 ****** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nTuesday 24 February 2026  00:43:30 +0000 (0:00:00.231)       0:00:15.727 ****** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nTuesday 24 February 2026  00:43:31 +0000 (0:00:00.210)       0:00:15.937 ****** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nTuesday 24 February 2026  00:43:31 +0000 (0:00:00.199)       0:00:16.136 ****** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nTuesday 24 February 2026  00:43:31 +0000 (0:00:00.249)       0:00:16.385 ****** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nTuesday 24 February 2026  00:43:32 +0000 (0:00:00.693)       0:00:17.079 ****** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nTuesday 24 February 2026  00:43:32 +0000 (0:00:00.211)       0:00:17.290 ****** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nTuesday 24 February 2026  00:43:32 +0000 (0:00:00.222)       0:00:17.513 ****** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nTuesday 24 February 2026  00:43:33 +0000 (0:00:00.261)       0:00:17.774 ****** \n\u001b[0;32mok: [testbed-node-4] => (item=scsi-0QEMU_QEMU_HARDDISK_47e600ec-05ca-44a6-93d0-5ca3104e3639)\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item=scsi-SQEMU_QEMU_HARDDISK_47e600ec-05ca-44a6-93d0-5ca3104e3639)\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nTuesday 24 February 2026  00:43:33 +0000 (0:00:00.517)       0:00:18.292 ****** \n\u001b[0;32mok: [testbed-node-4] => (item=scsi-0QEMU_QEMU_HARDDISK_f659b61f-c39f-426f-a640-6a753951661d)\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item=scsi-SQEMU_QEMU_HARDDISK_f659b61f-c39f-426f-a640-6a753951661d)\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nTuesday 24 February 2026  00:43:33 +0000 (0:00:00.447)       0:00:18.739 ****** \n\u001b[0;32mok: [testbed-node-4] => (item=scsi-0QEMU_QEMU_HARDDISK_fb58bf16-a1c6-4abb-925b-29c6a6a6f100)\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item=scsi-SQEMU_QEMU_HARDDISK_fb58bf16-a1c6-4abb-925b-29c6a6a6f100)\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nTuesday 24 February 2026  00:43:34 +0000 (0:00:00.459)       0:00:19.198 ****** \n\u001b[0;32mok: [testbed-node-4] => (item=scsi-0QEMU_QEMU_HARDDISK_5660bb62-e2c9-45b2-ae11-8ab0f829b194)\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item=scsi-SQEMU_QEMU_HARDDISK_5660bb62-e2c9-45b2-ae11-8ab0f829b194)\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nTuesday 24 February 2026  00:43:34 +0000 (0:00:00.445)       0:00:19.643 ****** \n\u001b[0;32mok: [testbed-node-4] => (item=ata-QEMU_DVD-ROM_QM00001)\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nTuesday 24 February 2026  00:43:35 +0000 (0:00:00.382)       0:00:20.025 ****** \n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-4 => (item=loop0)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-4 => (item=loop1)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-4 => (item=loop2)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-4 => (item=loop3)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-4 => (item=loop4)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-4 => (item=loop5)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-4 => (item=loop6)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-4 => (item=loop7)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-4 => (item=sda)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-4 => (item=sdb)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-4 => (item=sdc)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-4 => (item=sdd)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-4 => (item=sr0)\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nTuesday 24 February 2026  00:43:35 +0000 (0:00:00.492)       0:00:20.518 ****** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nTuesday 24 February 2026  00:43:36 +0000 (0:00:00.848)       0:00:21.366 ****** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nTuesday 24 February 2026  00:43:36 +0000 (0:00:00.212)       0:00:21.579 ****** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nTuesday 24 February 2026  00:43:37 +0000 (0:00:00.222)       0:00:21.801 ****** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nTuesday 24 February 2026  00:43:37 +0000 (0:00:00.208)       0:00:22.009 ****** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nTuesday 24 February 2026  00:43:37 +0000 (0:00:00.187)       0:00:22.197 ****** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nTuesday 24 February 2026  00:43:37 +0000 (0:00:00.197)       0:00:22.394 ****** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nTuesday 24 February 2026  00:43:37 +0000 (0:00:00.188)       0:00:22.583 ****** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nTuesday 24 February 2026  00:43:38 +0000 (0:00:00.193)       0:00:22.776 ****** \n\u001b[0;32mok: [testbed-node-4] => (item=sda1)\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item=sda14)\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item=sda15)\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item=sda16)\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nTuesday 24 February 2026  00:43:38 +0000 (0:00:00.902)       0:00:23.679 ****** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nTuesday 24 February 2026  00:43:39 +0000 (0:00:00.188)       0:00:23.867 ****** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nTuesday 24 February 2026  00:43:39 +0000 (0:00:00.203)       0:00:24.070 ****** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nTuesday 24 February 2026  00:43:39 +0000 (0:00:00.190)       0:00:24.261 ****** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Set UUIDs for OSD VGs/LVs] ***********************************************\nTuesday 24 February 2026  00:43:40 +0000 (0:00:00.742)       0:00:25.003 ****** \n\u001b[0;32mok: [testbed-node-4] => (item={'key': 'sdb', 'value': None})\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item={'key': 'sdc', 'value': None})\u001b[0m\n\nTASK [Generate WAL VG names] ***************************************************\nTuesday 24 February 2026  00:43:40 +0000 (0:00:00.175)       0:00:25.179 ****** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Generate DB VG names] ****************************************************\nTuesday 24 February 2026  00:43:40 +0000 (0:00:00.149)       0:00:25.328 ****** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Generate shared DB/WAL VG names] *****************************************\nTuesday 24 February 2026  00:43:40 +0000 (0:00:00.137)       0:00:25.466 ****** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Define lvm_volumes structures] *******************************************\nTuesday 24 February 2026  00:43:40 +0000 (0:00:00.134)       0:00:25.600 ****** \n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\nTASK [Generate lvm_volumes structure (block only)] *****************************\nTuesday 24 February 2026  00:43:40 +0000 (0:00:00.150)       0:00:25.750 ****** \n\u001b[0;32mok: [testbed-node-4] => (item={'key': 'sdb', 'value': {'osd_lvm_uuid': '294dd48f-9d6d-5079-8214-59e60af9e526'}})\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item={'key': 'sdc', 'value': {'osd_lvm_uuid': '7fbf4d63-7c6f-5c70-887c-2b73c978775b'}})\u001b[0m\n\nTASK [Generate lvm_volumes structure (block + db)] *****************************\nTuesday 24 February 2026  00:43:41 +0000 (0:00:00.145)       0:00:25.896 ****** \n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'sdb', 'value': {'osd_lvm_uuid': '294dd48f-9d6d-5079-8214-59e60af9e526'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'sdc', 'value': {'osd_lvm_uuid': '7fbf4d63-7c6f-5c70-887c-2b73c978775b'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Generate lvm_volumes structure (block + wal)] ****************************\nTuesday 24 February 2026  00:43:41 +0000 (0:00:00.129)       0:00:26.026 ****** \n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'sdb', 'value': {'osd_lvm_uuid': '294dd48f-9d6d-5079-8214-59e60af9e526'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'sdc', 'value': {'osd_lvm_uuid': '7fbf4d63-7c6f-5c70-887c-2b73c978775b'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Generate lvm_volumes structure (block + db + wal)] ***********************\nTuesday 24 February 2026  00:43:41 +0000 (0:00:00.153)       0:00:26.179 ****** \n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'sdb', 'value': {'osd_lvm_uuid': '294dd48f-9d6d-5079-8214-59e60af9e526'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'sdc', 'value': {'osd_lvm_uuid': '7fbf4d63-7c6f-5c70-887c-2b73c978775b'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Compile lvm_volumes] *****************************************************\nTuesday 24 February 2026  00:43:41 +0000 (0:00:00.133)       0:00:26.313 ****** \n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\nTASK [Set OSD devices config data] *********************************************\nTuesday 24 February 2026  00:43:41 +0000 (0:00:00.114)       0:00:26.427 ****** \n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\nTASK [Set DB devices config data] **********************************************\nTuesday 24 February 2026  00:43:41 +0000 (0:00:00.133)       0:00:26.561 ****** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Set WAL devices config data] *********************************************\nTuesday 24 February 2026  00:43:42 +0000 (0:00:00.238)       0:00:26.799 ****** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Set DB+WAL devices config data] ******************************************\nTuesday 24 February 2026  00:43:42 +0000 (0:00:00.116)       0:00:26.916 ****** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Print ceph_osd_devices] **************************************************\nTuesday 24 February 2026  00:43:42 +0000 (0:00:00.128)       0:00:27.044 ****** \n\u001b[0;32mok: [testbed-node-4] => {\u001b[0m\n\u001b[0;32m    \"ceph_osd_devices\": {\u001b[0m\n\u001b[0;32m        \"sdb\": {\u001b[0m\n\u001b[0;32m            \"osd_lvm_uuid\": \"294dd48f-9d6d-5079-8214-59e60af9e526\"\u001b[0m\n\u001b[0;32m        },\u001b[0m\n\u001b[0;32m        \"sdc\": {\u001b[0m\n\u001b[0;32m            \"osd_lvm_uuid\": \"7fbf4d63-7c6f-5c70-887c-2b73c978775b\"\u001b[0m\n\u001b[0;32m        }\u001b[0m\n\u001b[0;32m    }\u001b[0m\n\u001b[0;32m}\u001b[0m\n\nTASK [Print WAL devices] *******************************************************\nTuesday 24 February 2026  00:43:42 +0000 (0:00:00.142)       0:00:27.187 ****** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Print DB devices] ********************************************************\nTuesday 24 February 2026  00:43:42 +0000 (0:00:00.119)       0:00:27.307 ****** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Print shared DB/WAL devices] *********************************************\nTuesday 24 February 2026  00:43:42 +0000 (0:00:00.122)       0:00:27.429 ****** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Print configuration data] ************************************************\nTuesday 24 February 2026  00:43:42 +0000 (0:00:00.103)       0:00:27.532 ****** \n\u001b[0;33mchanged: [testbed-node-4] => {\u001b[0m\n\u001b[0;33m    \"_ceph_configure_lvm_config_data\": {\u001b[0m\n\u001b[0;33m        \"ceph_osd_devices\": {\u001b[0m\n\u001b[0;33m            \"sdb\": {\u001b[0m\n\u001b[0;33m                \"osd_lvm_uuid\": \"294dd48f-9d6d-5079-8214-59e60af9e526\"\u001b[0m\n\u001b[0;33m            },\u001b[0m\n\u001b[0;33m            \"sdc\": {\u001b[0m\n\u001b[0;33m                \"osd_lvm_uuid\": \"7fbf4d63-7c6f-5c70-887c-2b73c978775b\"\u001b[0m\n\u001b[0;33m            }\u001b[0m\n\u001b[0;33m        },\u001b[0m\n\u001b[0;33m        \"lvm_volumes\": [\u001b[0m\n\u001b[0;33m            {\u001b[0m\n\u001b[0;33m                \"data\": \"osd-block-294dd48f-9d6d-5079-8214-59e60af9e526\",\u001b[0m\n\u001b[0;33m                \"data_vg\": \"ceph-294dd48f-9d6d-5079-8214-59e60af9e526\"\u001b[0m\n\u001b[0;33m            },\u001b[0m\n\u001b[0;33m            {\u001b[0m\n\u001b[0;33m                \"data\": \"osd-block-7fbf4d63-7c6f-5c70-887c-2b73c978775b\",\u001b[0m\n\u001b[0;33m                \"data_vg\": \"ceph-7fbf4d63-7c6f-5c70-887c-2b73c978775b\"\u001b[0m\n\u001b[0;33m            }\u001b[0m\n\u001b[0;33m        ]\u001b[0m\n\u001b[0;33m    }\u001b[0m\n\u001b[0;33m}\u001b[0m\n\nRUNNING HANDLER [Write configuration file] *************************************\nTuesday 24 February 2026  00:43:42 +0000 (0:00:00.180)       0:00:27.713 ****** \n\u001b[0;33mchanged: [testbed-node-4 -> testbed-manager(192.168.16.5)]\u001b[0m\n\nPLAY [Ceph configure LVM] ******************************************************\n\nTASK [Get extra vars for Ceph configuration] ***********************************\nTuesday 24 February 2026  00:43:43 +0000 (0:00:01.006)       0:00:28.719 ****** \n\u001b[0;32mok: [testbed-node-5 -> testbed-manager(192.168.16.5)]\u001b[0m\n\nTASK [Get initial list of available block devices] *****************************\nTuesday 24 February 2026  00:43:44 +0000 (0:00:00.479)       0:00:29.198 ****** \n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nTuesday 24 February 2026  00:43:44 +0000 (0:00:00.203)       0:00:29.402 ****** \n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-5 => (item=loop0)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-5 => (item=loop1)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-5 => (item=loop2)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-5 => (item=loop3)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-5 => (item=loop4)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-5 => (item=loop5)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-5 => (item=loop6)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-5 => (item=loop7)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-5 => (item=sda)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-5 => (item=sdb)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-5 => (item=sdc)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-5 => (item=sdd)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-5 => (item=sr0)\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nTuesday 24 February 2026  00:43:44 +0000 (0:00:00.334)       0:00:29.737 ****** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nTuesday 24 February 2026  00:43:45 +0000 (0:00:00.179)       0:00:29.917 ****** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nTuesday 24 February 2026  00:43:45 +0000 (0:00:00.172)       0:00:30.089 ****** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nTuesday 24 February 2026  00:43:45 +0000 (0:00:00.137)       0:00:30.227 ****** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nTuesday 24 February 2026  00:43:45 +0000 (0:00:00.131)       0:00:30.358 ****** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nTuesday 24 February 2026  00:43:45 +0000 (0:00:00.192)       0:00:30.551 ****** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nTuesday 24 February 2026  00:43:45 +0000 (0:00:00.191)       0:00:30.743 ****** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nTuesday 24 February 2026  00:43:46 +0000 (0:00:00.171)       0:00:30.914 ****** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nTuesday 24 February 2026  00:43:46 +0000 (0:00:00.188)       0:00:31.103 ****** \n\u001b[0;32mok: [testbed-node-5] => (item=scsi-0QEMU_QEMU_HARDDISK_7a9487cb-ec9e-4508-9c5a-f692700fc5a4)\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item=scsi-SQEMU_QEMU_HARDDISK_7a9487cb-ec9e-4508-9c5a-f692700fc5a4)\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nTuesday 24 February 2026  00:43:47 +0000 (0:00:00.659)       0:00:31.763 ****** \n\u001b[0;32mok: [testbed-node-5] => (item=scsi-0QEMU_QEMU_HARDDISK_1558935b-4c66-4614-97e5-cda14abd2641)\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item=scsi-SQEMU_QEMU_HARDDISK_1558935b-4c66-4614-97e5-cda14abd2641)\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nTuesday 24 February 2026  00:43:47 +0000 (0:00:00.453)       0:00:32.217 ****** \n\u001b[0;32mok: [testbed-node-5] => (item=scsi-0QEMU_QEMU_HARDDISK_6f4643f0-329b-4ae2-97c7-d348648fd970)\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item=scsi-SQEMU_QEMU_HARDDISK_6f4643f0-329b-4ae2-97c7-d348648fd970)\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nTuesday 24 February 2026  00:43:47 +0000 (0:00:00.376)       0:00:32.593 ****** \n\u001b[0;32mok: [testbed-node-5] => (item=scsi-0QEMU_QEMU_HARDDISK_5f47789f-2a1a-442a-8a3c-644a4f92e5a8)\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item=scsi-SQEMU_QEMU_HARDDISK_5f47789f-2a1a-442a-8a3c-644a4f92e5a8)\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nTuesday 24 February 2026  00:43:48 +0000 (0:00:00.403)       0:00:32.997 ****** \n\u001b[0;32mok: [testbed-node-5] => (item=ata-QEMU_DVD-ROM_QM00001)\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nTuesday 24 February 2026  00:43:48 +0000 (0:00:00.311)       0:00:33.308 ****** \n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-5 => (item=loop0)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-5 => (item=loop1)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-5 => (item=loop2)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-5 => (item=loop3)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-5 => (item=loop4)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-5 => (item=loop5)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-5 => (item=loop6)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-5 => (item=loop7)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-5 => (item=sda)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-5 => (item=sdb)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-5 => (item=sdc)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-5 => (item=sdd)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-5 => (item=sr0)\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nTuesday 24 February 2026  00:43:48 +0000 (0:00:00.354)       0:00:33.662 ****** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nTuesday 24 February 2026  00:43:49 +0000 (0:00:00.168)       0:00:33.830 ****** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nTuesday 24 February 2026  00:43:49 +0000 (0:00:00.169)       0:00:33.999 ****** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nTuesday 24 February 2026  00:43:49 +0000 (0:00:00.172)       0:00:34.172 ****** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nTuesday 24 February 2026  00:43:49 +0000 (0:00:00.148)       0:00:34.321 ****** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nTuesday 24 February 2026  00:43:49 +0000 (0:00:00.134)       0:00:34.455 ****** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nTuesday 24 February 2026  00:43:50 +0000 (0:00:00.402)       0:00:34.857 ****** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nTuesday 24 February 2026  00:43:50 +0000 (0:00:00.163)       0:00:35.021 ****** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nTuesday 24 February 2026  00:43:50 +0000 (0:00:00.178)       0:00:35.199 ****** \n\u001b[0;32mok: [testbed-node-5] => (item=sda1)\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item=sda14)\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item=sda15)\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item=sda16)\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nTuesday 24 February 2026  00:43:51 +0000 (0:00:00.564)       0:00:35.764 ****** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nTuesday 24 February 2026  00:43:51 +0000 (0:00:00.149)       0:00:35.914 ****** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nTuesday 24 February 2026  00:43:51 +0000 (0:00:00.187)       0:00:36.101 ****** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nTuesday 24 February 2026  00:43:51 +0000 (0:00:00.191)       0:00:36.292 ****** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Set UUIDs for OSD VGs/LVs] ***********************************************\nTuesday 24 February 2026  00:43:51 +0000 (0:00:00.181)       0:00:36.473 ****** \n\u001b[0;32mok: [testbed-node-5] => (item={'key': 'sdb', 'value': None})\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item={'key': 'sdc', 'value': None})\u001b[0m\n\nTASK [Generate WAL VG names] ***************************************************\nTuesday 24 February 2026  00:43:51 +0000 (0:00:00.153)       0:00:36.626 ****** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Generate DB VG names] ****************************************************\nTuesday 24 February 2026  00:43:51 +0000 (0:00:00.116)       0:00:36.743 ****** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Generate shared DB/WAL VG names] *****************************************\nTuesday 24 February 2026  00:43:52 +0000 (0:00:00.104)       0:00:36.848 ****** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Define lvm_volumes structures] *******************************************\nTuesday 24 February 2026  00:43:52 +0000 (0:00:00.302)       0:00:37.151 ****** \n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [Generate lvm_volumes structure (block only)] *****************************\nTuesday 24 February 2026  00:43:52 +0000 (0:00:00.126)       0:00:37.277 ****** \n\u001b[0;32mok: [testbed-node-5] => (item={'key': 'sdb', 'value': {'osd_lvm_uuid': '752a9ec4-4bfc-5551-82cd-92b802bdad97'}})\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item={'key': 'sdc', 'value': {'osd_lvm_uuid': 'ba85a0e3-2f57-51d5-a9f1-072f0e9a8deb'}})\u001b[0m\n\nTASK [Generate lvm_volumes structure (block + db)] *****************************\nTuesday 24 February 2026  00:43:52 +0000 (0:00:00.146)       0:00:37.424 ****** \n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'sdb', 'value': {'osd_lvm_uuid': '752a9ec4-4bfc-5551-82cd-92b802bdad97'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'sdc', 'value': {'osd_lvm_uuid': 'ba85a0e3-2f57-51d5-a9f1-072f0e9a8deb'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Generate lvm_volumes structure (block + wal)] ****************************\nTuesday 24 February 2026  00:43:52 +0000 (0:00:00.134)       0:00:37.558 ****** \n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'sdb', 'value': {'osd_lvm_uuid': '752a9ec4-4bfc-5551-82cd-92b802bdad97'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'sdc', 'value': {'osd_lvm_uuid': 'ba85a0e3-2f57-51d5-a9f1-072f0e9a8deb'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Generate lvm_volumes structure (block + db + wal)] ***********************\nTuesday 24 February 2026  00:43:52 +0000 (0:00:00.136)       0:00:37.695 ****** \n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'sdb', 'value': {'osd_lvm_uuid': '752a9ec4-4bfc-5551-82cd-92b802bdad97'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'sdc', 'value': {'osd_lvm_uuid': 'ba85a0e3-2f57-51d5-a9f1-072f0e9a8deb'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Compile lvm_volumes] *****************************************************\nTuesday 24 February 2026  00:43:53 +0000 (0:00:00.140)       0:00:37.835 ****** \n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [Set OSD devices config data] *********************************************\nTuesday 24 February 2026  00:43:53 +0000 (0:00:00.114)       0:00:37.950 ****** \n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [Set DB devices config data] **********************************************\nTuesday 24 February 2026  00:43:53 +0000 (0:00:00.116)       0:00:38.066 ****** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Set WAL devices config data] *********************************************\nTuesday 24 February 2026  00:43:53 +0000 (0:00:00.148)       0:00:38.215 ****** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Set DB+WAL devices config data] ******************************************\nTuesday 24 February 2026  00:43:53 +0000 (0:00:00.116)       0:00:38.331 ****** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Print ceph_osd_devices] **************************************************\nTuesday 24 February 2026  00:43:53 +0000 (0:00:00.113)       0:00:38.445 ****** \n\u001b[0;32mok: [testbed-node-5] => {\u001b[0m\n\u001b[0;32m    \"ceph_osd_devices\": {\u001b[0m\n\u001b[0;32m        \"sdb\": {\u001b[0m\n\u001b[0;32m            \"osd_lvm_uuid\": \"752a9ec4-4bfc-5551-82cd-92b802bdad97\"\u001b[0m\n\u001b[0;32m        },\u001b[0m\n\u001b[0;32m        \"sdc\": {\u001b[0m\n\u001b[0;32m            \"osd_lvm_uuid\": \"ba85a0e3-2f57-51d5-a9f1-072f0e9a8deb\"\u001b[0m\n\u001b[0;32m        }\u001b[0m\n\u001b[0;32m    }\u001b[0m\n\u001b[0;32m}\u001b[0m\n\nTASK [Print WAL devices] *******************************************************\nTuesday 24 February 2026  00:43:53 +0000 (0:00:00.127)       0:00:38.573 ****** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Print DB devices] ********************************************************\nTuesday 24 February 2026  00:43:54 +0000 (0:00:00.307)       0:00:38.880 ****** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Print shared DB/WAL devices] *********************************************\nTuesday 24 February 2026  00:43:54 +0000 (0:00:00.131)       0:00:39.011 ****** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Print configuration data] ************************************************\nTuesday 24 February 2026  00:43:54 +0000 (0:00:00.121)       0:00:39.133 ****** \n\u001b[0;33mchanged: [testbed-node-5] => {\u001b[0m\n\u001b[0;33m    \"_ceph_configure_lvm_config_data\": {\u001b[0m\n\u001b[0;33m        \"ceph_osd_devices\": {\u001b[0m\n\u001b[0;33m            \"sdb\": {\u001b[0m\n\u001b[0;33m                \"osd_lvm_uuid\": \"752a9ec4-4bfc-5551-82cd-92b802bdad97\"\u001b[0m\n\u001b[0;33m            },\u001b[0m\n\u001b[0;33m            \"sdc\": {\u001b[0m\n\u001b[0;33m                \"osd_lvm_uuid\": \"ba85a0e3-2f57-51d5-a9f1-072f0e9a8deb\"\u001b[0m\n\u001b[0;33m            }\u001b[0m\n\u001b[0;33m        },\u001b[0m\n\u001b[0;33m        \"lvm_volumes\": [\u001b[0m\n\u001b[0;33m            {\u001b[0m\n\u001b[0;33m                \"data\": \"osd-block-752a9ec4-4bfc-5551-82cd-92b802bdad97\",\u001b[0m\n\u001b[0;33m                \"data_vg\": \"ceph-752a9ec4-4bfc-5551-82cd-92b802bdad97\"\u001b[0m\n\u001b[0;33m            },\u001b[0m\n\u001b[0;33m            {\u001b[0m\n\u001b[0;33m                \"data\": \"osd-block-ba85a0e3-2f57-51d5-a9f1-072f0e9a8deb\",\u001b[0m\n\u001b[0;33m                \"data_vg\": \"ceph-ba85a0e3-2f57-51d5-a9f1-072f0e9a8deb\"\u001b[0m\n\u001b[0;33m            }\u001b[0m\n\u001b[0;33m        ]\u001b[0m\n\u001b[0;33m    }\u001b[0m\n\u001b[0;33m}\u001b[0m\n\nRUNNING HANDLER [Write configuration file] *************************************\nTuesday 24 February 2026  00:43:54 +0000 (0:00:00.190)       0:00:39.323 ****** \n\u001b[0;33mchanged: [testbed-node-5 -> testbed-manager(192.168.16.5)]\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;33mtestbed-node-3\u001b[0m             : \u001b[0;32mok=42  \u001b[0m \u001b[0;33mchanged=2   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=32  \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-4\u001b[0m             : \u001b[0;32mok=42  \u001b[0m \u001b[0;33mchanged=2   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=32  \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-5\u001b[0m             : \u001b[0;32mok=42  \u001b[0m \u001b[0;33mchanged=2   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=32  \u001b[0m rescued=0    ignored=0   \n\n\n\nTASKS RECAP ********************************************************************\nTuesday 24 February 2026  00:43:55 +0000 (0:00:00.848)       0:00:40.172 ****** \n=============================================================================== \nWrite configuration file ------------------------------------------------ 3.74s\nAdd known partitions to the list of available block devices ------------- 1.22s\nAdd known links to the list of available block devices ------------------ 1.13s\nAdd known partitions to the list of available block devices ------------- 1.09s\nGet extra vars for Ceph configuration ----------------------------------- 0.96s\nAdd known partitions to the list of available block devices ------------- 0.90s\nAdd known links to the list of available block devices ------------------ 0.90s\nAdd known partitions to the list of available block devices ------------- 0.85s\nPrint configuration data ------------------------------------------------ 0.79s\nAdd known partitions to the list of available block devices ------------- 0.74s\nAdd known links to the list of available block devices ------------------ 0.69s\nGenerate lvm_volumes structure (block + wal) ---------------------------- 0.68s\nAdd known links to the list of available block devices ------------------ 0.66s\nGet initial list of available block devices ----------------------------- 0.65s\nAdd known links to the list of available block devices ------------------ 0.63s\nGenerate shared DB/WAL VG names ----------------------------------------- 0.60s\nAdd known links to the list of available block devices ------------------ 0.58s\nPrint WAL devices ------------------------------------------------------- 0.57s\nAdd known partitions to the list of available block devices ------------- 0.56s\nSet UUIDs for OSD VGs/LVs ----------------------------------------------- 0.55s\n\u001b[32m2026-02-24 00:44:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 627b5eb2-2e98-4413-86a1-6581671e4d11 (sync inventory) is running in background. Output coming soon.\u001b[0m\n\u001b[32m2026-02-24 00:44:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mStarting group_vars file reorganization\u001b[0m\n\u001b[32m2026-02-24 00:44:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mMoved 0 file(s) to their respective directories\u001b[0m\n\u001b[32m2026-02-24 00:44:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mGroup_vars file reorganization completed\u001b[0m\n\u001b[32m2026-02-24 00:44:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mStarting variable preparation from inventory\u001b[0m\n\u001b[32m2026-02-24 00:44:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWriting 050-kolla-ceph-rgw-hosts.yml with ceph_rgw_hosts\u001b[0m\n\u001b[32m2026-02-24 00:44:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWriting 050-infrastructure-cephclient-mons.yml with cephclient_mons\u001b[0m\n\u001b[32m2026-02-24 00:44:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWriting 050-ceph-cluster-fsid.yml with ceph_cluster_fsid\u001b[0m\n\u001b[32m2026-02-24 00:44:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m3 file(s) written, 6 host(s) processed\u001b[0m\n\u001b[32m2026-02-24 00:44:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mVariable preparation completed\u001b[0m\n\u001b[32m2026-02-24 00:44:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mStarting inventory overwrite handling\u001b[0m\n\u001b[32m2026-02-24 00:44:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mHandling group overwrites in 99-overwrite\u001b[0m\n\u001b[32m2026-02-24 00:44:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mRemoving group frr:children from 60-generic\u001b[0m\n\u001b[32m2026-02-24 00:44:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mRemoving group netbird:children from 50-infrastructure\u001b[0m\n\u001b[32m2026-02-24 00:44:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mRemoving group ceph-rgw from 50-ceph\u001b[0m\n\u001b[32m2026-02-24 00:44:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mRemoving group ceph-mds from 50-ceph\u001b[0m\n\u001b[32m2026-02-24 00:44:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mHandling group overwrites in 20-roles\u001b[0m\n\u001b[32m2026-02-24 00:44:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mRemoving group k3s_node from 50-infrastructure\u001b[0m\n\u001b[32m2026-02-24 00:44:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mRemoved 5 group(s) in total\u001b[0m\n\u001b[32m2026-02-24 00:44:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mInventory overwrite handling completed\u001b[0m\n\u001b[32m2026-02-24 00:44:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mStarting merge of inventory files\u001b[0m\n\u001b[32m2026-02-24 00:44:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mInventory files merged successfully\u001b[0m\n\u001b[32m2026-02-24 00:44:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mGenerating ClusterShell configuration from Ansible inventory\u001b[0m\n\u001b[32m2026-02-24 00:44:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSuccessfully wrote ClusterShell configuration\u001b[0m\n[master 26aab81] 2026-02-24-00-44\n 1 file changed, 30 insertions(+), 9 deletions(-)\n\u001b[32m2026-02-24 00:44:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mPrepare task for execution of ceph-create-lvm-devices.\u001b[0m\n\u001b[32m2026-02-24 00:44:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1c145c4b-6a66-4ef0-a436-014fc77a1571 (ceph-create-lvm-devices) was prepared for execution.\u001b[0m\n\u001b[32m2026-02-24 00:44:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task 1c145c4b-6a66-4ef0-a436-014fc77a1571 (ceph-create-lvm-devices) has been started and output is visible here.\u001b[0m\n\u001b[1;35m[WARNING]: Collection community.general does not support Ansible version\u001b[0m\n\u001b[1;35m2.16.14\u001b[0m\n\nPLAY [Ceph create LVM devices] *************************************************\n\nTASK [Get extra vars for Ceph configuration] ***********************************\nTuesday 24 February 2026  00:44:47 +0000 (0:00:00.223)       0:00:00.223 ****** \n\u001b[0;32mok: [testbed-node-3 -> testbed-manager(192.168.16.5)]\u001b[0m\n\nTASK [Get initial list of available block devices] *****************************\nTuesday 24 February 2026  00:44:47 +0000 (0:00:00.190)       0:00:00.413 ****** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nTuesday 24 February 2026  00:44:47 +0000 (0:00:00.179)       0:00:00.593 ****** \n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-3 => (item=loop0)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-3 => (item=loop1)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-3 => (item=loop2)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-3 => (item=loop3)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-3 => (item=loop4)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-3 => (item=loop5)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-3 => (item=loop6)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-3 => (item=loop7)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-3 => (item=sda)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-3 => (item=sdb)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-3 => (item=sdc)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-3 => (item=sdd)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-3 => (item=sr0)\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nTuesday 24 February 2026  00:44:48 +0000 (0:00:00.509)       0:00:01.102 ****** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nTuesday 24 February 2026  00:44:48 +0000 (0:00:00.184)       0:00:01.287 ****** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nTuesday 24 February 2026  00:44:48 +0000 (0:00:00.177)       0:00:01.464 ****** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nTuesday 24 February 2026  00:44:48 +0000 (0:00:00.183)       0:00:01.648 ****** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nTuesday 24 February 2026  00:44:48 +0000 (0:00:00.181)       0:00:01.830 ****** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nTuesday 24 February 2026  00:44:49 +0000 (0:00:00.176)       0:00:02.006 ****** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nTuesday 24 February 2026  00:44:49 +0000 (0:00:00.181)       0:00:02.188 ****** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nTuesday 24 February 2026  00:44:49 +0000 (0:00:00.189)       0:00:02.377 ****** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nTuesday 24 February 2026  00:44:49 +0000 (0:00:00.190)       0:00:02.568 ****** \n\u001b[0;32mok: [testbed-node-3] => (item=scsi-0QEMU_QEMU_HARDDISK_f90a4e80-1f36-4d14-b7da-de7b20e9fbca)\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item=scsi-SQEMU_QEMU_HARDDISK_f90a4e80-1f36-4d14-b7da-de7b20e9fbca)\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nTuesday 24 February 2026  00:44:50 +0000 (0:00:00.375)       0:00:02.943 ****** \n\u001b[0;32mok: [testbed-node-3] => (item=scsi-0QEMU_QEMU_HARDDISK_9c6b6763-9704-41ec-9b6d-97eb80a5145c)\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item=scsi-SQEMU_QEMU_HARDDISK_9c6b6763-9704-41ec-9b6d-97eb80a5145c)\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nTuesday 24 February 2026  00:44:50 +0000 (0:00:00.496)       0:00:03.440 ****** \n\u001b[0;32mok: [testbed-node-3] => (item=scsi-0QEMU_QEMU_HARDDISK_6644f191-69e9-4f3a-baff-cffe4152595b)\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item=scsi-SQEMU_QEMU_HARDDISK_6644f191-69e9-4f3a-baff-cffe4152595b)\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nTuesday 24 February 2026  00:44:51 +0000 (0:00:00.522)       0:00:03.963 ****** \n\u001b[0;32mok: [testbed-node-3] => (item=scsi-0QEMU_QEMU_HARDDISK_4534650b-23a0-46c8-8bf1-44520caee637)\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item=scsi-SQEMU_QEMU_HARDDISK_4534650b-23a0-46c8-8bf1-44520caee637)\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nTuesday 24 February 2026  00:44:51 +0000 (0:00:00.553)       0:00:04.516 ****** \n\u001b[0;32mok: [testbed-node-3] => (item=ata-QEMU_DVD-ROM_QM00001)\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nTuesday 24 February 2026  00:44:51 +0000 (0:00:00.296)       0:00:04.812 ****** \n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-3 => (item=loop0)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-3 => (item=loop1)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-3 => (item=loop2)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-3 => (item=loop3)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-3 => (item=loop4)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-3 => (item=loop5)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-3 => (item=loop6)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-3 => (item=loop7)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-3 => (item=sda)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-3 => (item=sdb)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-3 => (item=sdc)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-3 => (item=sdd)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-3 => (item=sr0)\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nTuesday 24 February 2026  00:44:52 +0000 (0:00:00.383)       0:00:05.196 ****** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nTuesday 24 February 2026  00:44:52 +0000 (0:00:00.201)       0:00:05.398 ****** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nTuesday 24 February 2026  00:44:52 +0000 (0:00:00.184)       0:00:05.582 ****** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nTuesday 24 February 2026  00:44:52 +0000 (0:00:00.195)       0:00:05.778 ****** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nTuesday 24 February 2026  00:44:53 +0000 (0:00:00.179)       0:00:05.958 ****** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nTuesday 24 February 2026  00:44:53 +0000 (0:00:00.190)       0:00:06.149 ****** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nTuesday 24 February 2026  00:44:53 +0000 (0:00:00.180)       0:00:06.329 ****** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nTuesday 24 February 2026  00:44:53 +0000 (0:00:00.202)       0:00:06.532 ****** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nTuesday 24 February 2026  00:44:53 +0000 (0:00:00.218)       0:00:06.750 ****** \n\u001b[0;32mok: [testbed-node-3] => (item=sda1)\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item=sda14)\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item=sda15)\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item=sda16)\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nTuesday 24 February 2026  00:44:54 +0000 (0:00:01.065)       0:00:07.816 ****** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nTuesday 24 February 2026  00:44:55 +0000 (0:00:00.306)       0:00:08.122 ****** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nTuesday 24 February 2026  00:44:55 +0000 (0:00:00.256)       0:00:08.379 ****** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nTuesday 24 February 2026  00:44:55 +0000 (0:00:00.233)       0:00:08.613 ****** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Check whether ceph_db_wal_devices is used exclusively] *******************\nTuesday 24 February 2026  00:44:55 +0000 (0:00:00.190)       0:00:08.803 ****** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Create dict of block VGs -> PVs from ceph_osd_devices] *******************\nTuesday 24 February 2026  00:44:56 +0000 (0:00:00.159)       0:00:08.963 ****** \n\u001b[0;32mok: [testbed-node-3] => (item={'key': 'sdb', 'value': {'osd_lvm_uuid': '9b92f24c-dac1-5bff-a390-34d9ff9e0f4f'}})\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item={'key': 'sdc', 'value': {'osd_lvm_uuid': '7f6a8fe5-6adc-52b4-b256-edcdbfd9bf4e'}})\u001b[0m\n\nTASK [Create block VGs] ********************************************************\nTuesday 24 February 2026  00:44:56 +0000 (0:00:00.197)       0:00:09.161 ****** \n\u001b[0;33mchanged: [testbed-node-3] => (item={'data': 'osd-block-9b92f24c-dac1-5bff-a390-34d9ff9e0f4f', 'data_vg': 'ceph-9b92f24c-dac1-5bff-a390-34d9ff9e0f4f'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'data': 'osd-block-7f6a8fe5-6adc-52b4-b256-edcdbfd9bf4e', 'data_vg': 'ceph-7f6a8fe5-6adc-52b4-b256-edcdbfd9bf4e'})\u001b[0m\n\nTASK [Print 'Create block VGs'] ************************************************\nTuesday 24 February 2026  00:44:58 +0000 (0:00:01.974)       0:00:11.135 ****** \n\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-9b92f24c-dac1-5bff-a390-34d9ff9e0f4f', 'data_vg': 'ceph-9b92f24c-dac1-5bff-a390-34d9ff9e0f4f'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-7f6a8fe5-6adc-52b4-b256-edcdbfd9bf4e', 'data_vg': 'ceph-7f6a8fe5-6adc-52b4-b256-edcdbfd9bf4e'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Create block LVs] ********************************************************\nTuesday 24 February 2026  00:44:58 +0000 (0:00:00.164)       0:00:11.299 ****** \n\u001b[0;33mchanged: [testbed-node-3] => (item={'data': 'osd-block-9b92f24c-dac1-5bff-a390-34d9ff9e0f4f', 'data_vg': 'ceph-9b92f24c-dac1-5bff-a390-34d9ff9e0f4f'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'data': 'osd-block-7f6a8fe5-6adc-52b4-b256-edcdbfd9bf4e', 'data_vg': 'ceph-7f6a8fe5-6adc-52b4-b256-edcdbfd9bf4e'})\u001b[0m\n\nTASK [Print 'Create block LVs'] ************************************************\nTuesday 24 February 2026  00:44:59 +0000 (0:00:01.528)       0:00:12.828 ****** \n\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-9b92f24c-dac1-5bff-a390-34d9ff9e0f4f', 'data_vg': 'ceph-9b92f24c-dac1-5bff-a390-34d9ff9e0f4f'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-7f6a8fe5-6adc-52b4-b256-edcdbfd9bf4e', 'data_vg': 'ceph-7f6a8fe5-6adc-52b4-b256-edcdbfd9bf4e'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Create DB VGs] ***********************************************************\nTuesday 24 February 2026  00:45:00 +0000 (0:00:00.150)       0:00:12.978 ****** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Print 'Create DB VGs'] ***************************************************\nTuesday 24 February 2026  00:45:00 +0000 (0:00:00.128)       0:00:13.107 ****** \n\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-9b92f24c-dac1-5bff-a390-34d9ff9e0f4f', 'data_vg': 'ceph-9b92f24c-dac1-5bff-a390-34d9ff9e0f4f'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-7f6a8fe5-6adc-52b4-b256-edcdbfd9bf4e', 'data_vg': 'ceph-7f6a8fe5-6adc-52b4-b256-edcdbfd9bf4e'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Create WAL VGs] **********************************************************\nTuesday 24 February 2026  00:45:00 +0000 (0:00:00.357)       0:00:13.464 ****** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Print 'Create WAL VGs'] **************************************************\nTuesday 24 February 2026  00:45:00 +0000 (0:00:00.135)       0:00:13.600 ****** \n\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-9b92f24c-dac1-5bff-a390-34d9ff9e0f4f', 'data_vg': 'ceph-9b92f24c-dac1-5bff-a390-34d9ff9e0f4f'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-7f6a8fe5-6adc-52b4-b256-edcdbfd9bf4e', 'data_vg': 'ceph-7f6a8fe5-6adc-52b4-b256-edcdbfd9bf4e'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Create DB+WAL VGs] *******************************************************\nTuesday 24 February 2026  00:45:00 +0000 (0:00:00.169)       0:00:13.769 ****** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Print 'Create DB+WAL VGs'] ***********************************************\nTuesday 24 February 2026  00:45:01 +0000 (0:00:00.142)       0:00:13.912 ****** \n\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-9b92f24c-dac1-5bff-a390-34d9ff9e0f4f', 'data_vg': 'ceph-9b92f24c-dac1-5bff-a390-34d9ff9e0f4f'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-7f6a8fe5-6adc-52b4-b256-edcdbfd9bf4e', 'data_vg': 'ceph-7f6a8fe5-6adc-52b4-b256-edcdbfd9bf4e'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Prepare variables for OSD count check] ***********************************\nTuesday 24 February 2026  00:45:01 +0000 (0:00:00.187)       0:00:14.100 ****** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\nTASK [Count OSDs put on ceph_db_devices defined in lvm_volumes] ****************\nTuesday 24 February 2026  00:45:01 +0000 (0:00:00.165)       0:00:14.265 ****** \n\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-9b92f24c-dac1-5bff-a390-34d9ff9e0f4f', 'data_vg': 'ceph-9b92f24c-dac1-5bff-a390-34d9ff9e0f4f'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-7f6a8fe5-6adc-52b4-b256-edcdbfd9bf4e', 'data_vg': 'ceph-7f6a8fe5-6adc-52b4-b256-edcdbfd9bf4e'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Count OSDs put on ceph_wal_devices defined in lvm_volumes] ***************\nTuesday 24 February 2026  00:45:01 +0000 (0:00:00.206)       0:00:14.471 ****** \n\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-9b92f24c-dac1-5bff-a390-34d9ff9e0f4f', 'data_vg': 'ceph-9b92f24c-dac1-5bff-a390-34d9ff9e0f4f'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-7f6a8fe5-6adc-52b4-b256-edcdbfd9bf4e', 'data_vg': 'ceph-7f6a8fe5-6adc-52b4-b256-edcdbfd9bf4e'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Count OSDs put on ceph_db_wal_devices defined in lvm_volumes] ************\nTuesday 24 February 2026  00:45:01 +0000 (0:00:00.168)       0:00:14.639 ****** \n\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-9b92f24c-dac1-5bff-a390-34d9ff9e0f4f', 'data_vg': 'ceph-9b92f24c-dac1-5bff-a390-34d9ff9e0f4f'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-7f6a8fe5-6adc-52b4-b256-edcdbfd9bf4e', 'data_vg': 'ceph-7f6a8fe5-6adc-52b4-b256-edcdbfd9bf4e'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Fail if number of OSDs exceeds num_osds for a DB VG] *********************\nTuesday 24 February 2026  00:45:01 +0000 (0:00:00.179)       0:00:14.819 ****** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Fail if number of OSDs exceeds num_osds for a WAL VG] ********************\nTuesday 24 February 2026  00:45:02 +0000 (0:00:00.142)       0:00:14.961 ****** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Fail if number of OSDs exceeds num_osds for a DB+WAL VG] *****************\nTuesday 24 February 2026  00:45:02 +0000 (0:00:00.159)       0:00:15.121 ****** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Print number of OSDs wanted per DB VG] ***********************************\nTuesday 24 February 2026  00:45:02 +0000 (0:00:00.131)       0:00:15.252 ****** \n\u001b[0;32mok: [testbed-node-3] => {\u001b[0m\n\u001b[0;32m    \"_num_osds_wanted_per_db_vg\": {}\u001b[0m\n\u001b[0;32m}\u001b[0m\n\nTASK [Print number of OSDs wanted per WAL VG] **********************************\nTuesday 24 February 2026  00:45:02 +0000 (0:00:00.435)       0:00:15.688 ****** \n\u001b[0;32mok: [testbed-node-3] => {\u001b[0m\n\u001b[0;32m    \"_num_osds_wanted_per_wal_vg\": {}\u001b[0m\n\u001b[0;32m}\u001b[0m\n\nTASK [Print number of OSDs wanted per DB+WAL VG] *******************************\nTuesday 24 February 2026  00:45:03 +0000 (0:00:00.164)       0:00:15.853 ****** \n\u001b[0;32mok: [testbed-node-3] => {\u001b[0m\n\u001b[0;32m    \"_num_osds_wanted_per_db_wal_vg\": {}\u001b[0m\n\u001b[0;32m}\u001b[0m\n\nTASK [Gather DB VGs with total and available size in bytes] ********************\nTuesday 24 February 2026  00:45:03 +0000 (0:00:00.157)       0:00:16.010 ****** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\nTASK [Gather WAL VGs with total and available size in bytes] *******************\nTuesday 24 February 2026  00:45:03 +0000 (0:00:00.682)       0:00:16.692 ****** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\nTASK [Gather DB+WAL VGs with total and available size in bytes] ****************\nTuesday 24 February 2026  00:45:04 +0000 (0:00:00.555)       0:00:17.248 ****** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\nTASK [Combine JSON from _db/wal/db_wal_vgs_cmd_output] *************************\nTuesday 24 February 2026  00:45:04 +0000 (0:00:00.552)       0:00:17.800 ****** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\nTASK [Calculate VG sizes (without buffer)] *************************************\nTuesday 24 February 2026  00:45:05 +0000 (0:00:00.167)       0:00:17.968 ****** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Calculate VG sizes (with buffer)] ****************************************\nTuesday 24 February 2026  00:45:05 +0000 (0:00:00.118)       0:00:18.086 ****** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Print LVM VGs report data] ***********************************************\nTuesday 24 February 2026  00:45:05 +0000 (0:00:00.105)       0:00:18.192 ****** \n\u001b[0;32mok: [testbed-node-3] => {\u001b[0m\n\u001b[0;32m    \"vgs_report\": {\u001b[0m\n\u001b[0;32m        \"vg\": []\u001b[0m\n\u001b[0;32m    }\u001b[0m\n\u001b[0;32m}\u001b[0m\n\nTASK [Print LVM VG sizes] ******************************************************\nTuesday 24 February 2026  00:45:05 +0000 (0:00:00.160)       0:00:18.353 ****** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Calculate size needed for LVs on ceph_db_devices] ************************\nTuesday 24 February 2026  00:45:05 +0000 (0:00:00.138)       0:00:18.491 ****** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Print size needed for LVs on ceph_db_devices] ****************************\nTuesday 24 February 2026  00:45:05 +0000 (0:00:00.157)       0:00:18.649 ****** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Fail if size of DB LVs on ceph_db_devices > available] *******************\nTuesday 24 February 2026  00:45:06 +0000 (0:00:00.353)       0:00:19.002 ****** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Calculate size needed for LVs on ceph_wal_devices] ***********************\nTuesday 24 February 2026  00:45:06 +0000 (0:00:00.140)       0:00:19.143 ****** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Print size needed for LVs on ceph_wal_devices] ***************************\nTuesday 24 February 2026  00:45:06 +0000 (0:00:00.162)       0:00:19.306 ****** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Fail if size of WAL LVs on ceph_wal_devices > available] *****************\nTuesday 24 February 2026  00:45:06 +0000 (0:00:00.187)       0:00:19.493 ****** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Calculate size needed for WAL LVs on ceph_db_wal_devices] ****************\nTuesday 24 February 2026  00:45:06 +0000 (0:00:00.137)       0:00:19.631 ****** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Print size needed for WAL LVs on ceph_db_wal_devices] ********************\nTuesday 24 February 2026  00:45:06 +0000 (0:00:00.138)       0:00:19.770 ****** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Calculate size needed for DB LVs on ceph_db_wal_devices] *****************\nTuesday 24 February 2026  00:45:07 +0000 (0:00:00.138)       0:00:19.908 ****** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Print size needed for DB LVs on ceph_db_wal_devices] *********************\nTuesday 24 February 2026  00:45:07 +0000 (0:00:00.146)       0:00:20.055 ****** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Fail if size of DB+WAL LVs on ceph_db_wal_devices > available] ***********\nTuesday 24 February 2026  00:45:07 +0000 (0:00:00.135)       0:00:20.191 ****** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Fail if DB LV size < 30 GiB for ceph_db_devices] *************************\nTuesday 24 February 2026  00:45:07 +0000 (0:00:00.126)       0:00:20.318 ****** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Fail if DB LV size < 30 GiB for ceph_db_wal_devices] *********************\nTuesday 24 February 2026  00:45:07 +0000 (0:00:00.131)       0:00:20.449 ****** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Create DB LVs for ceph_db_devices] ***************************************\nTuesday 24 February 2026  00:45:07 +0000 (0:00:00.136)       0:00:20.586 ****** \n\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-9b92f24c-dac1-5bff-a390-34d9ff9e0f4f', 'data_vg': 'ceph-9b92f24c-dac1-5bff-a390-34d9ff9e0f4f'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-7f6a8fe5-6adc-52b4-b256-edcdbfd9bf4e', 'data_vg': 'ceph-7f6a8fe5-6adc-52b4-b256-edcdbfd9bf4e'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Print 'Create DB LVs for ceph_db_devices'] *******************************\nTuesday 24 February 2026  00:45:08 +0000 (0:00:00.346)       0:00:20.932 ****** \n\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-9b92f24c-dac1-5bff-a390-34d9ff9e0f4f', 'data_vg': 'ceph-9b92f24c-dac1-5bff-a390-34d9ff9e0f4f'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-7f6a8fe5-6adc-52b4-b256-edcdbfd9bf4e', 'data_vg': 'ceph-7f6a8fe5-6adc-52b4-b256-edcdbfd9bf4e'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Create WAL LVs for ceph_wal_devices] *************************************\nTuesday 24 February 2026  00:45:08 +0000 (0:00:00.158)       0:00:21.091 ****** \n\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-9b92f24c-dac1-5bff-a390-34d9ff9e0f4f', 'data_vg': 'ceph-9b92f24c-dac1-5bff-a390-34d9ff9e0f4f'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-7f6a8fe5-6adc-52b4-b256-edcdbfd9bf4e', 'data_vg': 'ceph-7f6a8fe5-6adc-52b4-b256-edcdbfd9bf4e'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Print 'Create WAL LVs for ceph_wal_devices'] *****************************\nTuesday 24 February 2026  00:45:08 +0000 (0:00:00.152)       0:00:21.243 ****** \n\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-9b92f24c-dac1-5bff-a390-34d9ff9e0f4f', 'data_vg': 'ceph-9b92f24c-dac1-5bff-a390-34d9ff9e0f4f'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-7f6a8fe5-6adc-52b4-b256-edcdbfd9bf4e', 'data_vg': 'ceph-7f6a8fe5-6adc-52b4-b256-edcdbfd9bf4e'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Create WAL LVs for ceph_db_wal_devices] **********************************\nTuesday 24 February 2026  00:45:08 +0000 (0:00:00.161)       0:00:21.404 ****** \n\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-9b92f24c-dac1-5bff-a390-34d9ff9e0f4f', 'data_vg': 'ceph-9b92f24c-dac1-5bff-a390-34d9ff9e0f4f'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-7f6a8fe5-6adc-52b4-b256-edcdbfd9bf4e', 'data_vg': 'ceph-7f6a8fe5-6adc-52b4-b256-edcdbfd9bf4e'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Print 'Create WAL LVs for ceph_db_wal_devices'] **************************\nTuesday 24 February 2026  00:45:08 +0000 (0:00:00.165)       0:00:21.570 ****** \n\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-9b92f24c-dac1-5bff-a390-34d9ff9e0f4f', 'data_vg': 'ceph-9b92f24c-dac1-5bff-a390-34d9ff9e0f4f'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-7f6a8fe5-6adc-52b4-b256-edcdbfd9bf4e', 'data_vg': 'ceph-7f6a8fe5-6adc-52b4-b256-edcdbfd9bf4e'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Create DB LVs for ceph_db_wal_devices] ***********************************\nTuesday 24 February 2026  00:45:08 +0000 (0:00:00.146)       0:00:21.716 ****** \n\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-9b92f24c-dac1-5bff-a390-34d9ff9e0f4f', 'data_vg': 'ceph-9b92f24c-dac1-5bff-a390-34d9ff9e0f4f'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-7f6a8fe5-6adc-52b4-b256-edcdbfd9bf4e', 'data_vg': 'ceph-7f6a8fe5-6adc-52b4-b256-edcdbfd9bf4e'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Print 'Create DB LVs for ceph_db_wal_devices'] ***************************\nTuesday 24 February 2026  00:45:09 +0000 (0:00:00.138)       0:00:21.855 ****** \n\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-9b92f24c-dac1-5bff-a390-34d9ff9e0f4f', 'data_vg': 'ceph-9b92f24c-dac1-5bff-a390-34d9ff9e0f4f'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-7f6a8fe5-6adc-52b4-b256-edcdbfd9bf4e', 'data_vg': 'ceph-7f6a8fe5-6adc-52b4-b256-edcdbfd9bf4e'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Get list of Ceph LVs with associated VGs] ********************************\nTuesday 24 February 2026  00:45:09 +0000 (0:00:00.133)       0:00:21.989 ****** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\nTASK [Get list of Ceph PVs with associated VGs] ********************************\nTuesday 24 February 2026  00:45:09 +0000 (0:00:00.496)       0:00:22.485 ****** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\nTASK [Combine JSON from _lvs_cmd_output/_pvs_cmd_output] ***********************\nTuesday 24 February 2026  00:45:10 +0000 (0:00:00.502)       0:00:22.988 ****** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\nTASK [Create list of VG/LV names] **********************************************\nTuesday 24 February 2026  00:45:10 +0000 (0:00:00.137)       0:00:23.126 ****** \n\u001b[0;32mok: [testbed-node-3] => (item={'lv_name': 'osd-block-7f6a8fe5-6adc-52b4-b256-edcdbfd9bf4e', 'vg_name': 'ceph-7f6a8fe5-6adc-52b4-b256-edcdbfd9bf4e'})\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item={'lv_name': 'osd-block-9b92f24c-dac1-5bff-a390-34d9ff9e0f4f', 'vg_name': 'ceph-9b92f24c-dac1-5bff-a390-34d9ff9e0f4f'})\u001b[0m\n\nTASK [Fail if block LV defined in lvm_volumes is missing] **********************\nTuesday 24 February 2026  00:45:10 +0000 (0:00:00.146)       0:00:23.273 ****** \n\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-9b92f24c-dac1-5bff-a390-34d9ff9e0f4f', 'data_vg': 'ceph-9b92f24c-dac1-5bff-a390-34d9ff9e0f4f'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-7f6a8fe5-6adc-52b4-b256-edcdbfd9bf4e', 'data_vg': 'ceph-7f6a8fe5-6adc-52b4-b256-edcdbfd9bf4e'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Fail if DB LV defined in lvm_volumes is missing] *************************\nTuesday 24 February 2026  00:45:10 +0000 (0:00:00.256)       0:00:23.529 ****** \n\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-9b92f24c-dac1-5bff-a390-34d9ff9e0f4f', 'data_vg': 'ceph-9b92f24c-dac1-5bff-a390-34d9ff9e0f4f'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-7f6a8fe5-6adc-52b4-b256-edcdbfd9bf4e', 'data_vg': 'ceph-7f6a8fe5-6adc-52b4-b256-edcdbfd9bf4e'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Fail if WAL LV defined in lvm_volumes is missing] ************************\nTuesday 24 February 2026  00:45:10 +0000 (0:00:00.128)       0:00:23.658 ****** \n\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-9b92f24c-dac1-5bff-a390-34d9ff9e0f4f', 'data_vg': 'ceph-9b92f24c-dac1-5bff-a390-34d9ff9e0f4f'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-7f6a8fe5-6adc-52b4-b256-edcdbfd9bf4e', 'data_vg': 'ceph-7f6a8fe5-6adc-52b4-b256-edcdbfd9bf4e'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Print LVM report data] ***************************************************\nTuesday 24 February 2026  00:45:10 +0000 (0:00:00.134)       0:00:23.793 ****** \n\u001b[0;32mok: [testbed-node-3] => {\u001b[0m\n\u001b[0;32m    \"lvm_report\": {\u001b[0m\n\u001b[0;32m        \"lv\": [\u001b[0m\n\u001b[0;32m            {\u001b[0m\n\u001b[0;32m                \"lv_name\": \"osd-block-7f6a8fe5-6adc-52b4-b256-edcdbfd9bf4e\",\u001b[0m\n\u001b[0;32m                \"vg_name\": \"ceph-7f6a8fe5-6adc-52b4-b256-edcdbfd9bf4e\"\u001b[0m\n\u001b[0;32m            },\u001b[0m\n\u001b[0;32m            {\u001b[0m\n\u001b[0;32m                \"lv_name\": \"osd-block-9b92f24c-dac1-5bff-a390-34d9ff9e0f4f\",\u001b[0m\n\u001b[0;32m                \"vg_name\": \"ceph-9b92f24c-dac1-5bff-a390-34d9ff9e0f4f\"\u001b[0m\n\u001b[0;32m            }\u001b[0m\n\u001b[0;32m        ],\u001b[0m\n\u001b[0;32m        \"pv\": [\u001b[0m\n\u001b[0;32m            {\u001b[0m\n\u001b[0;32m                \"pv_name\": \"/dev/sdb\",\u001b[0m\n\u001b[0;32m                \"vg_name\": \"ceph-9b92f24c-dac1-5bff-a390-34d9ff9e0f4f\"\u001b[0m\n\u001b[0;32m            },\u001b[0m\n\u001b[0;32m            {\u001b[0m\n\u001b[0;32m                \"pv_name\": \"/dev/sdc\",\u001b[0m\n\u001b[0;32m                \"vg_name\": \"ceph-7f6a8fe5-6adc-52b4-b256-edcdbfd9bf4e\"\u001b[0m\n\u001b[0;32m            }\u001b[0m\n\u001b[0;32m        ]\u001b[0m\n\u001b[0;32m    }\u001b[0m\n\u001b[0;32m}\u001b[0m\n\nPLAY [Ceph create LVM devices] *************************************************\n\nTASK [Get extra vars for Ceph configuration] ***********************************\nTuesday 24 February 2026  00:45:11 +0000 (0:00:00.253)       0:00:24.047 ****** \n\u001b[0;32mok: [testbed-node-4 -> testbed-manager(192.168.16.5)]\u001b[0m\n\nTASK [Get initial list of available block devices] *****************************\nTuesday 24 February 2026  00:45:11 +0000 (0:00:00.281)       0:00:24.328 ****** \n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nTuesday 24 February 2026  00:45:11 +0000 (0:00:00.229)       0:00:24.558 ****** \n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-4 => (item=loop0)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-4 => (item=loop1)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-4 => (item=loop2)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-4 => (item=loop3)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-4 => (item=loop4)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-4 => (item=loop5)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-4 => (item=loop6)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-4 => (item=loop7)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-4 => (item=sda)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-4 => (item=sdb)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-4 => (item=sdc)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-4 => (item=sdd)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-4 => (item=sr0)\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nTuesday 24 February 2026  00:45:12 +0000 (0:00:00.378)       0:00:24.936 ****** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nTuesday 24 February 2026  00:45:12 +0000 (0:00:00.203)       0:00:25.140 ****** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nTuesday 24 February 2026  00:45:12 +0000 (0:00:00.186)       0:00:25.326 ****** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nTuesday 24 February 2026  00:45:12 +0000 (0:00:00.481)       0:00:25.808 ****** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nTuesday 24 February 2026  00:45:13 +0000 (0:00:00.188)       0:00:25.996 ****** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nTuesday 24 February 2026  00:45:13 +0000 (0:00:00.204)       0:00:26.201 ****** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nTuesday 24 February 2026  00:45:13 +0000 (0:00:00.175)       0:00:26.377 ****** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nTuesday 24 February 2026  00:45:13 +0000 (0:00:00.182)       0:00:26.559 ****** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nTuesday 24 February 2026  00:45:13 +0000 (0:00:00.174)       0:00:26.734 ****** \n\u001b[0;32mok: [testbed-node-4] => (item=scsi-0QEMU_QEMU_HARDDISK_47e600ec-05ca-44a6-93d0-5ca3104e3639)\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item=scsi-SQEMU_QEMU_HARDDISK_47e600ec-05ca-44a6-93d0-5ca3104e3639)\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nTuesday 24 February 2026  00:45:14 +0000 (0:00:00.393)       0:00:27.127 ****** \n\u001b[0;32mok: [testbed-node-4] => (item=scsi-0QEMU_QEMU_HARDDISK_f659b61f-c39f-426f-a640-6a753951661d)\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item=scsi-SQEMU_QEMU_HARDDISK_f659b61f-c39f-426f-a640-6a753951661d)\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nTuesday 24 February 2026  00:45:14 +0000 (0:00:00.378)       0:00:27.506 ****** \n\u001b[0;32mok: [testbed-node-4] => (item=scsi-0QEMU_QEMU_HARDDISK_fb58bf16-a1c6-4abb-925b-29c6a6a6f100)\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item=scsi-SQEMU_QEMU_HARDDISK_fb58bf16-a1c6-4abb-925b-29c6a6a6f100)\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nTuesday 24 February 2026  00:45:15 +0000 (0:00:00.401)       0:00:27.907 ****** \n\u001b[0;32mok: [testbed-node-4] => (item=scsi-0QEMU_QEMU_HARDDISK_5660bb62-e2c9-45b2-ae11-8ab0f829b194)\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item=scsi-SQEMU_QEMU_HARDDISK_5660bb62-e2c9-45b2-ae11-8ab0f829b194)\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nTuesday 24 February 2026  00:45:15 +0000 (0:00:00.517)       0:00:28.424 ****** \n\u001b[0;32mok: [testbed-node-4] => (item=ata-QEMU_DVD-ROM_QM00001)\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nTuesday 24 February 2026  00:45:16 +0000 (0:00:00.447)       0:00:28.872 ****** \n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-4 => (item=loop0)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-4 => (item=loop1)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-4 => (item=loop2)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-4 => (item=loop3)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-4 => (item=loop4)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-4 => (item=loop5)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-4 => (item=loop6)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-4 => (item=loop7)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-4 => (item=sda)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-4 => (item=sdb)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-4 => (item=sdc)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-4 => (item=sdd)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-4 => (item=sr0)\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nTuesday 24 February 2026  00:45:16 +0000 (0:00:00.679)       0:00:29.551 ****** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nTuesday 24 February 2026  00:45:16 +0000 (0:00:00.175)       0:00:29.727 ****** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nTuesday 24 February 2026  00:45:17 +0000 (0:00:00.188)       0:00:29.915 ****** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nTuesday 24 February 2026  00:45:17 +0000 (0:00:00.183)       0:00:30.099 ****** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nTuesday 24 February 2026  00:45:17 +0000 (0:00:00.187)       0:00:30.286 ****** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nTuesday 24 February 2026  00:45:17 +0000 (0:00:00.185)       0:00:30.472 ****** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nTuesday 24 February 2026  00:45:17 +0000 (0:00:00.178)       0:00:30.650 ****** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nTuesday 24 February 2026  00:45:18 +0000 (0:00:00.192)       0:00:30.843 ****** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nTuesday 24 February 2026  00:45:18 +0000 (0:00:00.270)       0:00:31.114 ****** \n\u001b[0;32mok: [testbed-node-4] => (item=sda1)\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item=sda14)\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item=sda15)\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item=sda16)\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nTuesday 24 February 2026  00:45:19 +0000 (0:00:00.737)       0:00:31.851 ****** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nTuesday 24 February 2026  00:45:19 +0000 (0:00:00.181)       0:00:32.033 ****** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nTuesday 24 February 2026  00:45:19 +0000 (0:00:00.475)       0:00:32.509 ****** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nTuesday 24 February 2026  00:45:19 +0000 (0:00:00.196)       0:00:32.705 ****** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Check whether ceph_db_wal_devices is used exclusively] *******************\nTuesday 24 February 2026  00:45:20 +0000 (0:00:00.174)       0:00:32.880 ****** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Create dict of block VGs -> PVs from ceph_osd_devices] *******************\nTuesday 24 February 2026  00:45:20 +0000 (0:00:00.112)       0:00:32.992 ****** \n\u001b[0;32mok: [testbed-node-4] => (item={'key': 'sdb', 'value': {'osd_lvm_uuid': '294dd48f-9d6d-5079-8214-59e60af9e526'}})\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item={'key': 'sdc', 'value': {'osd_lvm_uuid': '7fbf4d63-7c6f-5c70-887c-2b73c978775b'}})\u001b[0m\n\nTASK [Create block VGs] ********************************************************\nTuesday 24 February 2026  00:45:20 +0000 (0:00:00.176)       0:00:33.169 ****** \n\u001b[0;33mchanged: [testbed-node-4] => (item={'data': 'osd-block-294dd48f-9d6d-5079-8214-59e60af9e526', 'data_vg': 'ceph-294dd48f-9d6d-5079-8214-59e60af9e526'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'data': 'osd-block-7fbf4d63-7c6f-5c70-887c-2b73c978775b', 'data_vg': 'ceph-7fbf4d63-7c6f-5c70-887c-2b73c978775b'})\u001b[0m\n\nTASK [Print 'Create block VGs'] ************************************************\nTuesday 24 February 2026  00:45:22 +0000 (0:00:01.839)       0:00:35.009 ****** \n\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-294dd48f-9d6d-5079-8214-59e60af9e526', 'data_vg': 'ceph-294dd48f-9d6d-5079-8214-59e60af9e526'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-7fbf4d63-7c6f-5c70-887c-2b73c978775b', 'data_vg': 'ceph-7fbf4d63-7c6f-5c70-887c-2b73c978775b'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Create block LVs] ********************************************************\nTuesday 24 February 2026  00:45:22 +0000 (0:00:00.144)       0:00:35.154 ****** \n\u001b[0;33mchanged: [testbed-node-4] => (item={'data': 'osd-block-294dd48f-9d6d-5079-8214-59e60af9e526', 'data_vg': 'ceph-294dd48f-9d6d-5079-8214-59e60af9e526'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'data': 'osd-block-7fbf4d63-7c6f-5c70-887c-2b73c978775b', 'data_vg': 'ceph-7fbf4d63-7c6f-5c70-887c-2b73c978775b'})\u001b[0m\n\nTASK [Print 'Create block LVs'] ************************************************\nTuesday 24 February 2026  00:45:23 +0000 (0:00:01.290)       0:00:36.444 ****** \n\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-294dd48f-9d6d-5079-8214-59e60af9e526', 'data_vg': 'ceph-294dd48f-9d6d-5079-8214-59e60af9e526'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-7fbf4d63-7c6f-5c70-887c-2b73c978775b', 'data_vg': 'ceph-7fbf4d63-7c6f-5c70-887c-2b73c978775b'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Create DB VGs] ***********************************************************\nTuesday 24 February 2026  00:45:23 +0000 (0:00:00.137)       0:00:36.582 ****** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Print 'Create DB VGs'] ***************************************************\nTuesday 24 February 2026  00:45:23 +0000 (0:00:00.125)       0:00:36.707 ****** \n\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-294dd48f-9d6d-5079-8214-59e60af9e526', 'data_vg': 'ceph-294dd48f-9d6d-5079-8214-59e60af9e526'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-7fbf4d63-7c6f-5c70-887c-2b73c978775b', 'data_vg': 'ceph-7fbf4d63-7c6f-5c70-887c-2b73c978775b'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Create WAL VGs] **********************************************************\nTuesday 24 February 2026  00:45:24 +0000 (0:00:00.133)       0:00:36.841 ****** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Print 'Create WAL VGs'] **************************************************\nTuesday 24 February 2026  00:45:24 +0000 (0:00:00.100)       0:00:36.941 ****** \n\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-294dd48f-9d6d-5079-8214-59e60af9e526', 'data_vg': 'ceph-294dd48f-9d6d-5079-8214-59e60af9e526'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-7fbf4d63-7c6f-5c70-887c-2b73c978775b', 'data_vg': 'ceph-7fbf4d63-7c6f-5c70-887c-2b73c978775b'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Create DB+WAL VGs] *******************************************************\nTuesday 24 February 2026  00:45:24 +0000 (0:00:00.252)       0:00:37.194 ****** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Print 'Create DB+WAL VGs'] ***********************************************\nTuesday 24 February 2026  00:45:24 +0000 (0:00:00.121)       0:00:37.316 ****** \n\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-294dd48f-9d6d-5079-8214-59e60af9e526', 'data_vg': 'ceph-294dd48f-9d6d-5079-8214-59e60af9e526'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-7fbf4d63-7c6f-5c70-887c-2b73c978775b', 'data_vg': 'ceph-7fbf4d63-7c6f-5c70-887c-2b73c978775b'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Prepare variables for OSD count check] ***********************************\nTuesday 24 February 2026  00:45:24 +0000 (0:00:00.142)       0:00:37.459 ****** \n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\nTASK [Count OSDs put on ceph_db_devices defined in lvm_volumes] ****************\nTuesday 24 February 2026  00:45:24 +0000 (0:00:00.124)       0:00:37.584 ****** \n\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-294dd48f-9d6d-5079-8214-59e60af9e526', 'data_vg': 'ceph-294dd48f-9d6d-5079-8214-59e60af9e526'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-7fbf4d63-7c6f-5c70-887c-2b73c978775b', 'data_vg': 'ceph-7fbf4d63-7c6f-5c70-887c-2b73c978775b'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Count OSDs put on ceph_wal_devices defined in lvm_volumes] ***************\nTuesday 24 February 2026  00:45:24 +0000 (0:00:00.135)       0:00:37.719 ****** \n\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-294dd48f-9d6d-5079-8214-59e60af9e526', 'data_vg': 'ceph-294dd48f-9d6d-5079-8214-59e60af9e526'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-7fbf4d63-7c6f-5c70-887c-2b73c978775b', 'data_vg': 'ceph-7fbf4d63-7c6f-5c70-887c-2b73c978775b'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Count OSDs put on ceph_db_wal_devices defined in lvm_volumes] ************\nTuesday 24 February 2026  00:45:25 +0000 (0:00:00.138)       0:00:37.858 ****** \n\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-294dd48f-9d6d-5079-8214-59e60af9e526', 'data_vg': 'ceph-294dd48f-9d6d-5079-8214-59e60af9e526'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-7fbf4d63-7c6f-5c70-887c-2b73c978775b', 'data_vg': 'ceph-7fbf4d63-7c6f-5c70-887c-2b73c978775b'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Fail if number of OSDs exceeds num_osds for a DB VG] *********************\nTuesday 24 February 2026  00:45:25 +0000 (0:00:00.142)       0:00:38.000 ****** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Fail if number of OSDs exceeds num_osds for a WAL VG] ********************\nTuesday 24 February 2026  00:45:25 +0000 (0:00:00.132)       0:00:38.133 ****** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Fail if number of OSDs exceeds num_osds for a DB+WAL VG] *****************\nTuesday 24 February 2026  00:45:25 +0000 (0:00:00.125)       0:00:38.258 ****** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Print number of OSDs wanted per DB VG] ***********************************\nTuesday 24 February 2026  00:45:25 +0000 (0:00:00.121)       0:00:38.380 ****** \n\u001b[0;32mok: [testbed-node-4] => {\u001b[0m\n\u001b[0;32m    \"_num_osds_wanted_per_db_vg\": {}\u001b[0m\n\u001b[0;32m}\u001b[0m\n\nTASK [Print number of OSDs wanted per WAL VG] **********************************\nTuesday 24 February 2026  00:45:25 +0000 (0:00:00.137)       0:00:38.518 ****** \n\u001b[0;32mok: [testbed-node-4] => {\u001b[0m\n\u001b[0;32m    \"_num_osds_wanted_per_wal_vg\": {}\u001b[0m\n\u001b[0;32m}\u001b[0m\n\nTASK [Print number of OSDs wanted per DB+WAL VG] *******************************\nTuesday 24 February 2026  00:45:25 +0000 (0:00:00.108)       0:00:38.626 ****** \n\u001b[0;32mok: [testbed-node-4] => {\u001b[0m\n\u001b[0;32m    \"_num_osds_wanted_per_db_wal_vg\": {}\u001b[0m\n\u001b[0;32m}\u001b[0m\n\nTASK [Gather DB VGs with total and available size in bytes] ********************\nTuesday 24 February 2026  00:45:26 +0000 (0:00:00.230)       0:00:38.856 ****** \n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\nTASK [Gather WAL VGs with total and available size in bytes] *******************\nTuesday 24 February 2026  00:45:26 +0000 (0:00:00.502)       0:00:39.359 ****** \n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\nTASK [Gather DB+WAL VGs with total and available size in bytes] ****************\nTuesday 24 February 2026  00:45:26 +0000 (0:00:00.468)       0:00:39.827 ****** \n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\nTASK [Combine JSON from _db/wal/db_wal_vgs_cmd_output] *************************\nTuesday 24 February 2026  00:45:27 +0000 (0:00:00.480)       0:00:40.308 ****** \n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\nTASK [Calculate VG sizes (without buffer)] *************************************\nTuesday 24 February 2026  00:45:27 +0000 (0:00:00.140)       0:00:40.448 ****** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Calculate VG sizes (with buffer)] ****************************************\nTuesday 24 February 2026  00:45:27 +0000 (0:00:00.104)       0:00:40.552 ****** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Print LVM VGs report data] ***********************************************\nTuesday 24 February 2026  00:45:27 +0000 (0:00:00.089)       0:00:40.642 ****** \n\u001b[0;32mok: [testbed-node-4] => {\u001b[0m\n\u001b[0;32m    \"vgs_report\": {\u001b[0m\n\u001b[0;32m        \"vg\": []\u001b[0m\n\u001b[0;32m    }\u001b[0m\n\u001b[0;32m}\u001b[0m\n\nTASK [Print LVM VG sizes] ******************************************************\nTuesday 24 February 2026  00:45:27 +0000 (0:00:00.113)       0:00:40.755 ****** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Calculate size needed for LVs on ceph_db_devices] ************************\nTuesday 24 February 2026  00:45:28 +0000 (0:00:00.120)       0:00:40.876 ****** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Print size needed for LVs on ceph_db_devices] ****************************\nTuesday 24 February 2026  00:45:28 +0000 (0:00:00.125)       0:00:41.001 ****** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Fail if size of DB LVs on ceph_db_devices > available] *******************\nTuesday 24 February 2026  00:45:28 +0000 (0:00:00.115)       0:00:41.117 ****** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Calculate size needed for LVs on ceph_wal_devices] ***********************\nTuesday 24 February 2026  00:45:28 +0000 (0:00:00.114)       0:00:41.231 ****** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Print size needed for LVs on ceph_wal_devices] ***************************\nTuesday 24 February 2026  00:45:28 +0000 (0:00:00.235)       0:00:41.467 ****** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Fail if size of WAL LVs on ceph_wal_devices > available] *****************\nTuesday 24 February 2026  00:45:28 +0000 (0:00:00.114)       0:00:41.581 ****** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Calculate size needed for WAL LVs on ceph_db_wal_devices] ****************\nTuesday 24 February 2026  00:45:28 +0000 (0:00:00.116)       0:00:41.697 ****** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Print size needed for WAL LVs on ceph_db_wal_devices] ********************\nTuesday 24 February 2026  00:45:28 +0000 (0:00:00.126)       0:00:41.824 ****** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Calculate size needed for DB LVs on ceph_db_wal_devices] *****************\nTuesday 24 February 2026  00:45:29 +0000 (0:00:00.117)       0:00:41.942 ****** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Print size needed for DB LVs on ceph_db_wal_devices] *********************\nTuesday 24 February 2026  00:45:29 +0000 (0:00:00.117)       0:00:42.060 ****** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Fail if size of DB+WAL LVs on ceph_db_wal_devices > available] ***********\nTuesday 24 February 2026  00:45:29 +0000 (0:00:00.126)       0:00:42.186 ****** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Fail if DB LV size < 30 GiB for ceph_db_devices] *************************\nTuesday 24 February 2026  00:45:29 +0000 (0:00:00.126)       0:00:42.313 ****** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Fail if DB LV size < 30 GiB for ceph_db_wal_devices] *********************\nTuesday 24 February 2026  00:45:29 +0000 (0:00:00.111)       0:00:42.424 ****** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Create DB LVs for ceph_db_devices] ***************************************\nTuesday 24 February 2026  00:45:29 +0000 (0:00:00.130)       0:00:42.554 ****** \n\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-294dd48f-9d6d-5079-8214-59e60af9e526', 'data_vg': 'ceph-294dd48f-9d6d-5079-8214-59e60af9e526'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-7fbf4d63-7c6f-5c70-887c-2b73c978775b', 'data_vg': 'ceph-7fbf4d63-7c6f-5c70-887c-2b73c978775b'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Print 'Create DB LVs for ceph_db_devices'] *******************************\nTuesday 24 February 2026  00:45:29 +0000 (0:00:00.140)       0:00:42.694 ****** \n\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-294dd48f-9d6d-5079-8214-59e60af9e526', 'data_vg': 'ceph-294dd48f-9d6d-5079-8214-59e60af9e526'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-7fbf4d63-7c6f-5c70-887c-2b73c978775b', 'data_vg': 'ceph-7fbf4d63-7c6f-5c70-887c-2b73c978775b'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Create WAL LVs for ceph_wal_devices] *************************************\nTuesday 24 February 2026  00:45:29 +0000 (0:00:00.135)       0:00:42.829 ****** \n\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-294dd48f-9d6d-5079-8214-59e60af9e526', 'data_vg': 'ceph-294dd48f-9d6d-5079-8214-59e60af9e526'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-7fbf4d63-7c6f-5c70-887c-2b73c978775b', 'data_vg': 'ceph-7fbf4d63-7c6f-5c70-887c-2b73c978775b'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Print 'Create WAL LVs for ceph_wal_devices'] *****************************\nTuesday 24 February 2026  00:45:30 +0000 (0:00:00.257)       0:00:43.087 ****** \n\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-294dd48f-9d6d-5079-8214-59e60af9e526', 'data_vg': 'ceph-294dd48f-9d6d-5079-8214-59e60af9e526'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-7fbf4d63-7c6f-5c70-887c-2b73c978775b', 'data_vg': 'ceph-7fbf4d63-7c6f-5c70-887c-2b73c978775b'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Create WAL LVs for ceph_db_wal_devices] **********************************\nTuesday 24 February 2026  00:45:30 +0000 (0:00:00.116)       0:00:43.204 ****** \n\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-294dd48f-9d6d-5079-8214-59e60af9e526', 'data_vg': 'ceph-294dd48f-9d6d-5079-8214-59e60af9e526'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-7fbf4d63-7c6f-5c70-887c-2b73c978775b', 'data_vg': 'ceph-7fbf4d63-7c6f-5c70-887c-2b73c978775b'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Print 'Create WAL LVs for ceph_db_wal_devices'] **************************\nTuesday 24 February 2026  00:45:30 +0000 (0:00:00.106)       0:00:43.310 ****** \n\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-294dd48f-9d6d-5079-8214-59e60af9e526', 'data_vg': 'ceph-294dd48f-9d6d-5079-8214-59e60af9e526'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-7fbf4d63-7c6f-5c70-887c-2b73c978775b', 'data_vg': 'ceph-7fbf4d63-7c6f-5c70-887c-2b73c978775b'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Create DB LVs for ceph_db_wal_devices] ***********************************\nTuesday 24 February 2026  00:45:30 +0000 (0:00:00.136)       0:00:43.447 ****** \n\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-294dd48f-9d6d-5079-8214-59e60af9e526', 'data_vg': 'ceph-294dd48f-9d6d-5079-8214-59e60af9e526'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-7fbf4d63-7c6f-5c70-887c-2b73c978775b', 'data_vg': 'ceph-7fbf4d63-7c6f-5c70-887c-2b73c978775b'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Print 'Create DB LVs for ceph_db_wal_devices'] ***************************\nTuesday 24 February 2026  00:45:30 +0000 (0:00:00.150)       0:00:43.597 ****** \n\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-294dd48f-9d6d-5079-8214-59e60af9e526', 'data_vg': 'ceph-294dd48f-9d6d-5079-8214-59e60af9e526'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-7fbf4d63-7c6f-5c70-887c-2b73c978775b', 'data_vg': 'ceph-7fbf4d63-7c6f-5c70-887c-2b73c978775b'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Get list of Ceph LVs with associated VGs] ********************************\nTuesday 24 February 2026  00:45:30 +0000 (0:00:00.133)       0:00:43.731 ****** \n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\nTASK [Get list of Ceph PVs with associated VGs] ********************************\nTuesday 24 February 2026  00:45:31 +0000 (0:00:00.485)       0:00:44.216 ****** \n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\nTASK [Combine JSON from _lvs_cmd_output/_pvs_cmd_output] ***********************\nTuesday 24 February 2026  00:45:31 +0000 (0:00:00.475)       0:00:44.692 ****** \n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\nTASK [Create list of VG/LV names] **********************************************\nTuesday 24 February 2026  00:45:32 +0000 (0:00:00.142)       0:00:44.834 ****** \n\u001b[0;32mok: [testbed-node-4] => (item={'lv_name': 'osd-block-294dd48f-9d6d-5079-8214-59e60af9e526', 'vg_name': 'ceph-294dd48f-9d6d-5079-8214-59e60af9e526'})\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item={'lv_name': 'osd-block-7fbf4d63-7c6f-5c70-887c-2b73c978775b', 'vg_name': 'ceph-7fbf4d63-7c6f-5c70-887c-2b73c978775b'})\u001b[0m\n\nTASK [Fail if block LV defined in lvm_volumes is missing] **********************\nTuesday 24 February 2026  00:45:32 +0000 (0:00:00.162)       0:00:44.996 ****** \n\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-294dd48f-9d6d-5079-8214-59e60af9e526', 'data_vg': 'ceph-294dd48f-9d6d-5079-8214-59e60af9e526'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-7fbf4d63-7c6f-5c70-887c-2b73c978775b', 'data_vg': 'ceph-7fbf4d63-7c6f-5c70-887c-2b73c978775b'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Fail if DB LV defined in lvm_volumes is missing] *************************\nTuesday 24 February 2026  00:45:32 +0000 (0:00:00.138)       0:00:45.135 ****** \n\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-294dd48f-9d6d-5079-8214-59e60af9e526', 'data_vg': 'ceph-294dd48f-9d6d-5079-8214-59e60af9e526'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-7fbf4d63-7c6f-5c70-887c-2b73c978775b', 'data_vg': 'ceph-7fbf4d63-7c6f-5c70-887c-2b73c978775b'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Fail if WAL LV defined in lvm_volumes is missing] ************************\nTuesday 24 February 2026  00:45:32 +0000 (0:00:00.147)       0:00:45.282 ****** \n\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-294dd48f-9d6d-5079-8214-59e60af9e526', 'data_vg': 'ceph-294dd48f-9d6d-5079-8214-59e60af9e526'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-7fbf4d63-7c6f-5c70-887c-2b73c978775b', 'data_vg': 'ceph-7fbf4d63-7c6f-5c70-887c-2b73c978775b'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [Print LVM report data] ***************************************************\nTuesday 24 February 2026  00:45:32 +0000 (0:00:00.169)       0:00:45.452 ****** \n\u001b[0;32mok: [testbed-node-4] => {\u001b[0m\n\u001b[0;32m    \"lvm_report\": {\u001b[0m\n\u001b[0;32m        \"lv\": [\u001b[0m\n\u001b[0;32m            {\u001b[0m\n\u001b[0;32m                \"lv_name\": \"osd-block-294dd48f-9d6d-5079-8214-59e60af9e526\",\u001b[0m\n\u001b[0;32m                \"vg_name\": \"ceph-294dd48f-9d6d-5079-8214-59e60af9e526\"\u001b[0m\n\u001b[0;32m            },\u001b[0m\n\u001b[0;32m            {\u001b[0m\n\u001b[0;32m                \"lv_name\": \"osd-block-7fbf4d63-7c6f-5c70-887c-2b73c978775b\",\u001b[0m\n\u001b[0;32m                \"vg_name\": \"ceph-7fbf4d63-7c6f-5c70-887c-2b73c978775b\"\u001b[0m\n\u001b[0;32m            }\u001b[0m\n\u001b[0;32m        ],\u001b[0m\n\u001b[0;32m        \"pv\": [\u001b[0m\n\u001b[0;32m            {\u001b[0m\n\u001b[0;32m                \"pv_name\": \"/dev/sdb\",\u001b[0m\n\u001b[0;32m                \"vg_name\": \"ceph-294dd48f-9d6d-5079-8214-59e60af9e526\"\u001b[0m\n\u001b[0;32m            },\u001b[0m\n\u001b[0;32m            {\u001b[0m\n\u001b[0;32m                \"pv_name\": \"/dev/sdc\",\u001b[0m\n\u001b[0;32m                \"vg_name\": \"ceph-7fbf4d63-7c6f-5c70-887c-2b73c978775b\"\u001b[0m\n\u001b[0;32m            }\u001b[0m\n\u001b[0;32m        ]\u001b[0m\n\u001b[0;32m    }\u001b[0m\n\u001b[0;32m}\u001b[0m\n\nPLAY [Ceph create LVM devices] *************************************************\n\nTASK [Get extra vars for Ceph configuration] ***********************************\nTuesday 24 February 2026  00:45:33 +0000 (0:00:00.394)       0:00:45.847 ****** \n\u001b[0;32mok: [testbed-node-5 -> testbed-manager(192.168.16.5)]\u001b[0m\n\nTASK [Get initial list of available block devices] *****************************\nTuesday 24 February 2026  00:45:33 +0000 (0:00:00.287)       0:00:46.134 ****** \n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nTuesday 24 February 2026  00:45:33 +0000 (0:00:00.236)       0:00:46.371 ****** \n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-5 => (item=loop0)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-5 => (item=loop1)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-5 => (item=loop2)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-5 => (item=loop3)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-5 => (item=loop4)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-5 => (item=loop5)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-5 => (item=loop6)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-5 => (item=loop7)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-5 => (item=sda)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-5 => (item=sdb)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-5 => (item=sdc)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-5 => (item=sdd)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-5 => (item=sr0)\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nTuesday 24 February 2026  00:45:33 +0000 (0:00:00.401)       0:00:46.773 ****** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nTuesday 24 February 2026  00:45:34 +0000 (0:00:00.196)       0:00:46.969 ****** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nTuesday 24 February 2026  00:45:34 +0000 (0:00:00.203)       0:00:47.173 ****** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nTuesday 24 February 2026  00:45:34 +0000 (0:00:00.224)       0:00:47.398 ****** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nTuesday 24 February 2026  00:45:34 +0000 (0:00:00.191)       0:00:47.590 ****** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nTuesday 24 February 2026  00:45:35 +0000 (0:00:00.607)       0:00:48.197 ****** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nTuesday 24 February 2026  00:45:35 +0000 (0:00:00.194)       0:00:48.391 ****** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nTuesday 24 February 2026  00:45:35 +0000 (0:00:00.208)       0:00:48.600 ****** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nTuesday 24 February 2026  00:45:35 +0000 (0:00:00.198)       0:00:48.798 ****** \n\u001b[0;32mok: [testbed-node-5] => (item=scsi-0QEMU_QEMU_HARDDISK_7a9487cb-ec9e-4508-9c5a-f692700fc5a4)\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item=scsi-SQEMU_QEMU_HARDDISK_7a9487cb-ec9e-4508-9c5a-f692700fc5a4)\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nTuesday 24 February 2026  00:45:36 +0000 (0:00:00.411)       0:00:49.210 ****** \n\u001b[0;32mok: [testbed-node-5] => (item=scsi-0QEMU_QEMU_HARDDISK_1558935b-4c66-4614-97e5-cda14abd2641)\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item=scsi-SQEMU_QEMU_HARDDISK_1558935b-4c66-4614-97e5-cda14abd2641)\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nTuesday 24 February 2026  00:45:36 +0000 (0:00:00.424)       0:00:49.634 ****** \n\u001b[0;32mok: [testbed-node-5] => (item=scsi-0QEMU_QEMU_HARDDISK_6f4643f0-329b-4ae2-97c7-d348648fd970)\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item=scsi-SQEMU_QEMU_HARDDISK_6f4643f0-329b-4ae2-97c7-d348648fd970)\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nTuesday 24 February 2026  00:45:37 +0000 (0:00:00.418)       0:00:50.053 ****** \n\u001b[0;32mok: [testbed-node-5] => (item=scsi-0QEMU_QEMU_HARDDISK_5f47789f-2a1a-442a-8a3c-644a4f92e5a8)\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item=scsi-SQEMU_QEMU_HARDDISK_5f47789f-2a1a-442a-8a3c-644a4f92e5a8)\u001b[0m\n\nTASK [Add known links to the list of available block devices] ******************\nTuesday 24 February 2026  00:45:37 +0000 (0:00:00.414)       0:00:50.468 ****** \n\u001b[0;32mok: [testbed-node-5] => (item=ata-QEMU_DVD-ROM_QM00001)\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nTuesday 24 February 2026  00:45:37 +0000 (0:00:00.318)       0:00:50.786 ****** \n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-5 => (item=loop0)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-5 => (item=loop1)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-5 => (item=loop2)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-5 => (item=loop3)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-5 => (item=loop4)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-5 => (item=loop5)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-5 => (item=loop6)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-5 => (item=loop7)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-5 => (item=sda)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-5 => (item=sdb)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-5 => (item=sdc)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-5 => (item=sdd)\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-5 => (item=sr0)\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nTuesday 24 February 2026  00:45:38 +0000 (0:00:00.421)       0:00:51.207 ****** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nTuesday 24 February 2026  00:45:38 +0000 (0:00:00.206)       0:00:51.413 ****** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nTuesday 24 February 2026  00:45:39 +0000 (0:00:00.595)       0:00:52.009 ****** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nTuesday 24 February 2026  00:45:39 +0000 (0:00:00.195)       0:00:52.204 ****** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nTuesday 24 February 2026  00:45:39 +0000 (0:00:00.190)       0:00:52.394 ****** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nTuesday 24 February 2026  00:45:39 +0000 (0:00:00.205)       0:00:52.600 ****** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nTuesday 24 February 2026  00:45:39 +0000 (0:00:00.193)       0:00:52.794 ****** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nTuesday 24 February 2026  00:45:40 +0000 (0:00:00.210)       0:00:53.004 ****** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nTuesday 24 February 2026  00:45:40 +0000 (0:00:00.192)       0:00:53.196 ****** \n\u001b[0;32mok: [testbed-node-5] => (item=sda1)\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item=sda14)\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item=sda15)\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item=sda16)\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nTuesday 24 February 2026  00:45:40 +0000 (0:00:00.631)       0:00:53.828 ****** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nTuesday 24 February 2026  00:45:41 +0000 (0:00:00.196)       0:00:54.024 ****** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nTuesday 24 February 2026  00:45:41 +0000 (0:00:00.200)       0:00:54.225 ****** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Add known partitions to the list of available block devices] *************\nTuesday 24 February 2026  00:45:41 +0000 (0:00:00.180)       0:00:54.406 ****** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Check whether ceph_db_wal_devices is used exclusively] *******************\nTuesday 24 February 2026  00:45:41 +0000 (0:00:00.194)       0:00:54.600 ****** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Create dict of block VGs -> PVs from ceph_osd_devices] *******************\nTuesday 24 February 2026  00:45:42 +0000 (0:00:00.297)       0:00:54.898 ****** \n\u001b[0;32mok: [testbed-node-5] => (item={'key': 'sdb', 'value': {'osd_lvm_uuid': '752a9ec4-4bfc-5551-82cd-92b802bdad97'}})\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item={'key': 'sdc', 'value': {'osd_lvm_uuid': 'ba85a0e3-2f57-51d5-a9f1-072f0e9a8deb'}})\u001b[0m\n\nTASK [Create block VGs] ********************************************************\nTuesday 24 February 2026  00:45:42 +0000 (0:00:00.200)       0:00:55.099 ****** \n\u001b[0;33mchanged: [testbed-node-5] => (item={'data': 'osd-block-752a9ec4-4bfc-5551-82cd-92b802bdad97', 'data_vg': 'ceph-752a9ec4-4bfc-5551-82cd-92b802bdad97'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'data': 'osd-block-ba85a0e3-2f57-51d5-a9f1-072f0e9a8deb', 'data_vg': 'ceph-ba85a0e3-2f57-51d5-a9f1-072f0e9a8deb'})\u001b[0m\n\nTASK [Print 'Create block VGs'] ************************************************\nTuesday 24 February 2026  00:45:44 +0000 (0:00:01.937)       0:00:57.036 ****** \n\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-752a9ec4-4bfc-5551-82cd-92b802bdad97', 'data_vg': 'ceph-752a9ec4-4bfc-5551-82cd-92b802bdad97'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-ba85a0e3-2f57-51d5-a9f1-072f0e9a8deb', 'data_vg': 'ceph-ba85a0e3-2f57-51d5-a9f1-072f0e9a8deb'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Create block LVs] ********************************************************\nTuesday 24 February 2026  00:45:44 +0000 (0:00:00.153)       0:00:57.190 ****** \n\u001b[0;33mchanged: [testbed-node-5] => (item={'data': 'osd-block-752a9ec4-4bfc-5551-82cd-92b802bdad97', 'data_vg': 'ceph-752a9ec4-4bfc-5551-82cd-92b802bdad97'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'data': 'osd-block-ba85a0e3-2f57-51d5-a9f1-072f0e9a8deb', 'data_vg': 'ceph-ba85a0e3-2f57-51d5-a9f1-072f0e9a8deb'})\u001b[0m\n\nTASK [Print 'Create block LVs'] ************************************************\nTuesday 24 February 2026  00:45:45 +0000 (0:00:01.349)       0:00:58.539 ****** \n\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-752a9ec4-4bfc-5551-82cd-92b802bdad97', 'data_vg': 'ceph-752a9ec4-4bfc-5551-82cd-92b802bdad97'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-ba85a0e3-2f57-51d5-a9f1-072f0e9a8deb', 'data_vg': 'ceph-ba85a0e3-2f57-51d5-a9f1-072f0e9a8deb'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Create DB VGs] ***********************************************************\nTuesday 24 February 2026  00:45:45 +0000 (0:00:00.146)       0:00:58.686 ****** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Print 'Create DB VGs'] ***************************************************\nTuesday 24 February 2026  00:45:45 +0000 (0:00:00.132)       0:00:58.819 ****** \n\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-752a9ec4-4bfc-5551-82cd-92b802bdad97', 'data_vg': 'ceph-752a9ec4-4bfc-5551-82cd-92b802bdad97'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-ba85a0e3-2f57-51d5-a9f1-072f0e9a8deb', 'data_vg': 'ceph-ba85a0e3-2f57-51d5-a9f1-072f0e9a8deb'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Create WAL VGs] **********************************************************\nTuesday 24 February 2026  00:45:46 +0000 (0:00:00.149)       0:00:58.968 ****** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Print 'Create WAL VGs'] **************************************************\nTuesday 24 February 2026  00:45:46 +0000 (0:00:00.145)       0:00:59.113 ****** \n\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-752a9ec4-4bfc-5551-82cd-92b802bdad97', 'data_vg': 'ceph-752a9ec4-4bfc-5551-82cd-92b802bdad97'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-ba85a0e3-2f57-51d5-a9f1-072f0e9a8deb', 'data_vg': 'ceph-ba85a0e3-2f57-51d5-a9f1-072f0e9a8deb'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Create DB+WAL VGs] *******************************************************\nTuesday 24 February 2026  00:45:46 +0000 (0:00:00.148)       0:00:59.262 ****** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Print 'Create DB+WAL VGs'] ***********************************************\nTuesday 24 February 2026  00:45:46 +0000 (0:00:00.136)       0:00:59.398 ****** \n\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-752a9ec4-4bfc-5551-82cd-92b802bdad97', 'data_vg': 'ceph-752a9ec4-4bfc-5551-82cd-92b802bdad97'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-ba85a0e3-2f57-51d5-a9f1-072f0e9a8deb', 'data_vg': 'ceph-ba85a0e3-2f57-51d5-a9f1-072f0e9a8deb'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Prepare variables for OSD count check] ***********************************\nTuesday 24 February 2026  00:45:46 +0000 (0:00:00.147)       0:00:59.546 ****** \n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [Count OSDs put on ceph_db_devices defined in lvm_volumes] ****************\nTuesday 24 February 2026  00:45:47 +0000 (0:00:00.305)       0:00:59.851 ****** \n\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-752a9ec4-4bfc-5551-82cd-92b802bdad97', 'data_vg': 'ceph-752a9ec4-4bfc-5551-82cd-92b802bdad97'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-ba85a0e3-2f57-51d5-a9f1-072f0e9a8deb', 'data_vg': 'ceph-ba85a0e3-2f57-51d5-a9f1-072f0e9a8deb'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Count OSDs put on ceph_wal_devices defined in lvm_volumes] ***************\nTuesday 24 February 2026  00:45:47 +0000 (0:00:00.153)       0:01:00.004 ****** \n\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-752a9ec4-4bfc-5551-82cd-92b802bdad97', 'data_vg': 'ceph-752a9ec4-4bfc-5551-82cd-92b802bdad97'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-ba85a0e3-2f57-51d5-a9f1-072f0e9a8deb', 'data_vg': 'ceph-ba85a0e3-2f57-51d5-a9f1-072f0e9a8deb'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Count OSDs put on ceph_db_wal_devices defined in lvm_volumes] ************\nTuesday 24 February 2026  00:45:47 +0000 (0:00:00.151)       0:01:00.156 ****** \n\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-752a9ec4-4bfc-5551-82cd-92b802bdad97', 'data_vg': 'ceph-752a9ec4-4bfc-5551-82cd-92b802bdad97'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-ba85a0e3-2f57-51d5-a9f1-072f0e9a8deb', 'data_vg': 'ceph-ba85a0e3-2f57-51d5-a9f1-072f0e9a8deb'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Fail if number of OSDs exceeds num_osds for a DB VG] *********************\nTuesday 24 February 2026  00:45:47 +0000 (0:00:00.168)       0:01:00.324 ****** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Fail if number of OSDs exceeds num_osds for a WAL VG] ********************\nTuesday 24 February 2026  00:45:47 +0000 (0:00:00.137)       0:01:00.461 ****** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Fail if number of OSDs exceeds num_osds for a DB+WAL VG] *****************\nTuesday 24 February 2026  00:45:47 +0000 (0:00:00.123)       0:01:00.585 ****** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Print number of OSDs wanted per DB VG] ***********************************\nTuesday 24 February 2026  00:45:47 +0000 (0:00:00.134)       0:01:00.719 ****** \n\u001b[0;32mok: [testbed-node-5] => {\u001b[0m\n\u001b[0;32m    \"_num_osds_wanted_per_db_vg\": {}\u001b[0m\n\u001b[0;32m}\u001b[0m\n\nTASK [Print number of OSDs wanted per WAL VG] **********************************\nTuesday 24 February 2026  00:45:48 +0000 (0:00:00.138)       0:01:00.858 ****** \n\u001b[0;32mok: [testbed-node-5] => {\u001b[0m\n\u001b[0;32m    \"_num_osds_wanted_per_wal_vg\": {}\u001b[0m\n\u001b[0;32m}\u001b[0m\n\nTASK [Print number of OSDs wanted per DB+WAL VG] *******************************\nTuesday 24 February 2026  00:45:48 +0000 (0:00:00.141)       0:01:00.999 ****** \n\u001b[0;32mok: [testbed-node-5] => {\u001b[0m\n\u001b[0;32m    \"_num_osds_wanted_per_db_wal_vg\": {}\u001b[0m\n\u001b[0;32m}\u001b[0m\n\nTASK [Gather DB VGs with total and available size in bytes] ********************\nTuesday 24 February 2026  00:45:48 +0000 (0:00:00.132)       0:01:01.132 ****** \n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [Gather WAL VGs with total and available size in bytes] *******************\nTuesday 24 February 2026  00:45:48 +0000 (0:00:00.526)       0:01:01.658 ****** \n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [Gather DB+WAL VGs with total and available size in bytes] ****************\nTuesday 24 February 2026  00:45:49 +0000 (0:00:00.540)       0:01:02.198 ****** \n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [Combine JSON from _db/wal/db_wal_vgs_cmd_output] *************************\nTuesday 24 February 2026  00:45:50 +0000 (0:00:00.706)       0:01:02.904 ****** \n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [Calculate VG sizes (without buffer)] *************************************\nTuesday 24 February 2026  00:45:50 +0000 (0:00:00.136)       0:01:03.041 ****** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Calculate VG sizes (with buffer)] ****************************************\nTuesday 24 February 2026  00:45:50 +0000 (0:00:00.091)       0:01:03.133 ****** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Print LVM VGs report data] ***********************************************\nTuesday 24 February 2026  00:45:50 +0000 (0:00:00.097)       0:01:03.231 ****** \n\u001b[0;32mok: [testbed-node-5] => {\u001b[0m\n\u001b[0;32m    \"vgs_report\": {\u001b[0m\n\u001b[0;32m        \"vg\": []\u001b[0m\n\u001b[0;32m    }\u001b[0m\n\u001b[0;32m}\u001b[0m\n\nTASK [Print LVM VG sizes] ******************************************************\nTuesday 24 February 2026  00:45:50 +0000 (0:00:00.127)       0:01:03.359 ****** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Calculate size needed for LVs on ceph_db_devices] ************************\nTuesday 24 February 2026  00:45:50 +0000 (0:00:00.121)       0:01:03.480 ****** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Print size needed for LVs on ceph_db_devices] ****************************\nTuesday 24 February 2026  00:45:50 +0000 (0:00:00.115)       0:01:03.596 ****** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Fail if size of DB LVs on ceph_db_devices > available] *******************\nTuesday 24 February 2026  00:45:50 +0000 (0:00:00.117)       0:01:03.714 ****** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Calculate size needed for LVs on ceph_wal_devices] ***********************\nTuesday 24 February 2026  00:45:51 +0000 (0:00:00.129)       0:01:03.843 ****** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Print size needed for LVs on ceph_wal_devices] ***************************\nTuesday 24 February 2026  00:45:51 +0000 (0:00:00.116)       0:01:03.959 ****** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Fail if size of WAL LVs on ceph_wal_devices > available] *****************\nTuesday 24 February 2026  00:45:51 +0000 (0:00:00.112)       0:01:04.072 ****** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Calculate size needed for WAL LVs on ceph_db_wal_devices] ****************\nTuesday 24 February 2026  00:45:51 +0000 (0:00:00.099)       0:01:04.172 ****** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Print size needed for WAL LVs on ceph_db_wal_devices] ********************\nTuesday 24 February 2026  00:45:51 +0000 (0:00:00.228)       0:01:04.401 ****** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Calculate size needed for DB LVs on ceph_db_wal_devices] *****************\nTuesday 24 February 2026  00:45:51 +0000 (0:00:00.123)       0:01:04.524 ****** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Print size needed for DB LVs on ceph_db_wal_devices] *********************\nTuesday 24 February 2026  00:45:51 +0000 (0:00:00.119)       0:01:04.644 ****** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Fail if size of DB+WAL LVs on ceph_db_wal_devices > available] ***********\nTuesday 24 February 2026  00:45:51 +0000 (0:00:00.123)       0:01:04.767 ****** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Fail if DB LV size < 30 GiB for ceph_db_devices] *************************\nTuesday 24 February 2026  00:45:52 +0000 (0:00:00.107)       0:01:04.875 ****** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Fail if DB LV size < 30 GiB for ceph_db_wal_devices] *********************\nTuesday 24 February 2026  00:45:52 +0000 (0:00:00.109)       0:01:04.985 ****** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Create DB LVs for ceph_db_devices] ***************************************\nTuesday 24 February 2026  00:45:52 +0000 (0:00:00.118)       0:01:05.104 ****** \n\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-752a9ec4-4bfc-5551-82cd-92b802bdad97', 'data_vg': 'ceph-752a9ec4-4bfc-5551-82cd-92b802bdad97'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-ba85a0e3-2f57-51d5-a9f1-072f0e9a8deb', 'data_vg': 'ceph-ba85a0e3-2f57-51d5-a9f1-072f0e9a8deb'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Print 'Create DB LVs for ceph_db_devices'] *******************************\nTuesday 24 February 2026  00:45:52 +0000 (0:00:00.128)       0:01:05.232 ****** \n\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-752a9ec4-4bfc-5551-82cd-92b802bdad97', 'data_vg': 'ceph-752a9ec4-4bfc-5551-82cd-92b802bdad97'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-ba85a0e3-2f57-51d5-a9f1-072f0e9a8deb', 'data_vg': 'ceph-ba85a0e3-2f57-51d5-a9f1-072f0e9a8deb'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Create WAL LVs for ceph_wal_devices] *************************************\nTuesday 24 February 2026  00:45:52 +0000 (0:00:00.123)       0:01:05.355 ****** \n\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-752a9ec4-4bfc-5551-82cd-92b802bdad97', 'data_vg': 'ceph-752a9ec4-4bfc-5551-82cd-92b802bdad97'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-ba85a0e3-2f57-51d5-a9f1-072f0e9a8deb', 'data_vg': 'ceph-ba85a0e3-2f57-51d5-a9f1-072f0e9a8deb'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Print 'Create WAL LVs for ceph_wal_devices'] *****************************\nTuesday 24 February 2026  00:45:52 +0000 (0:00:00.142)       0:01:05.497 ****** \n\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-752a9ec4-4bfc-5551-82cd-92b802bdad97', 'data_vg': 'ceph-752a9ec4-4bfc-5551-82cd-92b802bdad97'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-ba85a0e3-2f57-51d5-a9f1-072f0e9a8deb', 'data_vg': 'ceph-ba85a0e3-2f57-51d5-a9f1-072f0e9a8deb'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Create WAL LVs for ceph_db_wal_devices] **********************************\nTuesday 24 February 2026  00:45:52 +0000 (0:00:00.135)       0:01:05.632 ****** \n\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-752a9ec4-4bfc-5551-82cd-92b802bdad97', 'data_vg': 'ceph-752a9ec4-4bfc-5551-82cd-92b802bdad97'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-ba85a0e3-2f57-51d5-a9f1-072f0e9a8deb', 'data_vg': 'ceph-ba85a0e3-2f57-51d5-a9f1-072f0e9a8deb'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Print 'Create WAL LVs for ceph_db_wal_devices'] **************************\nTuesday 24 February 2026  00:45:52 +0000 (0:00:00.138)       0:01:05.771 ****** \n\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-752a9ec4-4bfc-5551-82cd-92b802bdad97', 'data_vg': 'ceph-752a9ec4-4bfc-5551-82cd-92b802bdad97'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-ba85a0e3-2f57-51d5-a9f1-072f0e9a8deb', 'data_vg': 'ceph-ba85a0e3-2f57-51d5-a9f1-072f0e9a8deb'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Create DB LVs for ceph_db_wal_devices] ***********************************\nTuesday 24 February 2026  00:45:53 +0000 (0:00:00.261)       0:01:06.033 ****** \n\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-752a9ec4-4bfc-5551-82cd-92b802bdad97', 'data_vg': 'ceph-752a9ec4-4bfc-5551-82cd-92b802bdad97'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-ba85a0e3-2f57-51d5-a9f1-072f0e9a8deb', 'data_vg': 'ceph-ba85a0e3-2f57-51d5-a9f1-072f0e9a8deb'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Print 'Create DB LVs for ceph_db_wal_devices'] ***************************\nTuesday 24 February 2026  00:45:53 +0000 (0:00:00.143)       0:01:06.176 ****** \n\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-752a9ec4-4bfc-5551-82cd-92b802bdad97', 'data_vg': 'ceph-752a9ec4-4bfc-5551-82cd-92b802bdad97'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-ba85a0e3-2f57-51d5-a9f1-072f0e9a8deb', 'data_vg': 'ceph-ba85a0e3-2f57-51d5-a9f1-072f0e9a8deb'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Get list of Ceph LVs with associated VGs] ********************************\nTuesday 24 February 2026  00:45:53 +0000 (0:00:00.153)       0:01:06.330 ****** \n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [Get list of Ceph PVs with associated VGs] ********************************\nTuesday 24 February 2026  00:45:54 +0000 (0:00:00.531)       0:01:06.862 ****** \n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [Combine JSON from _lvs_cmd_output/_pvs_cmd_output] ***********************\nTuesday 24 February 2026  00:45:54 +0000 (0:00:00.484)       0:01:07.347 ****** \n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [Create list of VG/LV names] **********************************************\nTuesday 24 February 2026  00:45:54 +0000 (0:00:00.120)       0:01:07.468 ****** \n\u001b[0;32mok: [testbed-node-5] => (item={'lv_name': 'osd-block-752a9ec4-4bfc-5551-82cd-92b802bdad97', 'vg_name': 'ceph-752a9ec4-4bfc-5551-82cd-92b802bdad97'})\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item={'lv_name': 'osd-block-ba85a0e3-2f57-51d5-a9f1-072f0e9a8deb', 'vg_name': 'ceph-ba85a0e3-2f57-51d5-a9f1-072f0e9a8deb'})\u001b[0m\n\nTASK [Fail if block LV defined in lvm_volumes is missing] **********************\nTuesday 24 February 2026  00:45:54 +0000 (0:00:00.145)       0:01:07.613 ****** \n\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-752a9ec4-4bfc-5551-82cd-92b802bdad97', 'data_vg': 'ceph-752a9ec4-4bfc-5551-82cd-92b802bdad97'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-ba85a0e3-2f57-51d5-a9f1-072f0e9a8deb', 'data_vg': 'ceph-ba85a0e3-2f57-51d5-a9f1-072f0e9a8deb'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Fail if DB LV defined in lvm_volumes is missing] *************************\nTuesday 24 February 2026  00:45:54 +0000 (0:00:00.138)       0:01:07.752 ****** \n\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-752a9ec4-4bfc-5551-82cd-92b802bdad97', 'data_vg': 'ceph-752a9ec4-4bfc-5551-82cd-92b802bdad97'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-ba85a0e3-2f57-51d5-a9f1-072f0e9a8deb', 'data_vg': 'ceph-ba85a0e3-2f57-51d5-a9f1-072f0e9a8deb'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Fail if WAL LV defined in lvm_volumes is missing] ************************\nTuesday 24 February 2026  00:45:55 +0000 (0:00:00.137)       0:01:07.890 ****** \n\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-752a9ec4-4bfc-5551-82cd-92b802bdad97', 'data_vg': 'ceph-752a9ec4-4bfc-5551-82cd-92b802bdad97'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-ba85a0e3-2f57-51d5-a9f1-072f0e9a8deb', 'data_vg': 'ceph-ba85a0e3-2f57-51d5-a9f1-072f0e9a8deb'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Print LVM report data] ***************************************************\nTuesday 24 February 2026  00:45:55 +0000 (0:00:00.153)       0:01:08.044 ****** \n\u001b[0;32mok: [testbed-node-5] => {\u001b[0m\n\u001b[0;32m    \"lvm_report\": {\u001b[0m\n\u001b[0;32m        \"lv\": [\u001b[0m\n\u001b[0;32m            {\u001b[0m\n\u001b[0;32m                \"lv_name\": \"osd-block-752a9ec4-4bfc-5551-82cd-92b802bdad97\",\u001b[0m\n\u001b[0;32m                \"vg_name\": \"ceph-752a9ec4-4bfc-5551-82cd-92b802bdad97\"\u001b[0m\n\u001b[0;32m            },\u001b[0m\n\u001b[0;32m            {\u001b[0m\n\u001b[0;32m                \"lv_name\": \"osd-block-ba85a0e3-2f57-51d5-a9f1-072f0e9a8deb\",\u001b[0m\n\u001b[0;32m                \"vg_name\": \"ceph-ba85a0e3-2f57-51d5-a9f1-072f0e9a8deb\"\u001b[0m\n\u001b[0;32m            }\u001b[0m\n\u001b[0;32m        ],\u001b[0m\n\u001b[0;32m        \"pv\": [\u001b[0m\n\u001b[0;32m            {\u001b[0m\n\u001b[0;32m                \"pv_name\": \"/dev/sdb\",\u001b[0m\n\u001b[0;32m                \"vg_name\": \"ceph-752a9ec4-4bfc-5551-82cd-92b802bdad97\"\u001b[0m\n\u001b[0;32m            },\u001b[0m\n\u001b[0;32m            {\u001b[0m\n\u001b[0;32m                \"pv_name\": \"/dev/sdc\",\u001b[0m\n\u001b[0;32m                \"vg_name\": \"ceph-ba85a0e3-2f57-51d5-a9f1-072f0e9a8deb\"\u001b[0m\n\u001b[0;32m            }\u001b[0m\n\u001b[0;32m        ]\u001b[0m\n\u001b[0;32m    }\u001b[0m\n\u001b[0;32m}\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;33mtestbed-node-3\u001b[0m             : \u001b[0;32mok=51  \u001b[0m \u001b[0;33mchanged=2   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=62  \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-4\u001b[0m             : \u001b[0;32mok=51  \u001b[0m \u001b[0;33mchanged=2   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=62  \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-5\u001b[0m             : \u001b[0;32mok=51  \u001b[0m \u001b[0;33mchanged=2   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=62  \u001b[0m rescued=0    ignored=0   \n\n\n\nTASKS RECAP ********************************************************************\nTuesday 24 February 2026  00:45:55 +0000 (0:00:00.135)       0:01:08.179 ****** \n=============================================================================== \nCreate block VGs -------------------------------------------------------- 5.75s\nCreate block LVs -------------------------------------------------------- 4.17s\nGather DB+WAL VGs with total and available size in bytes ---------------- 1.74s\nGather DB VGs with total and available size in bytes -------------------- 1.71s\nGather WAL VGs with total and available size in bytes ------------------- 1.56s\nGet list of Ceph LVs with associated VGs -------------------------------- 1.51s\nAdd known partitions to the list of available block devices ------------- 1.48s\nGet list of Ceph PVs with associated VGs -------------------------------- 1.46s\nAdd known links to the list of available block devices ------------------ 1.29s\nAdd known partitions to the list of available block devices ------------- 1.07s\nPrint LVM report data --------------------------------------------------- 0.78s\nGet extra vars for Ceph configuration ----------------------------------- 0.76s\nAdd known partitions to the list of available block devices ------------- 0.74s\nPrint number of OSDs wanted per DB VG ----------------------------------- 0.71s\nGet initial list of available block devices ----------------------------- 0.65s\nPrint 'Create DB VGs' --------------------------------------------------- 0.64s\nAdd known partitions to the list of available block devices ------------- 0.63s\nCreate DB LVs for ceph_db_devices --------------------------------------- 0.62s\nAdd known links to the list of available block devices ------------------ 0.61s\nAdd known partitions to the list of available block devices ------------- 0.60s\n\u001b[32m2026-02-24 00:46:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mPrepare task for execution of facts.\u001b[0m\n\u001b[32m2026-02-24 00:46:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask eb8126f1-e933-4fbf-9464-6ddb35899e54 (facts) was prepared for execution.\u001b[0m\n\u001b[32m2026-02-24 00:46:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task eb8126f1-e933-4fbf-9464-6ddb35899e54 (facts) has been started and output is visible here.\u001b[0m\n\nPLAY [Apply role facts] ********************************************************\n\nTASK [osism.commons.facts : Create custom facts directory] *********************\nTuesday 24 February 2026  00:46:12 +0000 (0:00:00.240)       0:00:00.240 ****** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [osism.commons.facts : Copy fact files] ***********************************\nTuesday 24 February 2026  00:46:13 +0000 (0:00:01.011)       0:00:01.252 ****** \n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nPLAY [Gather facts for all hosts] **********************************************\n\nTASK [Gathers facts about hosts] ***********************************************\nTuesday 24 February 2026  00:46:14 +0000 (0:00:01.065)       0:00:02.318 ****** \n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nPLAY [Gather facts for all hosts if using --limit] *****************************\n\nTASK [Gather facts for all hosts] **********************************************\nTuesday 24 February 2026  00:46:19 +0000 (0:00:04.998)       0:00:07.316 ****** \n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;32mtestbed-manager\u001b[0m            : \u001b[0;32mok=2   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   \n\u001b[0;32mtestbed-node-0\u001b[0m             : \u001b[0;32mok=2   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   \n\u001b[0;32mtestbed-node-1\u001b[0m             : \u001b[0;32mok=2   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   \n\u001b[0;32mtestbed-node-2\u001b[0m             : \u001b[0;32mok=2   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   \n\u001b[0;32mtestbed-node-3\u001b[0m             : \u001b[0;32mok=2   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   \n\u001b[0;32mtestbed-node-4\u001b[0m             : \u001b[0;32mok=2   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   \n\u001b[0;32mtestbed-node-5\u001b[0m             : \u001b[0;32mok=2   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nTuesday 24 February 2026  00:46:19 +0000 (0:00:00.481)       0:00:07.797 ****** \n=============================================================================== \nGathers facts about hosts ----------------------------------------------- 5.00s\nosism.commons.facts : Copy fact files ----------------------------------- 1.07s\nosism.commons.facts : Create custom facts directory --------------------- 1.01s\nGather facts for all hosts ---------------------------------------------- 0.48s\n\u001b[32m2026-02-24 00:46:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mPrepare task for execution of frr.\u001b[0m\n\u001b[32m2026-02-24 00:46:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 7e87fc91-2dc6-4330-b4c9-6b6256ab6f40 (frr) was prepared for execution.\u001b[0m\n\u001b[32m2026-02-24 00:46:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task 7e87fc91-2dc6-4330-b4c9-6b6256ab6f40 (frr) has been started and output is visible here.\u001b[0m\n\nPLAY [Apply role frr] **********************************************************\n\nTASK [osism.services.frr : Include distribution specific install tasks] ********\nTuesday 24 February 2026  00:46:36 +0000 (0:00:00.232)       0:00:00.232 ****** \n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/services/roles/frr/tasks/install-Debian-family.yml for testbed-manager\u001b[0m\n\nTASK [osism.services.frr : Pin frr package version] ****************************\nTuesday 24 February 2026  00:46:37 +0000 (0:00:00.239)       0:00:00.472 ****** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nTASK [osism.services.frr : Install frr package] ********************************\nTuesday 24 February 2026  00:46:38 +0000 (0:00:01.048)       0:00:01.520 ****** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nTASK [osism.services.frr : Copy file: /etc/frr/vtysh.conf] *********************\nTuesday 24 February 2026  00:46:46 +0000 (0:00:08.738)       0:00:10.258 ****** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\nTASK [osism.services.frr : Copy file: /etc/frr/daemons] ************************\nTuesday 24 February 2026  00:46:47 +0000 (0:00:01.082)       0:00:11.341 ****** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nTASK [osism.services.frr : Set _frr_uplinks fact] ******************************\nTuesday 24 February 2026  00:46:48 +0000 (0:00:00.968)       0:00:12.309 ****** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\nTASK [osism.services.frr : Write frr_config_template to temporary file] ********\nTuesday 24 February 2026  00:46:50 +0000 (0:00:01.179)       0:00:13.489 ****** \n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\nTASK [osism.services.frr : Render frr.conf from frr_config_template variable] ***\nTuesday 24 February 2026  00:46:50 +0000 (0:00:00.148)       0:00:13.638 ****** \n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\nTASK [osism.services.frr : Remove temporary frr_config_template file] **********\nTuesday 24 February 2026  00:46:50 +0000 (0:00:00.153)       0:00:13.792 ****** \n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\nTASK [osism.services.frr : Check for frr.conf file in the configuration repository] ***\nTuesday 24 February 2026  00:46:50 +0000 (0:00:00.155)       0:00:13.948 ****** \n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\nTASK [osism.services.frr : Copy frr.conf file from the configuration repository] ***\nTuesday 24 February 2026  00:46:50 +0000 (0:00:00.144)       0:00:14.092 ****** \n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\nTASK [osism.services.frr : Copy default frr.conf file of type k3s_cilium] ******\nTuesday 24 February 2026  00:46:50 +0000 (0:00:00.152)       0:00:14.244 ****** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nTASK [osism.services.frr : Set sysctl parameters] ******************************\nTuesday 24 February 2026  00:46:51 +0000 (0:00:01.113)       0:00:15.357 ****** \n\u001b[0;33mchanged: [testbed-manager] => (item={'name': 'net.ipv4.ip_forward', 'value': 1})\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item={'name': 'net.ipv4.conf.all.send_redirects', 'value': 0})\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item={'name': 'net.ipv4.conf.all.accept_redirects', 'value': 0})\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item={'name': 'net.ipv4.fib_multipath_hash_policy', 'value': 1})\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item={'name': 'net.ipv4.conf.default.ignore_routes_with_linkdown', 'value': 1})\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item={'name': 'net.ipv4.conf.all.rp_filter', 'value': 2})\u001b[0m\n\nTASK [osism.services.frr : Manage frr service] *********************************\nTuesday 24 February 2026  00:46:54 +0000 (0:00:02.262)       0:00:17.620 ****** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\nRUNNING HANDLER [osism.services.frr : Restart frr service] *********************\nTuesday 24 February 2026  00:46:55 +0000 (0:00:01.205)       0:00:18.825 ****** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;33mtestbed-manager\u001b[0m            : \u001b[0;32mok=10  \u001b[0m \u001b[0;33mchanged=6   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=5   \u001b[0m rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nTuesday 24 February 2026  00:46:56 +0000 (0:00:01.295)       0:00:20.121 ****** \n=============================================================================== \nosism.services.frr : Install frr package -------------------------------- 8.74s\nosism.services.frr : Set sysctl parameters ------------------------------ 2.26s\nosism.services.frr : Restart frr service -------------------------------- 1.30s\nosism.services.frr : Manage frr service --------------------------------- 1.21s\nosism.services.frr : Set _frr_uplinks fact ------------------------------ 1.18s\nosism.services.frr : Copy default frr.conf file of type k3s_cilium ------ 1.11s\nosism.services.frr : Copy file: /etc/frr/vtysh.conf --------------------- 1.08s\nosism.services.frr : Pin frr package version ---------------------------- 1.05s\nosism.services.frr : Copy file: /etc/frr/daemons ------------------------ 0.97s\nosism.services.frr : Include distribution specific install tasks -------- 0.24s\nosism.services.frr : Remove temporary frr_config_template file ---------- 0.16s\nosism.services.frr : Render frr.conf from frr_config_template variable --- 0.15s\nosism.services.frr : Copy frr.conf file from the configuration repository --- 0.15s\nosism.services.frr : Write frr_config_template to temporary file -------- 0.15s\nosism.services.frr : Check for frr.conf file in the configuration repository --- 0.14s\n\n--> DEPLOY IN A NUTSHELL -- START -- Tue Feb 24 00:46:57 UTC 2026\n\n\u001b[32m2026-02-24 00:46:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mCollection nutshell is prepared for execution\u001b[0m\n\u001b[32m2026-02-24 00:46:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mA [0] - dotfiles\u001b[0m\n\u001b[32m2026-02-24 00:47:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mA [0] - homer\u001b[0m\n\u001b[32m2026-02-24 00:47:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mA [0] - netdata\u001b[0m\n\u001b[32m2026-02-24 00:47:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mA [0] - openstackclient\u001b[0m\n\u001b[32m2026-02-24 00:47:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mA [0] - phpmyadmin\u001b[0m\n\u001b[32m2026-02-24 00:47:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mA [0] - common\u001b[0m\n\u001b[32m2026-02-24 00:47:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mA [1] -- loadbalancer\u001b[0m\n\u001b[32m2026-02-24 00:47:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mA [2] --- opensearch\u001b[0m\n\u001b[32m2026-02-24 00:47:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mA [2] --- mariadb-ng\u001b[0m\n\u001b[32m2026-02-24 00:47:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mA [3] ---- horizon\u001b[0m\n\u001b[32m2026-02-24 00:47:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mA [3] ---- keystone\u001b[0m\n\u001b[32m2026-02-24 00:47:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mA [4] ----- neutron\u001b[0m\n\u001b[32m2026-02-24 00:47:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mA [5] ------ wait-for-nova\u001b[0m\n\u001b[32m2026-02-24 00:47:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mA [6] ------- octavia\u001b[0m\n\u001b[32m2026-02-24 00:47:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mA [4] ----- barbican\u001b[0m\n\u001b[32m2026-02-24 00:47:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mA [4] ----- designate\u001b[0m\n\u001b[32m2026-02-24 00:47:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mA [4] ----- ironic\u001b[0m\n\u001b[32m2026-02-24 00:47:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mA [4] ----- placement\u001b[0m\n\u001b[32m2026-02-24 00:47:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mA [4] ----- magnum\u001b[0m\n\u001b[32m2026-02-24 00:47:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mA [1] -- openvswitch\u001b[0m\n\u001b[32m2026-02-24 00:47:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mA [2] --- ovn\u001b[0m\n\u001b[32m2026-02-24 00:47:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mA [1] -- memcached\u001b[0m\n\u001b[32m2026-02-24 00:47:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mA [1] -- redis\u001b[0m\n\u001b[32m2026-02-24 00:47:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mA [1] -- rabbitmq-ng\u001b[0m\n\u001b[32m2026-02-24 00:47:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mA [0] - kubernetes\u001b[0m\n\u001b[32m2026-02-24 00:47:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mA [1] -- kubeconfig\u001b[0m\n\u001b[32m2026-02-24 00:47:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mA [1] -- copy-kubeconfig\u001b[0m\n\u001b[32m2026-02-24 00:47:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mA [0] - ceph\u001b[0m\n\u001b[32m2026-02-24 00:47:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mA [1] -- ceph-pools\u001b[0m\n\u001b[32m2026-02-24 00:47:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mA [2] --- copy-ceph-keys\u001b[0m\n\u001b[32m2026-02-24 00:47:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mA [3] ---- cephclient\u001b[0m\n\u001b[32m2026-02-24 00:47:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mA [4] ----- ceph-bootstrap-dashboard\u001b[0m\n\u001b[32m2026-02-24 00:47:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mA [4] ----- wait-for-keystone\u001b[0m\n\u001b[32m2026-02-24 00:47:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mA [5] ------ kolla-ceph-rgw\u001b[0m\n\u001b[32m2026-02-24 00:47:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mA [5] ------ glance\u001b[0m\n\u001b[32m2026-02-24 00:47:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mA [5] ------ cinder\u001b[0m\n\u001b[32m2026-02-24 00:47:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mA [5] ------ nova\u001b[0m\n\u001b[32m2026-02-24 00:47:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mA [4] ----- prometheus\u001b[0m\n\u001b[32m2026-02-24 00:47:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mA [5] ------ grafana\u001b[0m\n\u001b[32m2026-02-24 00:47:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mAll tasks of the collection nutshell are prepared for execution\u001b[0m\n\u001b[32m2026-02-24 00:47:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTasks are running in the background\u001b[0m\n\u001b[32m2026-02-24 00:47:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mNo task IDs specified, wait for all currently running tasks\u001b[0m\n\u001b[32m2026-02-24 00:47:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ff4e45b7-3128-4ce1-95d9-909f33d92301 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:47:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f826982c-1ebc-4d5d-843f-4287d6696757 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:47:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f60b9a97-c893-4bee-b88b-e953f42f36db is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:47:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c72fddd9-9c8a-4ed1-9b6f-116f1787f9e8 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:47:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 991890e5-75ae-4ceb-aafe-9a43e3178763 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:47:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:47:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3803c886-08de-4d66-ab8f-1da8b65e33ca is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:47:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 00:47:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ff4e45b7-3128-4ce1-95d9-909f33d92301 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:47:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f826982c-1ebc-4d5d-843f-4287d6696757 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:47:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f60b9a97-c893-4bee-b88b-e953f42f36db is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:47:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c72fddd9-9c8a-4ed1-9b6f-116f1787f9e8 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:47:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 991890e5-75ae-4ceb-aafe-9a43e3178763 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:47:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:47:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3803c886-08de-4d66-ab8f-1da8b65e33ca is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:47:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 00:47:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ff4e45b7-3128-4ce1-95d9-909f33d92301 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:47:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f826982c-1ebc-4d5d-843f-4287d6696757 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:47:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f60b9a97-c893-4bee-b88b-e953f42f36db is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:47:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c72fddd9-9c8a-4ed1-9b6f-116f1787f9e8 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:47:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 991890e5-75ae-4ceb-aafe-9a43e3178763 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:47:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:47:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3803c886-08de-4d66-ab8f-1da8b65e33ca is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:47:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 00:47:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ff4e45b7-3128-4ce1-95d9-909f33d92301 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:47:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f826982c-1ebc-4d5d-843f-4287d6696757 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:47:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f60b9a97-c893-4bee-b88b-e953f42f36db is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:47:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c72fddd9-9c8a-4ed1-9b6f-116f1787f9e8 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:47:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 991890e5-75ae-4ceb-aafe-9a43e3178763 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:47:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:47:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3803c886-08de-4d66-ab8f-1da8b65e33ca is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:47:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 00:47:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ff4e45b7-3128-4ce1-95d9-909f33d92301 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:47:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f826982c-1ebc-4d5d-843f-4287d6696757 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:47:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f60b9a97-c893-4bee-b88b-e953f42f36db is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:47:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c72fddd9-9c8a-4ed1-9b6f-116f1787f9e8 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:47:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 991890e5-75ae-4ceb-aafe-9a43e3178763 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:47:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:47:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3803c886-08de-4d66-ab8f-1da8b65e33ca is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:47:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 00:47:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ff4e45b7-3128-4ce1-95d9-909f33d92301 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:47:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f826982c-1ebc-4d5d-843f-4287d6696757 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:47:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f60b9a97-c893-4bee-b88b-e953f42f36db is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:47:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c72fddd9-9c8a-4ed1-9b6f-116f1787f9e8 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:47:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 991890e5-75ae-4ceb-aafe-9a43e3178763 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:47:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:47:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3803c886-08de-4d66-ab8f-1da8b65e33ca is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:47:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 00:47:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ff4e45b7-3128-4ce1-95d9-909f33d92301 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:47:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f826982c-1ebc-4d5d-843f-4287d6696757 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:47:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f60b9a97-c893-4bee-b88b-e953f42f36db is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:47:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c72fddd9-9c8a-4ed1-9b6f-116f1787f9e8 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:47:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 991890e5-75ae-4ceb-aafe-9a43e3178763 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:47:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:47:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3803c886-08de-4d66-ab8f-1da8b65e33ca is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:47:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 00:47:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ff4e45b7-3128-4ce1-95d9-909f33d92301 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:47:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f826982c-1ebc-4d5d-843f-4287d6696757 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:47:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f60b9a97-c893-4bee-b88b-e953f42f36db is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:47:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c72fddd9-9c8a-4ed1-9b6f-116f1787f9e8 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:47:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 991890e5-75ae-4ceb-aafe-9a43e3178763 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:47:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:47:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3803c886-08de-4d66-ab8f-1da8b65e33ca is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:47:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\nPLAY [Apply role geerlingguy.dotfiles] *****************************************\n\nTASK [geerlingguy.dotfiles : Ensure dotfiles repository is cloned locally.] ****\nTuesday 24 February 2026  00:47:23 +0000 (0:00:00.524)       0:00:00.524 ****** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nTASK [geerlingguy.dotfiles : Ensure all configured dotfiles are links.] ********\nTuesday 24 February 2026  00:47:27 +0000 (0:00:03.858)       0:00:04.382 ****** \n\u001b[0;32mok: [testbed-manager] => (item=.tmux.conf)\u001b[0m\n\u001b[0;32mok: [testbed-node-0] => (item=.tmux.conf)\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item=.tmux.conf)\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item=.tmux.conf)\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item=.tmux.conf)\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item=.tmux.conf)\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item=.tmux.conf)\u001b[0m\n\nTASK [geerlingguy.dotfiles : Remove existing dotfiles file if a replacement is being linked.] ***\nTuesday 24 February 2026  00:47:29 +0000 (0:00:01.831)       0:00:06.214 ****** \n\u001b[0;32mok: [testbed-manager] => (item=[0, {'changed': False, 'stdout': '', 'stderr': \"ls: cannot access '/home/dragon/.tmux.conf': No such file or directory\", 'rc': 2, 'cmd': ['ls', '-F', '~/.tmux.conf'], 'start': '2026-02-24 00:47:28.113823', 'end': '2026-02-24 00:47:28.117993', 'delta': '0:00:00.004170', 'failed': False, 'msg': 'non-zero return code', 'invocation': {'module_args': {'_raw_params': 'ls -F ~/.tmux.conf', '_uses_shell': False, 'expand_argument_vars': True, 'stdin_add_newline': True, 'strip_empty_ends': True, 'argv': None, 'chdir': None, 'executable': None, 'creates': None, 'removes': None, 'stdin': None}}, 'stdout_lines': [], 'stderr_lines': [\"ls: cannot access '/home/dragon/.tmux.conf': No such file or directory\"], 'failed_when_result': False, 'item': '.tmux.conf', 'ansible_loop_var': 'item'}])\u001b[0m\n\u001b[0;32mok: [testbed-node-0] => (item=[0, {'changed': False, 'stdout': '', 'stderr': \"ls: cannot access '/home/dragon/.tmux.conf': No such file or directory\", 'rc': 2, 'cmd': ['ls', '-F', '~/.tmux.conf'], 'start': '2026-02-24 00:47:28.319974', 'end': '2026-02-24 00:47:28.328652', 'delta': '0:00:00.008678', 'failed': False, 'msg': 'non-zero return code', 'invocation': {'module_args': {'_raw_params': 'ls -F ~/.tmux.conf', '_uses_shell': False, 'expand_argument_vars': True, 'stdin_add_newline': True, 'strip_empty_ends': True, 'argv': None, 'chdir': None, 'executable': None, 'creates': None, 'removes': None, 'stdin': None}}, 'stdout_lines': [], 'stderr_lines': [\"ls: cannot access '/home/dragon/.tmux.conf': No such file or directory\"], 'failed_when_result': False, 'item': '.tmux.conf', 'ansible_loop_var': 'item'}])\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item=[0, {'changed': False, 'stdout': '', 'stderr': \"ls: cannot access '/home/dragon/.tmux.conf': No such file or directory\", 'rc': 2, 'cmd': ['ls', '-F', '~/.tmux.conf'], 'start': '2026-02-24 00:47:28.608254', 'end': '2026-02-24 00:47:28.617681', 'delta': '0:00:00.009427', 'failed': False, 'msg': 'non-zero return code', 'invocation': {'module_args': {'_raw_params': 'ls -F ~/.tmux.conf', '_uses_shell': False, 'expand_argument_vars': True, 'stdin_add_newline': True, 'strip_empty_ends': True, 'argv': None, 'chdir': None, 'executable': None, 'creates': None, 'removes': None, 'stdin': None}}, 'stdout_lines': [], 'stderr_lines': [\"ls: cannot access '/home/dragon/.tmux.conf': No such file or directory\"], 'failed_when_result': False, 'item': '.tmux.conf', 'ansible_loop_var': 'item'}])\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item=[0, {'changed': False, 'stdout': '', 'stderr': \"ls: cannot access '/home/dragon/.tmux.conf': No such file or directory\", 'rc': 2, 'cmd': ['ls', '-F', '~/.tmux.conf'], 'start': '2026-02-24 00:47:28.831013', 'end': '2026-02-24 00:47:28.838516', 'delta': '0:00:00.007503', 'failed': False, 'msg': 'non-zero return code', 'invocation': {'module_args': {'_raw_params': 'ls -F ~/.tmux.conf', '_uses_shell': False, 'expand_argument_vars': True, 'stdin_add_newline': True, 'strip_empty_ends': True, 'argv': None, 'chdir': None, 'executable': None, 'creates': None, 'removes': None, 'stdin': None}}, 'stdout_lines': [], 'stderr_lines': [\"ls: cannot access '/home/dragon/.tmux.conf': No such file or directory\"], 'failed_when_result': False, 'item': '.tmux.conf', 'ansible_loop_var': 'item'}])\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item=[0, {'changed': False, 'stdout': '', 'stderr': \"ls: cannot access '/home/dragon/.tmux.conf': No such file or directory\", 'rc': 2, 'cmd': ['ls', '-F', '~/.tmux.conf'], 'start': '2026-02-24 00:47:29.019007', 'end': '2026-02-24 00:47:29.030269', 'delta': '0:00:00.011262', 'failed': False, 'msg': 'non-zero return code', 'invocation': {'module_args': {'_raw_params': 'ls -F ~/.tmux.conf', '_uses_shell': False, 'expand_argument_vars': True, 'stdin_add_newline': True, 'strip_empty_ends': True, 'argv': None, 'chdir': None, 'executable': None, 'creates': None, 'removes': None, 'stdin': None}}, 'stdout_lines': [], 'stderr_lines': [\"ls: cannot access '/home/dragon/.tmux.conf': No such file or directory\"], 'failed_when_result': False, 'item': '.tmux.conf', 'ansible_loop_var': 'item'}])\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item=[0, {'changed': False, 'stdout': '', 'stderr': \"ls: cannot access '/home/dragon/.tmux.conf': No such file or directory\", 'rc': 2, 'cmd': ['ls', '-F', '~/.tmux.conf'], 'start': '2026-02-24 00:47:29.029611', 'end': '2026-02-24 00:47:29.037746', 'delta': '0:00:00.008135', 'failed': False, 'msg': 'non-zero return code', 'invocation': {'module_args': {'_raw_params': 'ls -F ~/.tmux.conf', '_uses_shell': False, 'expand_argument_vars': True, 'stdin_add_newline': True, 'strip_empty_ends': True, 'argv': None, 'chdir': None, 'executable': None, 'creates': None, 'removes': None, 'stdin': None}}, 'stdout_lines': [], 'stderr_lines': [\"ls: cannot access '/home/dragon/.tmux.conf': No such file or directory\"], 'failed_when_result': False, 'item': '.tmux.conf', 'ansible_loop_var': 'item'}])\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item=[0, {'changed': False, 'stdout': '', 'stderr': \"ls: cannot access '/home/dragon/.tmux.conf': No such file or directory\", 'rc': 2, 'cmd': ['ls', '-F', '~/.tmux.conf'], 'start': '2026-02-24 00:47:29.083881', 'end': '2026-02-24 00:47:29.090232', 'delta': '0:00:00.006351', 'failed': False, 'msg': 'non-zero return code', 'invocation': {'module_args': {'_raw_params': 'ls -F ~/.tmux.conf', '_uses_shell': False, 'expand_argument_vars': True, 'stdin_add_newline': True, 'strip_empty_ends': True, 'argv': None, 'chdir': None, 'executable': None, 'creates': None, 'removes': None, 'stdin': None}}, 'stdout_lines': [], 'stderr_lines': [\"ls: cannot access '/home/dragon/.tmux.conf': No such file or directory\"], 'failed_when_result': False, 'item': '.tmux.conf', 'ansible_loop_var': 'item'}])\u001b[0m\n\nTASK [geerlingguy.dotfiles : Ensure parent folders of link dotfiles exist.] ****\nTuesday 24 February 2026  00:47:32 +0000 (0:00:02.917)       0:00:09.131 ****** \n\u001b[0;32mok: [testbed-node-0] => (item=.tmux.conf)\u001b[0m\n\u001b[0;32mok: [testbed-manager] => (item=.tmux.conf)\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item=.tmux.conf)\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item=.tmux.conf)\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item=.tmux.conf)\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item=.tmux.conf)\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item=.tmux.conf)\u001b[0m\n\nTASK [geerlingguy.dotfiles : Link dotfiles into home folder.] ******************\nTuesday 24 February 2026  00:47:34 +0000 (0:00:01.821)       0:00:10.953 ****** \n\u001b[0;33mchanged: [testbed-manager] => (item=.tmux.conf)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=.tmux.conf)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=.tmux.conf)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=.tmux.conf)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item=.tmux.conf)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=.tmux.conf)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=.tmux.conf)\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;33mtestbed-manager\u001b[0m            : \u001b[0;32mok=5   \u001b[0m \u001b[0;33mchanged=2   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=5   \u001b[0m \u001b[0;33mchanged=2   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=5   \u001b[0m \u001b[0;33mchanged=2   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=5   \u001b[0m \u001b[0;33mchanged=2   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-3\u001b[0m             : \u001b[0;32mok=5   \u001b[0m \u001b[0;33mchanged=2   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-4\u001b[0m             : \u001b[0;32mok=5   \u001b[0m \u001b[0;33mchanged=2   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-5\u001b[0m             : \u001b[0;32mok=5   \u001b[0m \u001b[0;33mchanged=2   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nTuesday 24 February 2026  00:47:35 +0000 (0:00:01.862)       0:00:12.816 ****** \n=============================================================================== \ngeerlingguy.dotfiles : Ensure dotfiles repository is cloned locally. ---- 3.86s\ngeerlingguy.dotfiles : Remove existing dotfiles file if a replacement is being linked. --- 2.92s\ngeerlingguy.dotfiles : Link dotfiles into home folder. ------------------ 1.86s\ngeerlingguy.dotfiles : Ensure all configured dotfiles are links. -------- 1.83s\ngeerlingguy.dotfiles : Ensure parent folders of link dotfiles exist. ---- 1.82s\n\u001b[32m2026-02-24 00:47:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ff4e45b7-3128-4ce1-95d9-909f33d92301 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:47:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f826982c-1ebc-4d5d-843f-4287d6696757 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:47:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f60b9a97-c893-4bee-b88b-e953f42f36db is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:47:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask da7eb094-0467-4cde-920c-d33e5527beb2 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:47:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c72fddd9-9c8a-4ed1-9b6f-116f1787f9e8 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:47:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 991890e5-75ae-4ceb-aafe-9a43e3178763 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:47:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:47:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3803c886-08de-4d66-ab8f-1da8b65e33ca is in state SUCCESS\u001b[0m\n\u001b[32m2026-02-24 00:47:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 00:47:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ff4e45b7-3128-4ce1-95d9-909f33d92301 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:47:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f826982c-1ebc-4d5d-843f-4287d6696757 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:47:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f60b9a97-c893-4bee-b88b-e953f42f36db is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:47:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask da7eb094-0467-4cde-920c-d33e5527beb2 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:47:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c72fddd9-9c8a-4ed1-9b6f-116f1787f9e8 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:47:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 991890e5-75ae-4ceb-aafe-9a43e3178763 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:47:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:47:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 00:47:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ff4e45b7-3128-4ce1-95d9-909f33d92301 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:47:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f826982c-1ebc-4d5d-843f-4287d6696757 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:47:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f60b9a97-c893-4bee-b88b-e953f42f36db is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:47:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask da7eb094-0467-4cde-920c-d33e5527beb2 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:47:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c72fddd9-9c8a-4ed1-9b6f-116f1787f9e8 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:47:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 991890e5-75ae-4ceb-aafe-9a43e3178763 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:47:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:47:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 00:47:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ff4e45b7-3128-4ce1-95d9-909f33d92301 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:47:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f826982c-1ebc-4d5d-843f-4287d6696757 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:47:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f60b9a97-c893-4bee-b88b-e953f42f36db is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:47:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask da7eb094-0467-4cde-920c-d33e5527beb2 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:47:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c72fddd9-9c8a-4ed1-9b6f-116f1787f9e8 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:47:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 991890e5-75ae-4ceb-aafe-9a43e3178763 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:47:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:47:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 00:47:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ff4e45b7-3128-4ce1-95d9-909f33d92301 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:47:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f826982c-1ebc-4d5d-843f-4287d6696757 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:47:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f60b9a97-c893-4bee-b88b-e953f42f36db is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:47:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask da7eb094-0467-4cde-920c-d33e5527beb2 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:47:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c72fddd9-9c8a-4ed1-9b6f-116f1787f9e8 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:47:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 991890e5-75ae-4ceb-aafe-9a43e3178763 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:47:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:47:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 00:47:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ff4e45b7-3128-4ce1-95d9-909f33d92301 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:47:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f826982c-1ebc-4d5d-843f-4287d6696757 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:47:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f60b9a97-c893-4bee-b88b-e953f42f36db is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:47:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask da7eb094-0467-4cde-920c-d33e5527beb2 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:47:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c72fddd9-9c8a-4ed1-9b6f-116f1787f9e8 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:47:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 991890e5-75ae-4ceb-aafe-9a43e3178763 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:47:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:47:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 00:47:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ff4e45b7-3128-4ce1-95d9-909f33d92301 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:47:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f826982c-1ebc-4d5d-843f-4287d6696757 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:47:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f60b9a97-c893-4bee-b88b-e953f42f36db is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:47:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask da7eb094-0467-4cde-920c-d33e5527beb2 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:47:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c72fddd9-9c8a-4ed1-9b6f-116f1787f9e8 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:47:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 991890e5-75ae-4ceb-aafe-9a43e3178763 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:47:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:47:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 00:48:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ff4e45b7-3128-4ce1-95d9-909f33d92301 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:48:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f826982c-1ebc-4d5d-843f-4287d6696757 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:48:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f60b9a97-c893-4bee-b88b-e953f42f36db is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:48:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask da7eb094-0467-4cde-920c-d33e5527beb2 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:48:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c72fddd9-9c8a-4ed1-9b6f-116f1787f9e8 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:48:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 991890e5-75ae-4ceb-aafe-9a43e3178763 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:48:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:48:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 00:48:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ff4e45b7-3128-4ce1-95d9-909f33d92301 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:48:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f826982c-1ebc-4d5d-843f-4287d6696757 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:48:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f60b9a97-c893-4bee-b88b-e953f42f36db is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:48:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask da7eb094-0467-4cde-920c-d33e5527beb2 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:48:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c72fddd9-9c8a-4ed1-9b6f-116f1787f9e8 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:48:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 991890e5-75ae-4ceb-aafe-9a43e3178763 is in state SUCCESS\u001b[0m\n\u001b[32m2026-02-24 00:48:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:48:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 00:48:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ff4e45b7-3128-4ce1-95d9-909f33d92301 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:48:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f826982c-1ebc-4d5d-843f-4287d6696757 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:48:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f60b9a97-c893-4bee-b88b-e953f42f36db is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:48:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask da7eb094-0467-4cde-920c-d33e5527beb2 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:48:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c72fddd9-9c8a-4ed1-9b6f-116f1787f9e8 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:48:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:48:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 00:48:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ff4e45b7-3128-4ce1-95d9-909f33d92301 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:48:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f826982c-1ebc-4d5d-843f-4287d6696757 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:48:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f60b9a97-c893-4bee-b88b-e953f42f36db is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:48:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask da7eb094-0467-4cde-920c-d33e5527beb2 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:48:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c72fddd9-9c8a-4ed1-9b6f-116f1787f9e8 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:48:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:48:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 00:48:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ff4e45b7-3128-4ce1-95d9-909f33d92301 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:48:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f826982c-1ebc-4d5d-843f-4287d6696757 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:48:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f60b9a97-c893-4bee-b88b-e953f42f36db is in state SUCCESS\u001b[0m\n\u001b[32m2026-02-24 00:48:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask da7eb094-0467-4cde-920c-d33e5527beb2 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:48:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c72fddd9-9c8a-4ed1-9b6f-116f1787f9e8 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:48:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:48:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 00:48:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ff4e45b7-3128-4ce1-95d9-909f33d92301 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:48:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f826982c-1ebc-4d5d-843f-4287d6696757 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:48:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask da7eb094-0467-4cde-920c-d33e5527beb2 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:48:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c72fddd9-9c8a-4ed1-9b6f-116f1787f9e8 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:48:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:48:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 00:48:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ff4e45b7-3128-4ce1-95d9-909f33d92301 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:48:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f826982c-1ebc-4d5d-843f-4287d6696757 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:48:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask da7eb094-0467-4cde-920c-d33e5527beb2 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:48:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c72fddd9-9c8a-4ed1-9b6f-116f1787f9e8 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:48:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:48:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 00:48:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ff4e45b7-3128-4ce1-95d9-909f33d92301 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:48:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f826982c-1ebc-4d5d-843f-4287d6696757 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:48:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask da7eb094-0467-4cde-920c-d33e5527beb2 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:48:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c72fddd9-9c8a-4ed1-9b6f-116f1787f9e8 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:48:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:48:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 00:48:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ff4e45b7-3128-4ce1-95d9-909f33d92301 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:48:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f826982c-1ebc-4d5d-843f-4287d6696757 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:48:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask da7eb094-0467-4cde-920c-d33e5527beb2 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:48:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c72fddd9-9c8a-4ed1-9b6f-116f1787f9e8 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:48:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:48:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 00:48:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ff4e45b7-3128-4ce1-95d9-909f33d92301 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:48:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f826982c-1ebc-4d5d-843f-4287d6696757 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:48:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask da7eb094-0467-4cde-920c-d33e5527beb2 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:48:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c72fddd9-9c8a-4ed1-9b6f-116f1787f9e8 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:48:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:48:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 00:48:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ff4e45b7-3128-4ce1-95d9-909f33d92301 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:48:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f826982c-1ebc-4d5d-843f-4287d6696757 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:48:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask da7eb094-0467-4cde-920c-d33e5527beb2 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:48:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c72fddd9-9c8a-4ed1-9b6f-116f1787f9e8 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:48:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:48:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 00:48:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ff4e45b7-3128-4ce1-95d9-909f33d92301 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:48:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f826982c-1ebc-4d5d-843f-4287d6696757 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:48:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask da7eb094-0467-4cde-920c-d33e5527beb2 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:48:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c72fddd9-9c8a-4ed1-9b6f-116f1787f9e8 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:48:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:48:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 00:48:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ff4e45b7-3128-4ce1-95d9-909f33d92301 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:48:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f826982c-1ebc-4d5d-843f-4287d6696757 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:48:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask da7eb094-0467-4cde-920c-d33e5527beb2 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:48:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c72fddd9-9c8a-4ed1-9b6f-116f1787f9e8 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:48:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:48:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 00:48:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ff4e45b7-3128-4ce1-95d9-909f33d92301 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:48:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f826982c-1ebc-4d5d-843f-4287d6696757 is in state STARTED\u001b[0m\n\n\nPLAY [Apply role homer] ********************************************************\n\nTASK [osism.services.homer : Inform about new parameter homer_url_opensearch_dashboards] ***\nTuesday 24 February 2026  00:47:23 +0000 (0:00:00.536)       0:00:00.536 ****** \n\u001b[0;32mok: [testbed-manager] => {\u001b[0m\n\u001b[0;32m    \"msg\": \"The support for the homer_url_kibana has been removed. Please use the homer_url_opensearch_dashboards parameter.\"\u001b[0m\n\u001b[0;32m}\u001b[0m\n\nTASK [osism.services.homer : Create traefik external network] ******************\nTuesday 24 February 2026  00:47:23 +0000 (0:00:00.145)       0:00:00.681 ****** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\nTASK [osism.services.homer : Create required directories] **********************\nTuesday 24 February 2026  00:47:26 +0000 (0:00:02.987)       0:00:03.669 ****** \n\u001b[0;33mchanged: [testbed-manager] => (item=/opt/homer/configuration)\u001b[0m\n\u001b[0;32mok: [testbed-manager] => (item=/opt/homer)\u001b[0m\n\nTASK [osism.services.homer : Copy config.yml configuration file] ***************\nTuesday 24 February 2026  00:47:28 +0000 (0:00:02.054)       0:00:05.723 ****** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nTASK [osism.services.homer : Copy docker-compose.yml file] *********************\nTuesday 24 February 2026  00:47:31 +0000 (0:00:03.272)       0:00:08.995 ****** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nTASK [osism.services.homer : Manage homer service] *****************************\nTuesday 24 February 2026  00:47:33 +0000 (0:00:01.655)       0:00:10.651 ****** \n\u001b[1;30mFAILED - RETRYING: [testbed-manager]: Manage homer service (10 retries left).\u001b[0m\n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\nRUNNING HANDLER [osism.services.homer : Restart homer service] *****************\nTuesday 24 February 2026  00:48:00 +0000 (0:00:27.336)       0:00:37.987 ****** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;33mtestbed-manager\u001b[0m            : \u001b[0;32mok=7   \u001b[0m \u001b[0;33mchanged=4   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nTuesday 24 February 2026  00:48:03 +0000 (0:00:02.318)       0:00:40.306 ****** \n=============================================================================== \nosism.services.homer : Manage homer service ---------------------------- 27.34s\nosism.services.homer : Copy config.yml configuration file --------------- 3.27s\nosism.services.homer : Create traefik external network ------------------ 2.99s\nosism.services.homer : Restart homer service ---------------------------- 2.32s\nosism.services.homer : Create required directories ---------------------- 2.05s\nosism.services.homer : Copy docker-compose.yml file --------------------- 1.66s\nosism.services.homer : Inform about new parameter homer_url_opensearch_dashboards --- 0.15s\n\n\nPLAY [Apply role openstackclient] **********************************************\n\nTASK [osism.services.openstackclient : Include tasks] **************************\nTuesday 24 February 2026  00:47:24 +0000 (0:00:00.850)       0:00:00.850 ****** \n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/services/roles/openstackclient/tasks/container-Debian-family.yml for testbed-manager\u001b[0m\n\nTASK [osism.services.openstackclient : Create required directories] ************\nTuesday 24 February 2026  00:47:25 +0000 (0:00:00.868)       0:00:01.718 ****** \n\u001b[0;33mchanged: [testbed-manager] => (item=/opt/configuration/environments/openstack)\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item=/opt/openstackclient/data)\u001b[0m\n\u001b[0;32mok: [testbed-manager] => (item=/opt/openstackclient)\u001b[0m\n\nTASK [osism.services.openstackclient : Copy docker-compose.yml file] ***********\nTuesday 24 February 2026  00:47:27 +0000 (0:00:02.235)       0:00:03.954 ****** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nTASK [osism.services.openstackclient : Manage openstackclient service] *********\nTuesday 24 February 2026  00:47:30 +0000 (0:00:03.090)       0:00:07.044 ****** \n\u001b[1;30mFAILED - RETRYING: [testbed-manager]: Manage openstackclient service (10 retries left).\u001b[0m\n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\nTASK [osism.services.openstackclient : Copy openstack wrapper script] **********\nTuesday 24 February 2026  00:48:03 +0000 (0:00:33.482)       0:00:40.527 ****** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nTASK [osism.services.openstackclient : Remove ospurge wrapper script] **********\nTuesday 24 February 2026  00:48:05 +0000 (0:00:02.038)       0:00:42.566 ****** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\nRUNNING HANDLER [osism.services.openstackclient : Restart openstackclient service] ***\nTuesday 24 February 2026  00:48:07 +0000 (0:00:01.616)       0:00:44.182 ****** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nRUNNING HANDLER [osism.services.openstackclient : Ensure that all containers are up] ***\nTuesday 24 February 2026  00:48:11 +0000 (0:00:03.591)       0:00:47.774 ****** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nRUNNING HANDLER [osism.services.openstackclient : Wait for an healthy service] ***\nTuesday 24 February 2026  00:48:12 +0000 (0:00:01.063)       0:00:48.837 ****** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nRUNNING HANDLER [osism.services.openstackclient : Copy bash completion script] ***\nTuesday 24 February 2026  00:48:12 +0000 (0:00:00.486)       0:00:49.323 ****** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;33mtestbed-manager\u001b[0m            : \u001b[0;32mok=10  \u001b[0m \u001b[0;33mchanged=6   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nTuesday 24 February 2026  00:48:13 +0000 (0:00:00.434)       0:00:49.758 ****** \n=============================================================================== \nosism.services.openstackclient : Manage openstackclient service -------- 33.48s\nosism.services.openstackclient : Restart openstackclient service -------- 3.59s\nosism.services.openstackclient : Copy docker-compose.yml file ----------- 3.09s\nosism.services.openstackclient : Create required directories ------------ 2.24s\nosism.services.openstackclient : Copy openstack wrapper script ---------- 2.04s\nosism.services.openstackclient : Remove ospurge wrapper script ---------- 1.62s\nosism.services.openstackclient : Ensure that all containers are up ------ 1.06s\nosism.services.openstackclient : Include tasks -------------------------- 0.87s\nosism.services.openstackclient : Wait for an healthy service ------------ 0.49s\nosism.services.openstackclient : Copy bash completion script ------------ 0.43s\n\n\nPLAY [Apply role phpmyadmin] ***************************************************\n\nTASK [osism.services.phpmyadmin : Create traefik external network] *************\nTuesday 24 February 2026  00:47:42 +0000 (0:00:00.401)       0:00:00.401 ****** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\nTASK [osism.services.phpmyadmin : Create required directories] *****************\nTuesday 24 February 2026  00:47:43 +0000 (0:00:01.010)       0:00:01.411 ****** \n\u001b[0;33mchanged: [testbed-manager] => (item=/opt/phpmyadmin)\u001b[0m\n\nTASK [osism.services.phpmyadmin : Copy docker-compose.yml file] ****************\nTuesday 24 February 2026  00:47:44 +0000 (0:00:00.538)       0:00:01.950 ****** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nTASK [osism.services.phpmyadmin : Manage phpmyadmin service] *******************\nTuesday 24 February 2026  00:47:45 +0000 (0:00:01.092)       0:00:03.042 ****** \n\u001b[1;30mFAILED - RETRYING: [testbed-manager]: Manage phpmyadmin service (10 retries left).\u001b[0m\n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\nRUNNING HANDLER [osism.services.phpmyadmin : Restart phpmyadmin service] *******\nTuesday 24 February 2026  00:48:36 +0000 (0:00:50.758)       0:00:53.800 ****** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;33mtestbed-manager\u001b[0m            : \u001b[0;32mok=5   \u001b[0m \u001b[0;33mchanged=3   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nTuesday 24 February 2026  00:48:39 +0000 (0:00:03.634)       0:00:57.435 ****** \n=============================================================================== \nosism.services.phpmyadmin : Manage phpmyadmin service ------------------ 50.76s\nosism.services.phpmyadmin : Restart phpmyadmin service ------------------ 3.63s\nosism.services.phpmyadmin : Copy docker-compose.yml file ---------------- 1.09s\nosism.services.phpmyadmin : Create traefik external network ------------- 1.01s\nosism.services.phpmyadmin : Create required directories ----------------- 0.54s\n\u001b[32m2026-02-24 00:48:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask da7eb094-0467-4cde-920c-d33e5527beb2 is in state SUCCESS\u001b[0m\n\u001b[32m2026-02-24 00:48:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c72fddd9-9c8a-4ed1-9b6f-116f1787f9e8 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:48:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:48:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 00:48:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ff4e45b7-3128-4ce1-95d9-909f33d92301 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:48:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f826982c-1ebc-4d5d-843f-4287d6696757 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:48:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c72fddd9-9c8a-4ed1-9b6f-116f1787f9e8 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:48:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:48:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 00:48:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ff4e45b7-3128-4ce1-95d9-909f33d92301 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:48:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f826982c-1ebc-4d5d-843f-4287d6696757 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:48:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c72fddd9-9c8a-4ed1-9b6f-116f1787f9e8 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:48:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:48:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 00:48:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ff4e45b7-3128-4ce1-95d9-909f33d92301 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:48:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f826982c-1ebc-4d5d-843f-4287d6696757 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:48:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c72fddd9-9c8a-4ed1-9b6f-116f1787f9e8 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:48:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:48:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 00:48:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ff4e45b7-3128-4ce1-95d9-909f33d92301 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:48:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f826982c-1ebc-4d5d-843f-4287d6696757 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:48:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c72fddd9-9c8a-4ed1-9b6f-116f1787f9e8 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:48:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:48:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 00:48:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ff4e45b7-3128-4ce1-95d9-909f33d92301 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:48:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f826982c-1ebc-4d5d-843f-4287d6696757 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:48:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c72fddd9-9c8a-4ed1-9b6f-116f1787f9e8 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:48:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:48:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\n\nPLAY [Group hosts based on configuration] **************************************\n\nTASK [Group hosts based on enabled services] ***********************************\nTuesday 24 February 2026  00:47:23 +0000 (0:00:00.461)       0:00:00.461 ****** \n\u001b[0;33mchanged: [testbed-manager] => (item=enable_netdata_True)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=enable_netdata_True)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=enable_netdata_True)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=enable_netdata_True)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item=enable_netdata_True)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=enable_netdata_True)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=enable_netdata_True)\u001b[0m\n\nPLAY [Apply role netdata] ******************************************************\n\nTASK [osism.services.netdata : Include distribution specific install tasks] ****\nTuesday 24 February 2026  00:47:25 +0000 (0:00:01.694)       0:00:02.156 ****** \n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/services/roles/netdata/tasks/install-Debian-family.yml for testbed-manager, testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m\n\nTASK [osism.services.netdata : Remove old architecture-dependent repository] ***\nTuesday 24 February 2026  00:47:26 +0000 (0:00:01.552)       0:00:03.709 ****** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [osism.services.netdata : Install apt-transport-https package] ************\nTuesday 24 February 2026  00:47:28 +0000 (0:00:02.134)       0:00:05.843 ****** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [osism.services.netdata : Add repository gpg key] *************************\nTuesday 24 February 2026  00:47:32 +0000 (0:00:03.965)       0:00:09.809 ****** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\nTASK [osism.services.netdata : Add repository] *********************************\nTuesday 24 February 2026  00:47:35 +0000 (0:00:02.722)       0:00:12.531 ****** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nTASK [osism.services.netdata : Install package netdata] ************************\nTuesday 24 February 2026  00:47:47 +0000 (0:00:11.477)       0:00:24.009 ****** \n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nTASK [osism.services.netdata : Include config tasks] ***************************\nTuesday 24 February 2026  00:48:29 +0000 (0:00:42.307)       0:01:06.316 ****** \n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/services/roles/netdata/tasks/config.yml for testbed-manager, testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m\n\nTASK [osism.services.netdata : Copy configuration files] ***********************\nTuesday 24 February 2026  00:48:30 +0000 (0:00:01.326)       0:01:07.643 ****** \n\u001b[0;33mchanged: [testbed-manager] => (item=netdata.conf)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=netdata.conf)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=netdata.conf)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=netdata.conf)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item=netdata.conf)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=netdata.conf)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=netdata.conf)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=stream.conf)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item=stream.conf)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=stream.conf)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=stream.conf)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=stream.conf)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=stream.conf)\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item=stream.conf)\u001b[0m\n\nTASK [osism.services.netdata : Retrieve /etc/netdata/.opt-out-from-anonymous-statistics status] ***\nTuesday 24 February 2026  00:48:36 +0000 (0:00:05.356)       0:01:12.999 ****** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [osism.services.netdata : Opt out from anonymous statistics] **************\nTuesday 24 February 2026  00:48:37 +0000 (0:00:01.346)       0:01:14.346 ****** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nTASK [osism.services.netdata : Add netdata user to docker group] ***************\nTuesday 24 February 2026  00:48:39 +0000 (0:00:02.002)       0:01:16.349 ****** \n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [osism.services.netdata : Manage service netdata] *************************\nTuesday 24 February 2026  00:48:41 +0000 (0:00:01.752)       0:01:18.101 ****** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\nTASK [osism.services.netdata : Include host type specific tasks] ***************\nTuesday 24 February 2026  00:48:43 +0000 (0:00:01.935)       0:01:20.036 ****** \n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/services/roles/netdata/tasks/server.yml for testbed-manager\u001b[0m\n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/services/roles/netdata/tasks/client.yml for testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m\n\nTASK [osism.services.netdata : Set sysctl vm.max_map_count parameter] **********\nTuesday 24 February 2026  00:48:44 +0000 (0:00:01.167)       0:01:21.204 ****** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nRUNNING HANDLER [osism.services.netdata : Restart service netdata] *************\nTuesday 24 February 2026  00:48:45 +0000 (0:00:01.736)       0:01:22.941 ****** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;33mtestbed-manager\u001b[0m            : \u001b[0;32mok=16  \u001b[0m \u001b[0;33mchanged=8   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=15  \u001b[0m \u001b[0;33mchanged=7   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=15  \u001b[0m \u001b[0;33mchanged=7   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=15  \u001b[0m \u001b[0;33mchanged=7   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-3\u001b[0m             : \u001b[0;32mok=15  \u001b[0m \u001b[0;33mchanged=7   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-4\u001b[0m             : \u001b[0;32mok=15  \u001b[0m \u001b[0;33mchanged=7   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-5\u001b[0m             : \u001b[0;32mok=15  \u001b[0m \u001b[0;33mchanged=7   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nTuesday 24 February 2026  00:48:57 +0000 (0:00:11.256)       0:01:34.197 ****** \n=============================================================================== \nosism.services.netdata : Install package netdata ----------------------- 42.31s\nosism.services.netdata : Add repository -------------------------------- 11.48s\nosism.services.netdata : Restart service netdata ----------------------- 11.26s\nosism.services.netdata : Copy configuration files ----------------------- 5.36s\nosism.services.netdata : Install apt-transport-https package ------------ 3.97s\nosism.services.netdata : Add repository gpg key ------------------------- 2.72s\nosism.services.netdata : Remove old architecture-dependent repository --- 2.13s\nosism.services.netdata : Opt out from anonymous statistics -------------- 2.00s\nosism.services.netdata : Manage service netdata ------------------------- 1.94s\nosism.services.netdata : Add netdata user to docker group --------------- 1.75s\nosism.services.netdata : Set sysctl vm.max_map_count parameter ---------- 1.74s\nGroup hosts based on enabled services ----------------------------------- 1.69s\nosism.services.netdata : Include distribution specific install tasks ---- 1.55s\nosism.services.netdata : Retrieve /etc/netdata/.opt-out-from-anonymous-statistics status --- 1.35s\nosism.services.netdata : Include config tasks --------------------------- 1.33s\nosism.services.netdata : Include host type specific tasks --------------- 1.17s\n\u001b[32m2026-02-24 00:48:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ff4e45b7-3128-4ce1-95d9-909f33d92301 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:48:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f826982c-1ebc-4d5d-843f-4287d6696757 is in state SUCCESS\u001b[0m\n\u001b[32m2026-02-24 00:48:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c72fddd9-9c8a-4ed1-9b6f-116f1787f9e8 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:48:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:48:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 00:49:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ff4e45b7-3128-4ce1-95d9-909f33d92301 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:49:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c72fddd9-9c8a-4ed1-9b6f-116f1787f9e8 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:49:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:49:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 00:49:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ff4e45b7-3128-4ce1-95d9-909f33d92301 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:49:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c72fddd9-9c8a-4ed1-9b6f-116f1787f9e8 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:49:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:49:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 00:49:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ff4e45b7-3128-4ce1-95d9-909f33d92301 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:49:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c72fddd9-9c8a-4ed1-9b6f-116f1787f9e8 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:49:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:49:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 00:49:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ff4e45b7-3128-4ce1-95d9-909f33d92301 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:49:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c72fddd9-9c8a-4ed1-9b6f-116f1787f9e8 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:49:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:49:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 00:49:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ff4e45b7-3128-4ce1-95d9-909f33d92301 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:49:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c72fddd9-9c8a-4ed1-9b6f-116f1787f9e8 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:49:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:49:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 00:49:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ff4e45b7-3128-4ce1-95d9-909f33d92301 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:49:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c72fddd9-9c8a-4ed1-9b6f-116f1787f9e8 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:49:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:49:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 00:49:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ff4e45b7-3128-4ce1-95d9-909f33d92301 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:49:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c72fddd9-9c8a-4ed1-9b6f-116f1787f9e8 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:49:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:49:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 00:49:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ff4e45b7-3128-4ce1-95d9-909f33d92301 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:49:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c72fddd9-9c8a-4ed1-9b6f-116f1787f9e8 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:49:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:49:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 00:49:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ff4e45b7-3128-4ce1-95d9-909f33d92301 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:49:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c72fddd9-9c8a-4ed1-9b6f-116f1787f9e8 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:49:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:49:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 00:49:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ff4e45b7-3128-4ce1-95d9-909f33d92301 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:49:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c72fddd9-9c8a-4ed1-9b6f-116f1787f9e8 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:49:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:49:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 00:49:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ff4e45b7-3128-4ce1-95d9-909f33d92301 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:49:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f8694a2a-a9eb-4540-948b-ced7e58054f8 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:49:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c72fddd9-9c8a-4ed1-9b6f-116f1787f9e8 is in state SUCCESS\u001b[0m\n\n\nPLAY [Apply role common] *******************************************************\n\nTASK [common : include_tasks] **************************************************\nTuesday 24 February 2026  00:47:14 +0000 (0:00:00.237)       0:00:00.238 ****** \n\u001b[0;36mincluded: /ansible/roles/common/tasks/deploy.yml for testbed-manager, testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m\n\nTASK [common : Ensuring config directories exist] ******************************\nTuesday 24 February 2026  00:47:15 +0000 (0:00:01.442)       0:00:01.680 ****** \n\u001b[0;33mchanged: [testbed-node-0] => (item=[{'service_name': 'cron'}, 'cron'])\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=[{'service_name': 'cron'}, 'cron'])\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item=[{'service_name': 'cron'}, 'cron'])\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=[{'service_name': 'fluentd'}, 'fluentd'])\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=[{'service_name': 'cron'}, 'cron'])\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item=[{'service_name': 'cron'}, 'cron'])\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=[{'service_name': 'cron'}, 'cron'])\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=[{'service_name': 'cron'}, 'cron'])\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=[{'service_name': 'fluentd'}, 'fluentd'])\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=[{'service_name': 'kolla-toolbox'}, 'kolla-toolbox'])\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item=[{'service_name': 'fluentd'}, 'fluentd'])\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=[{'service_name': 'fluentd'}, 'fluentd'])\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item=[{'service_name': 'fluentd'}, 'fluentd'])\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=[{'service_name': 'fluentd'}, 'fluentd'])\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=[{'service_name': 'fluentd'}, 'fluentd'])\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=[{'service_name': 'kolla-toolbox'}, 'kolla-toolbox'])\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item=[{'service_name': 'kolla-toolbox'}, 'kolla-toolbox'])\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item=[{'service_name': 'kolla-toolbox'}, 'kolla-toolbox'])\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=[{'service_name': 'kolla-toolbox'}, 'kolla-toolbox'])\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=[{'service_name': 'kolla-toolbox'}, 'kolla-toolbox'])\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=[{'service_name': 'kolla-toolbox'}, 'kolla-toolbox'])\u001b[0m\n\nTASK [common : include_tasks] **************************************************\nTuesday 24 February 2026  00:47:19 +0000 (0:00:04.142)       0:00:05.822 ****** \n\u001b[0;36mincluded: /ansible/roles/common/tasks/copy-certs.yml for testbed-manager, testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m\n\nTASK [service-cert-copy : common | Copying over extra CA certificates] *********\nTuesday 24 February 2026  00:47:21 +0000 (0:00:01.559)       0:00:07.382 ****** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2024.2', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2024.2', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2024.2', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2024.2', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2024.2', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2024.2', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2024.2', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2024.2', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2024.2', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2024.2', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2024.2', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2024.2', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2024.2', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2024.2', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2024.2', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2024.2', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2024.2', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2024.2', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2024.2', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2024.2', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2024.2', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\nTASK [service-cert-copy : common | Copying over backend internal TLS certificate] ***\nTuesday 24 February 2026  00:47:26 +0000 (0:00:05.066)       0:00:12.448 ****** \n\u001b[0;36mskipping: [testbed-manager] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2024.2', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2024.2', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2024.2', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2024.2', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2024.2', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2024.2', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2024.2', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2024.2', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2024.2', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2024.2', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2024.2', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2024.2', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2024.2', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2024.2', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2024.2', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2024.2', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2024.2', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2024.2', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2024.2', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2024.2', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2024.2', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [service-cert-copy : common | Copying over backend internal TLS key] ******\nTuesday 24 February 2026  00:47:28 +0000 (0:00:02.141)       0:00:14.590 ****** \n\u001b[0;36mskipping: [testbed-manager] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2024.2', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2024.2', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2024.2', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2024.2', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2024.2', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2024.2', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2024.2', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2024.2', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2024.2', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2024.2', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2024.2', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2024.2', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2024.2', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2024.2', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2024.2', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2024.2', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2024.2', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2024.2', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2024.2', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2024.2', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2024.2', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [common : Copying over /run subdirectories conf] **************************\nTuesday 24 February 2026  00:47:30 +0000 (0:00:02.241)       0:00:16.832 ****** \n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [common : Restart systemd-tmpfiles] ***************************************\nTuesday 24 February 2026  00:47:31 +0000 (0:00:00.783)       0:00:17.616 ****** \n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [common : Copying over config.json files for services] ********************\nTuesday 24 February 2026  00:47:32 +0000 (0:00:01.089)       0:00:18.705 ****** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2024.2', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2024.2', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2024.2', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2024.2', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2024.2', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2024.2', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2024.2', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2024.2', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2024.2', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2024.2', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2024.2', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2024.2', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2024.2', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2024.2', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2024.2', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2024.2', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2024.2', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2024.2', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2024.2', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2024.2', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2024.2', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\nTASK [common : Find custom fluentd input config files] *************************\nTuesday 24 February 2026  00:47:41 +0000 (0:00:08.714)       0:00:27.419 ****** \n\u001b[1;35m[WARNING]: Skipped\u001b[0m\n\u001b[1;35m'/opt/configuration/environments/kolla/files/overlays/fluentd/input' path due\u001b[0m\n\u001b[1;35mto this access issue:\u001b[0m\n\u001b[1;35m'/opt/configuration/environments/kolla/files/overlays/fluentd/input' is not a\u001b[0m\n\u001b[1;35mdirectory\u001b[0m\n\u001b[0;32mok: [testbed-manager -> localhost]\u001b[0m\n\nTASK [common : Find custom fluentd filter config files] ************************\nTuesday 24 February 2026  00:47:43 +0000 (0:00:01.752)       0:00:29.172 ****** \n\u001b[1;35m[WARNING]: Skipped\u001b[0m\n\u001b[1;35m'/opt/configuration/environments/kolla/files/overlays/fluentd/filter' path due\u001b[0m\n\u001b[1;35mto this access issue:\u001b[0m\n\u001b[1;35m'/opt/configuration/environments/kolla/files/overlays/fluentd/filter' is not a\u001b[0m\n\u001b[1;35mdirectory\u001b[0m\n\u001b[0;32mok: [testbed-manager -> localhost]\u001b[0m\n\nTASK [common : Find custom fluentd format config files] ************************\nTuesday 24 February 2026  00:47:44 +0000 (0:00:01.199)       0:00:30.371 ****** \n\u001b[1;35m[WARNING]: Skipped\u001b[0m\n\u001b[1;35m'/opt/configuration/environments/kolla/files/overlays/fluentd/format' path due\u001b[0m\n\u001b[1;35mto this access issue:\u001b[0m\n\u001b[1;35m'/opt/configuration/environments/kolla/files/overlays/fluentd/format' is not a\u001b[0m\n\u001b[1;35mdirectory\u001b[0m\n\u001b[0;32mok: [testbed-manager -> localhost]\u001b[0m\n\nTASK [common : Find custom fluentd output config files] ************************\nTuesday 24 February 2026  00:47:45 +0000 (0:00:00.791)       0:00:31.163 ****** \n\u001b[1;35m[WARNING]: Skipped\u001b[0m\n\u001b[1;35m'/opt/configuration/environments/kolla/files/overlays/fluentd/output' path due\u001b[0m\n\u001b[1;35mto this access issue:\u001b[0m\n\u001b[1;35m'/opt/configuration/environments/kolla/files/overlays/fluentd/output' is not a\u001b[0m\n\u001b[1;35mdirectory\u001b[0m\n\u001b[0;32mok: [testbed-manager -> localhost]\u001b[0m\n\nTASK [common : Copying over fluentd.conf] **************************************\nTuesday 24 February 2026  00:47:45 +0000 (0:00:00.679)       0:00:31.842 ****** \n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [common : Copying over cron logrotate config file] ************************\nTuesday 24 February 2026  00:47:50 +0000 (0:00:04.267)       0:00:36.110 ****** \n\u001b[0;33mchanged: [testbed-node-0] => (item=/ansible/roles/common/templates/cron-logrotate-global.conf.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item=/ansible/roles/common/templates/cron-logrotate-global.conf.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=/ansible/roles/common/templates/cron-logrotate-global.conf.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=/ansible/roles/common/templates/cron-logrotate-global.conf.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item=/ansible/roles/common/templates/cron-logrotate-global.conf.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=/ansible/roles/common/templates/cron-logrotate-global.conf.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=/ansible/roles/common/templates/cron-logrotate-global.conf.j2)\u001b[0m\n\nTASK [common : Ensure RabbitMQ Erlang cookie exists] ***************************\nTuesday 24 February 2026  00:47:54 +0000 (0:00:04.599)       0:00:40.709 ****** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [common : Ensuring config directories have correct owner and permission] ***\nTuesday 24 February 2026  00:47:57 +0000 (0:00:02.928)       0:00:43.638 ****** \n\u001b[0;32mok: [testbed-node-0] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2024.2', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2024.2', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;32mok: [testbed-manager] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2024.2', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-manager] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2024.2', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;32mok: [testbed-node-0] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2024.2', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2024.2', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2024.2', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2024.2', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2024.2', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2024.2', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2024.2', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2024.2', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2024.2', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2024.2', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2024.2', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2024.2', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;32mok: [testbed-manager] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2024.2', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2024.2', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2024.2', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2024.2', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2024.2', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\nTASK [common : Copy rabbitmq-env.conf to kolla toolbox] ************************\nTuesday 24 February 2026  00:48:01 +0000 (0:00:03.343)       0:00:46.981 ****** \n\u001b[0;33mchanged: [testbed-node-0] => (item=/ansible/roles/common/templates/rabbitmq-env.conf.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=/ansible/roles/common/templates/rabbitmq-env.conf.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item=/ansible/roles/common/templates/rabbitmq-env.conf.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=/ansible/roles/common/templates/rabbitmq-env.conf.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=/ansible/roles/common/templates/rabbitmq-env.conf.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item=/ansible/roles/common/templates/rabbitmq-env.conf.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=/ansible/roles/common/templates/rabbitmq-env.conf.j2)\u001b[0m\n\nTASK [common : Copy rabbitmq erl_inetrc to kolla toolbox] **********************\nTuesday 24 February 2026  00:48:03 +0000 (0:00:02.665)       0:00:49.647 ****** \n\u001b[0;33mchanged: [testbed-node-1] => (item=/ansible/roles/common/templates/erl_inetrc.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=/ansible/roles/common/templates/erl_inetrc.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=/ansible/roles/common/templates/erl_inetrc.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item=/ansible/roles/common/templates/erl_inetrc.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=/ansible/roles/common/templates/erl_inetrc.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=/ansible/roles/common/templates/erl_inetrc.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item=/ansible/roles/common/templates/erl_inetrc.j2)\u001b[0m\n\nTASK [common : Check common containers] ****************************************\nTuesday 24 February 2026  00:48:06 +0000 (0:00:02.582)       0:00:52.229 ****** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2024.2', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2024.2', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2024.2', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2024.2', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2024.2', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2024.2', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2024.2', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2024.2', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2024.2', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2024.2', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2024.2', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2024.2', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2024.2', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2024.2', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2024.2', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2024.2', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2024.2', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2024.2', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2024.2', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2024.2', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2024.2', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\nTASK [common : Creating log volume] ********************************************\nTuesday 24 February 2026  00:48:10 +0000 (0:00:04.623)       0:00:56.854 ****** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nTASK [common : Link kolla_logs volume to /var/log/kolla] ***********************\nTuesday 24 February 2026  00:48:13 +0000 (0:00:02.092)       0:00:58.946 ****** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nTASK [common : Flush handlers] *************************************************\nTuesday 24 February 2026  00:48:14 +0000 (0:00:01.316)       0:01:00.263 ****** \n\nTASK [common : Flush handlers] *************************************************\nTuesday 24 February 2026  00:48:14 +0000 (0:00:00.064)       0:01:00.327 ****** \n\nTASK [common : Flush handlers] *************************************************\nTuesday 24 February 2026  00:48:14 +0000 (0:00:00.061)       0:01:00.388 ****** \n\nTASK [common : Flush handlers] *************************************************\nTuesday 24 February 2026  00:48:14 +0000 (0:00:00.229)       0:01:00.618 ****** \n\nTASK [common : Flush handlers] *************************************************\nTuesday 24 February 2026  00:48:14 +0000 (0:00:00.075)       0:01:00.693 ****** \n\nTASK [common : Flush handlers] *************************************************\nTuesday 24 February 2026  00:48:14 +0000 (0:00:00.066)       0:01:00.760 ****** \n\nTASK [common : Flush handlers] *************************************************\nTuesday 24 February 2026  00:48:14 +0000 (0:00:00.059)       0:01:00.819 ****** \n\nRUNNING HANDLER [common : Restart fluentd container] ***************************\nTuesday 24 February 2026  00:48:14 +0000 (0:00:00.081)       0:01:00.901 ****** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [common : Restart kolla-toolbox container] *********************\nTuesday 24 February 2026  00:48:46 +0000 (0:00:31.962)       0:01:32.863 ****** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\nRUNNING HANDLER [common : Initializing toolbox container using normal user] ****\nTuesday 24 February 2026  00:49:17 +0000 (0:00:30.979)       0:02:03.843 ****** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\nRUNNING HANDLER [common : Restart cron container] ******************************\nTuesday 24 February 2026  00:49:20 +0000 (0:00:02.533)       0:02:06.376 ****** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;33mtestbed-manager\u001b[0m            : \u001b[0;32mok=22  \u001b[0m \u001b[0;33mchanged=14  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=4   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=18  \u001b[0m \u001b[0;33mchanged=14  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=4   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=18  \u001b[0m \u001b[0;33mchanged=14  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=4   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=18  \u001b[0m \u001b[0;33mchanged=14  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=4   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-3\u001b[0m             : \u001b[0;32mok=18  \u001b[0m \u001b[0;33mchanged=14  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=4   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-4\u001b[0m             : \u001b[0;32mok=18  \u001b[0m \u001b[0;33mchanged=14  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=4   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-5\u001b[0m             : \u001b[0;32mok=18  \u001b[0m \u001b[0;33mchanged=14  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=4   \u001b[0m rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nTuesday 24 February 2026  00:49:29 +0000 (0:00:09.227)       0:02:15.603 ****** \n=============================================================================== \ncommon : Restart fluentd container ------------------------------------- 31.96s\ncommon : Restart kolla-toolbox container ------------------------------- 30.98s\ncommon : Restart cron container ----------------------------------------- 9.23s\ncommon : Copying over config.json files for services -------------------- 8.71s\nservice-cert-copy : common | Copying over extra CA certificates --------- 5.07s\ncommon : Check common containers ---------------------------------------- 4.62s\ncommon : Copying over cron logrotate config file ------------------------ 4.60s\ncommon : Copying over fluentd.conf -------------------------------------- 4.27s\ncommon : Ensuring config directories exist ------------------------------ 4.14s\ncommon : Ensuring config directories have correct owner and permission --- 3.34s\ncommon : Ensure RabbitMQ Erlang cookie exists --------------------------- 2.93s\ncommon : Copy rabbitmq-env.conf to kolla toolbox ------------------------ 2.67s\ncommon : Copy rabbitmq erl_inetrc to kolla toolbox ---------------------- 2.58s\ncommon : Initializing toolbox container using normal user --------------- 2.53s\nservice-cert-copy : common | Copying over backend internal TLS key ------ 2.24s\nservice-cert-copy : common | Copying over backend internal TLS certificate --- 2.14s\ncommon : Creating log volume -------------------------------------------- 2.09s\ncommon : Find custom fluentd input config files ------------------------- 1.75s\ncommon : include_tasks -------------------------------------------------- 1.56s\ncommon : include_tasks -------------------------------------------------- 1.44s\n\u001b[32m2026-02-24 00:49:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:49:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 404270fa-50db-4a78-8c44-21df48acba39 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:49:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 343b9f6a-0e5e-4af4-8877-c8c90685dcbb is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:49:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 0f568c96-7c6a-46f2-9e12-c74cfe06cd40 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:49:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 00:49:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ff4e45b7-3128-4ce1-95d9-909f33d92301 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:49:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f8694a2a-a9eb-4540-948b-ced7e58054f8 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:49:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:49:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 404270fa-50db-4a78-8c44-21df48acba39 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:49:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 343b9f6a-0e5e-4af4-8877-c8c90685dcbb is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:49:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 0f568c96-7c6a-46f2-9e12-c74cfe06cd40 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:49:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 00:49:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ff4e45b7-3128-4ce1-95d9-909f33d92301 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:49:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f8694a2a-a9eb-4540-948b-ced7e58054f8 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:49:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:49:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 404270fa-50db-4a78-8c44-21df48acba39 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:49:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 343b9f6a-0e5e-4af4-8877-c8c90685dcbb is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:49:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 0f568c96-7c6a-46f2-9e12-c74cfe06cd40 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:49:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 00:49:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ff4e45b7-3128-4ce1-95d9-909f33d92301 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:49:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f8694a2a-a9eb-4540-948b-ced7e58054f8 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:49:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:49:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 404270fa-50db-4a78-8c44-21df48acba39 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:49:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 343b9f6a-0e5e-4af4-8877-c8c90685dcbb is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:49:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 0f568c96-7c6a-46f2-9e12-c74cfe06cd40 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:49:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 00:49:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ff4e45b7-3128-4ce1-95d9-909f33d92301 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:49:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f8694a2a-a9eb-4540-948b-ced7e58054f8 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:49:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:49:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 404270fa-50db-4a78-8c44-21df48acba39 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:49:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 343b9f6a-0e5e-4af4-8877-c8c90685dcbb is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:49:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 0f568c96-7c6a-46f2-9e12-c74cfe06cd40 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:49:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 00:49:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ff4e45b7-3128-4ce1-95d9-909f33d92301 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:49:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f8694a2a-a9eb-4540-948b-ced7e58054f8 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:49:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:49:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 404270fa-50db-4a78-8c44-21df48acba39 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:49:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 343b9f6a-0e5e-4af4-8877-c8c90685dcbb is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:49:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 0f568c96-7c6a-46f2-9e12-c74cfe06cd40 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:49:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 00:49:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ff4e45b7-3128-4ce1-95d9-909f33d92301 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:49:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f8694a2a-a9eb-4540-948b-ced7e58054f8 is in state SUCCESS\u001b[0m\n\u001b[32m2026-02-24 00:49:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:49:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 404270fa-50db-4a78-8c44-21df48acba39 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:49:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 343b9f6a-0e5e-4af4-8877-c8c90685dcbb is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:49:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 0f568c96-7c6a-46f2-9e12-c74cfe06cd40 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:49:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 00:49:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ff4e45b7-3128-4ce1-95d9-909f33d92301 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:49:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e7b4207c-f95e-49ed-815c-ab81afbfeb3f is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:49:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:49:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 404270fa-50db-4a78-8c44-21df48acba39 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:49:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 343b9f6a-0e5e-4af4-8877-c8c90685dcbb is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:49:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 0f568c96-7c6a-46f2-9e12-c74cfe06cd40 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:49:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 00:49:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ff4e45b7-3128-4ce1-95d9-909f33d92301 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:49:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e7b4207c-f95e-49ed-815c-ab81afbfeb3f is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:49:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:49:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 404270fa-50db-4a78-8c44-21df48acba39 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:49:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 343b9f6a-0e5e-4af4-8877-c8c90685dcbb is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:49:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 0f568c96-7c6a-46f2-9e12-c74cfe06cd40 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:49:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 00:50:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ff4e45b7-3128-4ce1-95d9-909f33d92301 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:50:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e7b4207c-f95e-49ed-815c-ab81afbfeb3f is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:50:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:50:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 404270fa-50db-4a78-8c44-21df48acba39 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:50:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 343b9f6a-0e5e-4af4-8877-c8c90685dcbb is in state SUCCESS\u001b[0m\n\n\nPLAY [Group hosts based on configuration] **************************************\n\nTASK [Group hosts based on Kolla action] ***************************************\nTuesday 24 February 2026  00:49:36 +0000 (0:00:00.473)       0:00:00.473 ****** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [Group hosts based on enabled services] ***********************************\nTuesday 24 February 2026  00:49:37 +0000 (0:00:00.621)       0:00:01.094 ****** \n\u001b[0;32mok: [testbed-node-0] => (item=enable_memcached_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item=enable_memcached_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item=enable_memcached_True)\u001b[0m\n\nPLAY [Apply role memcached] ****************************************************\n\nTASK [memcached : include_tasks] ***********************************************\nTuesday 24 February 2026  00:49:37 +0000 (0:00:00.796)       0:00:01.891 ****** \n\u001b[0;36mincluded: /ansible/roles/memcached/tasks/deploy.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [memcached : Ensuring config directories exist] ***************************\nTuesday 24 February 2026  00:49:38 +0000 (0:00:00.501)       0:00:02.392 ****** \n\u001b[0;33mchanged: [testbed-node-2] => (item=memcached)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=memcached)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=memcached)\u001b[0m\n\nTASK [memcached : Copying over config.json files for services] *****************\nTuesday 24 February 2026  00:49:39 +0000 (0:00:00.877)       0:00:03.270 ****** \n\u001b[0;33mchanged: [testbed-node-2] => (item=memcached)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=memcached)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=memcached)\u001b[0m\n\nTASK [memcached : Check memcached container] ***********************************\nTuesday 24 February 2026  00:49:41 +0000 (0:00:02.108)       0:00:05.378 ****** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\nRUNNING HANDLER [memcached : Restart memcached container] **********************\nTuesday 24 February 2026  00:49:43 +0000 (0:00:01.897)       0:00:07.276 ****** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=7   \u001b[0m \u001b[0;33mchanged=4   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=7   \u001b[0m \u001b[0;33mchanged=4   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=7   \u001b[0m \u001b[0;33mchanged=4   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nTuesday 24 February 2026  00:49:50 +0000 (0:00:07.173)       0:00:14.449 ****** \n=============================================================================== \nmemcached : Restart memcached container --------------------------------- 7.17s\nmemcached : Copying over config.json files for services ----------------- 2.11s\nmemcached : Check memcached container ----------------------------------- 1.90s\nmemcached : Ensuring config directories exist --------------------------- 0.88s\nGroup hosts based on enabled services ----------------------------------- 0.80s\nGroup hosts based on Kolla action --------------------------------------- 0.62s\nmemcached : include_tasks ----------------------------------------------- 0.50s\n\n\nPLAY [Group hosts based on configuration] **************************************\n\nTASK [Group hosts based on Kolla action] ***************************************\nTuesday 24 February 2026  00:49:35 +0000 (0:00:00.429)       0:00:00.429 ****** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [Group hosts based on enabled services] ***********************************\nTuesday 24 February 2026  00:49:35 +0000 (0:00:00.265)       0:00:00.695 ****** \n\u001b[0;32mok: [testbed-node-0] => (item=enable_redis_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item=enable_redis_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item=enable_redis_True)\u001b[0m\n\nPLAY [Apply role redis] ********************************************************\n\nTASK [redis : include_tasks] ***************************************************\nTuesday 24 February 2026  00:49:36 +0000 (0:00:00.799)       0:00:01.495 ****** \n\u001b[0;36mincluded: /ansible/roles/redis/tasks/deploy.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [redis : Ensuring config directories exist] *******************************\nTuesday 24 February 2026  00:49:37 +0000 (0:00:00.547)       0:00:02.042 ****** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'redis', 'value': {'container_name': 'redis', 'group': 'redis', 'enabled': True, 'image': 'registry.osism.tech/kolla/redis:2024.2', 'volumes': ['/etc/kolla/redis/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'redis:/var/lib/redis/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-server 6379'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'redis', 'value': {'container_name': 'redis', 'group': 'redis', 'enabled': True, 'image': 'registry.osism.tech/kolla/redis:2024.2', 'volumes': ['/etc/kolla/redis/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'redis:/var/lib/redis/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-server 6379'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'redis', 'value': {'container_name': 'redis', 'group': 'redis', 'enabled': True, 'image': 'registry.osism.tech/kolla/redis:2024.2', 'volumes': ['/etc/kolla/redis/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'redis:/var/lib/redis/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-server 6379'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'redis-sentinel', 'value': {'container_name': 'redis_sentinel', 'group': 'redis', 'environment': {'REDIS_CONF': '/etc/redis/redis.conf', 'REDIS_GEN_CONF': '/etc/redis/redis-regenerated-by-config-rewrite.conf'}, 'enabled': True, 'image': 'registry.osism.tech/kolla/redis-sentinel:2024.2', 'volumes': ['/etc/kolla/redis-sentinel/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-sentinel 26379'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'redis-sentinel', 'value': {'container_name': 'redis_sentinel', 'group': 'redis', 'environment': {'REDIS_CONF': '/etc/redis/redis.conf', 'REDIS_GEN_CONF': '/etc/redis/redis-regenerated-by-config-rewrite.conf'}, 'enabled': True, 'image': 'registry.osism.tech/kolla/redis-sentinel:2024.2', 'volumes': ['/etc/kolla/redis-sentinel/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-sentinel 26379'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'redis-sentinel', 'value': {'container_name': 'redis_sentinel', 'group': 'redis', 'environment': {'REDIS_CONF': '/etc/redis/redis.conf', 'REDIS_GEN_CONF': '/etc/redis/redis-regenerated-by-config-rewrite.conf'}, 'enabled': True, 'image': 'registry.osism.tech/kolla/redis-sentinel:2024.2', 'volumes': ['/etc/kolla/redis-sentinel/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-sentinel 26379'], 'timeout': '30'}}})\u001b[0m\n\nTASK [redis : Copying over default config.json files] **************************\nTuesday 24 February 2026  00:49:38 +0000 (0:00:01.585)       0:00:03.628 ****** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'redis', 'value': {'container_name': 'redis', 'group': 'redis', 'enabled': True, 'image': 'registry.osism.tech/kolla/redis:2024.2', 'volumes': ['/etc/kolla/redis/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'redis:/var/lib/redis/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-server 6379'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'redis', 'value': {'container_name': 'redis', 'group': 'redis', 'enabled': True, 'image': 'registry.osism.tech/kolla/redis:2024.2', 'volumes': ['/etc/kolla/redis/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'redis:/var/lib/redis/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-server 6379'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'redis', 'value': {'container_name': 'redis', 'group': 'redis', 'enabled': True, 'image': 'registry.osism.tech/kolla/redis:2024.2', 'volumes': ['/etc/kolla/redis/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'redis:/var/lib/redis/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-server 6379'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'redis-sentinel', 'value': {'container_name': 'redis_sentinel', 'group': 'redis', 'environment': {'REDIS_CONF': '/etc/redis/redis.conf', 'REDIS_GEN_CONF': '/etc/redis/redis-regenerated-by-config-rewrite.conf'}, 'enabled': True, 'image': 'registry.osism.tech/kolla/redis-sentinel:2024.2', 'volumes': ['/etc/kolla/redis-sentinel/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-sentinel 26379'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'redis-sentinel', 'value': {'container_name': 'redis_sentinel', 'group': 'redis', 'environment': {'REDIS_CONF': '/etc/redis/redis.conf', 'REDIS_GEN_CONF': '/etc/redis/redis-regenerated-by-config-rewrite.conf'}, 'enabled': True, 'image': 'registry.osism.tech/kolla/redis-sentinel:2024.2', 'volumes': ['/etc/kolla/redis-sentinel/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-sentinel 26379'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'redis-sentinel', 'value': {'container_name': 'redis_sentinel', 'group': 'redis', 'environment': {'REDIS_CONF': '/etc/redis/redis.conf', 'REDIS_GEN_CONF': '/etc/redis/redis-regenerated-by-config-rewrite.conf'}, 'enabled': True, 'image': 'registry.osism.tech/kolla/redis-sentinel:2024.2', 'volumes': ['/etc/kolla/redis-sentinel/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-sentinel 26379'], 'timeout': '30'}}})\u001b[0m\n\nTASK [redis : Copying over redis config files] *********************************\nTuesday 24 February 2026  00:49:41 +0000 (0:00:03.088)       0:00:06.717 ****** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'redis', 'value': {'container_name': 'redis', 'group': 'redis', 'enabled': True, 'image': 'registry.osism.tech/kolla/redis:2024.2', 'volumes': ['/etc/kolla/redis/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'redis:/var/lib/redis/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-server 6379'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'redis', 'value': {'container_name': 'redis', 'group': 'redis', 'enabled': True, 'image': 'registry.osism.tech/kolla/redis:2024.2', 'volumes': ['/etc/kolla/redis/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'redis:/var/lib/redis/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-server 6379'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'redis', 'value': {'container_name': 'redis', 'group': 'redis', 'enabled': True, 'image': 'registry.osism.tech/kolla/redis:2024.2', 'volumes': ['/etc/kolla/redis/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'redis:/var/lib/redis/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-server 6379'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'redis-sentinel', 'value': {'container_name': 'redis_sentinel', 'group': 'redis', 'environment': {'REDIS_CONF': '/etc/redis/redis.conf', 'REDIS_GEN_CONF': '/etc/redis/redis-regenerated-by-config-rewrite.conf'}, 'enabled': True, 'image': 'registry.osism.tech/kolla/redis-sentinel:2024.2', 'volumes': ['/etc/kolla/redis-sentinel/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-sentinel 26379'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'redis-sentinel', 'value': {'container_name': 'redis_sentinel', 'group': 'redis', 'environment': {'REDIS_CONF': '/etc/redis/redis.conf', 'REDIS_GEN_CONF': '/etc/redis/redis-regenerated-by-config-rewrite.conf'}, 'enabled': True, 'image': 'registry.osism.tech/kolla/redis-sentinel:2024.2', 'volumes': ['/etc/kolla/redis-sentinel/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-sentinel 26379'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'redis-sentinel', 'value': {'container_name': 'redis_sentinel', 'group': 'redis', 'environment': {'REDIS_CONF': '/etc/redis/redis.conf', 'REDIS_GEN_CONF': '/etc/redis/redis-regenerated-by-config-rewrite.conf'}, 'enabled': True, 'image': 'registry.osism.tech/kolla/redis-sentinel:2024.2', 'volumes': ['/etc/kolla/redis-sentinel/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-sentinel 26379'], 'timeout': '30'}}})\u001b[0m\n\nTASK [redis : Check redis containers] ******************************************\nTuesday 24 February 2026  00:49:44 +0000 (0:00:02.893)       0:00:09.610 ****** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'redis', 'value': {'container_name': 'redis', 'group': 'redis', 'enabled': True, 'image': 'registry.osism.tech/kolla/redis:2024.2', 'volumes': ['/etc/kolla/redis/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'redis:/var/lib/redis/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-server 6379'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'redis', 'value': {'container_name': 'redis', 'group': 'redis', 'enabled': True, 'image': 'registry.osism.tech/kolla/redis:2024.2', 'volumes': ['/etc/kolla/redis/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'redis:/var/lib/redis/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-server 6379'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'redis', 'value': {'container_name': 'redis', 'group': 'redis', 'enabled': True, 'image': 'registry.osism.tech/kolla/redis:2024.2', 'volumes': ['/etc/kolla/redis/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'redis:/var/lib/redis/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-server 6379'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'redis-sentinel', 'value': {'container_name': 'redis_sentinel', 'group': 'redis', 'environment': {'REDIS_CONF': '/etc/redis/redis.conf', 'REDIS_GEN_CONF': '/etc/redis/redis-regenerated-by-config-rewrite.conf'}, 'enabled': True, 'image': 'registry.osism.tech/kolla/redis-sentinel:2024.2', 'volumes': ['/etc/kolla/redis-sentinel/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-sentinel 26379'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'redis-sentinel', 'value': {'container_name': 'redis_sentinel', 'group': 'redis', 'environment': {'REDIS_CONF': '/etc/redis/redis.conf', 'REDIS_GEN_CONF': '/etc/redis/redis-regenerated-by-config-rewrite.conf'}, 'enabled': True, 'image': 'registry.osism.tech/kolla/redis-sentinel:2024.2', 'volumes': ['/etc/kolla/redis-sentinel/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-sentinel 26379'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'redis-sentinel', 'value': {'container_name': 'redis_sentinel', 'group': 'redis', 'environment': {'REDIS_CONF': '/etc/redis/redis.conf', 'REDIS_GEN_CONF': '/etc/redis/redis-regenerated-by-config-rewrite.conf'}, 'enabled': True, 'image': 'registry.osism.tech/kolla/redis-sentinel:2024.2', 'volumes': ['/etc/kolla/redis-sentinel/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-sentinel 26379'], 'timeout': '30'}}})\u001b[0m\n\nTASK [redis : Flush handlers] **************************************************\nTuesday 24 February 2026  00:49:46 +0000 (0:00:01.754)       0:00:11.365 ****** \n\nTASK [redis : Flush handlers] **************************************************\nTuesday 24 February 2026  00:49:46 +0000 (0:00:00.143)       0:00:11.508 ****** \n\nTASK [redis : Flush handlers] **************************************************\nTuesday 24 February 2026  00:49:46 +0000 (0:00:00.113)       0:00:11.622 ****** \n\nRUNNING HANDLER [redis : Restart redis container] ******************************\nTuesday 24 February 2026  00:49:46 +0000 (0:00:00.077)       0:00:11.699 ****** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\nRUNNING HANDLER [redis : Restart redis-sentinel container] *********************\nTuesday 24 February 2026  00:49:55 +0000 (0:00:09.001)       0:00:20.701 ****** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=9   \u001b[0m \u001b[0;33mchanged=6   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=9   \u001b[0m \u001b[0;33mchanged=6   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=9   \u001b[0m \u001b[0;33mchanged=6   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nTuesday 24 February 2026  00:49:58 +0000 (0:00:03.207)       0:00:23.908 ****** \n=============================================================================== \nredis : Restart redis container ----------------------------------------- 9.00s\nredis : Restart redis-sentinel container -------------------------------- 3.21s\nredis : Copying over default config.json files -------------------------- 3.09s\nredis : Copying over redis config files --------------------------------- 2.89s\nredis : Check redis containers ------------------------------------------ 1.75s\nredis : Ensuring config directories exist ------------------------------- 1.59s\nGroup hosts based on enabled services ----------------------------------- 0.80s\nredis : include_tasks --------------------------------------------------- 0.55s\nredis : Flush handlers -------------------------------------------------- 0.33s\nGroup hosts based on Kolla action --------------------------------------- 0.27s\n\u001b[32m2026-02-24 00:50:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 0f568c96-7c6a-46f2-9e12-c74cfe06cd40 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:50:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 00:50:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ff4e45b7-3128-4ce1-95d9-909f33d92301 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:50:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e7b4207c-f95e-49ed-815c-ab81afbfeb3f is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:50:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:50:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 404270fa-50db-4a78-8c44-21df48acba39 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:50:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 0f568c96-7c6a-46f2-9e12-c74cfe06cd40 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:50:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 00:50:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ff4e45b7-3128-4ce1-95d9-909f33d92301 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:50:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e7b4207c-f95e-49ed-815c-ab81afbfeb3f is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:50:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:50:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 404270fa-50db-4a78-8c44-21df48acba39 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:50:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 0f568c96-7c6a-46f2-9e12-c74cfe06cd40 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:50:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 00:50:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ff4e45b7-3128-4ce1-95d9-909f33d92301 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:50:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e7b4207c-f95e-49ed-815c-ab81afbfeb3f is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:50:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:50:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 404270fa-50db-4a78-8c44-21df48acba39 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:50:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 0f568c96-7c6a-46f2-9e12-c74cfe06cd40 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:50:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 00:50:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ff4e45b7-3128-4ce1-95d9-909f33d92301 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:50:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e7b4207c-f95e-49ed-815c-ab81afbfeb3f is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:50:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:50:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 404270fa-50db-4a78-8c44-21df48acba39 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:50:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 0f568c96-7c6a-46f2-9e12-c74cfe06cd40 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:50:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 00:50:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ff4e45b7-3128-4ce1-95d9-909f33d92301 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:50:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e7b4207c-f95e-49ed-815c-ab81afbfeb3f is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:50:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:50:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 404270fa-50db-4a78-8c44-21df48acba39 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:50:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 0f568c96-7c6a-46f2-9e12-c74cfe06cd40 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:50:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 00:50:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ff4e45b7-3128-4ce1-95d9-909f33d92301 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:50:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e7b4207c-f95e-49ed-815c-ab81afbfeb3f is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:50:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:50:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 404270fa-50db-4a78-8c44-21df48acba39 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:50:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 0f568c96-7c6a-46f2-9e12-c74cfe06cd40 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:50:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 00:50:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ff4e45b7-3128-4ce1-95d9-909f33d92301 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:50:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e7b4207c-f95e-49ed-815c-ab81afbfeb3f is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:50:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:50:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 404270fa-50db-4a78-8c44-21df48acba39 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:50:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 0f568c96-7c6a-46f2-9e12-c74cfe06cd40 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:50:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 00:50:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ff4e45b7-3128-4ce1-95d9-909f33d92301 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:50:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e7b4207c-f95e-49ed-815c-ab81afbfeb3f is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:50:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:50:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 404270fa-50db-4a78-8c44-21df48acba39 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:50:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 0f568c96-7c6a-46f2-9e12-c74cfe06cd40 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:50:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 00:50:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ff4e45b7-3128-4ce1-95d9-909f33d92301 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:50:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e7b4207c-f95e-49ed-815c-ab81afbfeb3f is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:50:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:50:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 404270fa-50db-4a78-8c44-21df48acba39 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:50:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 0f568c96-7c6a-46f2-9e12-c74cfe06cd40 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:50:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 00:50:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ff4e45b7-3128-4ce1-95d9-909f33d92301 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:50:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e7b4207c-f95e-49ed-815c-ab81afbfeb3f is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:50:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:50:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 404270fa-50db-4a78-8c44-21df48acba39 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:50:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 0f568c96-7c6a-46f2-9e12-c74cfe06cd40 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:50:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 00:50:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ff4e45b7-3128-4ce1-95d9-909f33d92301 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:50:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e7b4207c-f95e-49ed-815c-ab81afbfeb3f is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:50:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:50:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 404270fa-50db-4a78-8c44-21df48acba39 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:50:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 0f568c96-7c6a-46f2-9e12-c74cfe06cd40 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:50:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 00:50:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ff4e45b7-3128-4ce1-95d9-909f33d92301 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:50:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e7b4207c-f95e-49ed-815c-ab81afbfeb3f is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:50:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:50:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 404270fa-50db-4a78-8c44-21df48acba39 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:50:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 0f568c96-7c6a-46f2-9e12-c74cfe06cd40 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:50:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 00:50:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ff4e45b7-3128-4ce1-95d9-909f33d92301 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:50:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e7b4207c-f95e-49ed-815c-ab81afbfeb3f is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:50:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ba7d4726-e54a-4b5b-b060-92175ab177a2 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:50:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:50:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 404270fa-50db-4a78-8c44-21df48acba39 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:50:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 0f568c96-7c6a-46f2-9e12-c74cfe06cd40 is in state SUCCESS\u001b[0m\n\n\nPLAY [Group hosts based on configuration] **************************************\n\nTASK [Group hosts based on Kolla action] ***************************************\nTuesday 24 February 2026  00:49:35 +0000 (0:00:00.236)       0:00:00.236 ****** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [Group hosts based on enabled services] ***********************************\nTuesday 24 February 2026  00:49:36 +0000 (0:00:00.798)       0:00:01.034 ****** \n\u001b[0;32mok: [testbed-node-0] => (item=enable_openvswitch_True_enable_ovs_dpdk_False)\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item=enable_openvswitch_True_enable_ovs_dpdk_False)\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item=enable_openvswitch_True_enable_ovs_dpdk_False)\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item=enable_openvswitch_True_enable_ovs_dpdk_False)\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item=enable_openvswitch_True_enable_ovs_dpdk_False)\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item=enable_openvswitch_True_enable_ovs_dpdk_False)\u001b[0m\n\nPLAY [Apply role openvswitch] **************************************************\n\nTASK [openvswitch : include_tasks] *********************************************\nTuesday 24 February 2026  00:49:37 +0000 (0:00:01.031)       0:00:02.066 ****** \n\u001b[0;36mincluded: /ansible/roles/openvswitch/tasks/deploy.yml for testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m\n\nTASK [module-load : Load modules] **********************************************\nTuesday 24 February 2026  00:49:38 +0000 (0:00:01.459)       0:00:03.526 ****** \n\u001b[0;33mchanged: [testbed-node-0] => (item=openvswitch)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=openvswitch)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=openvswitch)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item=openvswitch)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=openvswitch)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=openvswitch)\u001b[0m\n\nTASK [module-load : Persist modules via modules-load.d] ************************\nTuesday 24 February 2026  00:49:40 +0000 (0:00:01.801)       0:00:05.327 ****** \n\u001b[0;33mchanged: [testbed-node-0] => (item=openvswitch)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=openvswitch)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=openvswitch)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item=openvswitch)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=openvswitch)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=openvswitch)\u001b[0m\n\nTASK [module-load : Drop module persistence] ***********************************\nTuesday 24 February 2026  00:49:42 +0000 (0:00:01.659)       0:00:06.987 ****** \n\u001b[0;36mskipping: [testbed-node-0] => (item=openvswitch) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=openvswitch) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=openvswitch) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=openvswitch) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=openvswitch) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=openvswitch) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [openvswitch : Create /run/openvswitch directory on host] *****************\nTuesday 24 February 2026  00:49:43 +0000 (0:00:01.095)       0:00:08.082 ****** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [openvswitch : Ensuring config directories exist] *************************\nTuesday 24 February 2026  00:49:44 +0000 (0:00:00.912)       0:00:08.994 ****** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'openvswitch-db-server', 'value': {'container_name': 'openvswitch_db', 'image': 'registry.osism.tech/kolla/openvswitch-db-server:2024.2', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'volumes': ['/etc/kolla/openvswitch-db-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', 'openvswitch_db:/var/lib/openvswitch/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovsdb-client list-dbs'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'openvswitch-db-server', 'value': {'container_name': 'openvswitch_db', 'image': 'registry.osism.tech/kolla/openvswitch-db-server:2024.2', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'volumes': ['/etc/kolla/openvswitch-db-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', 'openvswitch_db:/var/lib/openvswitch/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovsdb-client list-dbs'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'openvswitch-db-server', 'value': {'container_name': 'openvswitch_db', 'image': 'registry.osism.tech/kolla/openvswitch-db-server:2024.2', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'volumes': ['/etc/kolla/openvswitch-db-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', 'openvswitch_db:/var/lib/openvswitch/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovsdb-client list-dbs'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'openvswitch-db-server', 'value': {'container_name': 'openvswitch_db', 'image': 'registry.osism.tech/kolla/openvswitch-db-server:2024.2', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'volumes': ['/etc/kolla/openvswitch-db-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', 'openvswitch_db:/var/lib/openvswitch/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovsdb-client list-dbs'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'openvswitch-vswitchd', 'value': {'container_name': 'openvswitch_vswitchd', 'image': 'registry.osism.tech/kolla/openvswitch-vswitchd:2024.2', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'privileged': True, 'volumes': ['/etc/kolla/openvswitch-vswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovs-appctl version'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'openvswitch-db-server', 'value': {'container_name': 'openvswitch_db', 'image': 'registry.osism.tech/kolla/openvswitch-db-server:2024.2', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'volumes': ['/etc/kolla/openvswitch-db-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', 'openvswitch_db:/var/lib/openvswitch/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovsdb-client list-dbs'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'openvswitch-vswitchd', 'value': {'container_name': 'openvswitch_vswitchd', 'image': 'registry.osism.tech/kolla/openvswitch-vswitchd:2024.2', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'privileged': True, 'volumes': ['/etc/kolla/openvswitch-vswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovs-appctl version'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'openvswitch-vswitchd', 'value': {'container_name': 'openvswitch_vswitchd', 'image': 'registry.osism.tech/kolla/openvswitch-vswitchd:2024.2', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'privileged': True, 'volumes': ['/etc/kolla/openvswitch-vswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovs-appctl version'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'openvswitch-db-server', 'value': {'container_name': 'openvswitch_db', 'image': 'registry.osism.tech/kolla/openvswitch-db-server:2024.2', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'volumes': ['/etc/kolla/openvswitch-db-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', 'openvswitch_db:/var/lib/openvswitch/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovsdb-client list-dbs'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'openvswitch-vswitchd', 'value': {'container_name': 'openvswitch_vswitchd', 'image': 'registry.osism.tech/kolla/openvswitch-vswitchd:2024.2', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'privileged': True, 'volumes': ['/etc/kolla/openvswitch-vswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovs-appctl version'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'openvswitch-vswitchd', 'value': {'container_name': 'openvswitch_vswitchd', 'image': 'registry.osism.tech/kolla/openvswitch-vswitchd:2024.2', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'privileged': True, 'volumes': ['/etc/kolla/openvswitch-vswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovs-appctl version'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'openvswitch-vswitchd', 'value': {'container_name': 'openvswitch_vswitchd', 'image': 'registry.osism.tech/kolla/openvswitch-vswitchd:2024.2', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'privileged': True, 'volumes': ['/etc/kolla/openvswitch-vswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovs-appctl version'], 'timeout': '30'}}})\u001b[0m\n\nTASK [openvswitch : Copying over config.json files for services] ***************\nTuesday 24 February 2026  00:49:45 +0000 (0:00:01.723)       0:00:10.717 ****** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'openvswitch-db-server', 'value': {'container_name': 'openvswitch_db', 'image': 'registry.osism.tech/kolla/openvswitch-db-server:2024.2', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'volumes': ['/etc/kolla/openvswitch-db-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', 'openvswitch_db:/var/lib/openvswitch/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovsdb-client list-dbs'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'openvswitch-db-server', 'value': {'container_name': 'openvswitch_db', 'image': 'registry.osism.tech/kolla/openvswitch-db-server:2024.2', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'volumes': ['/etc/kolla/openvswitch-db-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', 'openvswitch_db:/var/lib/openvswitch/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovsdb-client list-dbs'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'openvswitch-db-server', 'value': {'container_name': 'openvswitch_db', 'image': 'registry.osism.tech/kolla/openvswitch-db-server:2024.2', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'volumes': ['/etc/kolla/openvswitch-db-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', 'openvswitch_db:/var/lib/openvswitch/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovsdb-client list-dbs'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'openvswitch-db-server', 'value': {'container_name': 'openvswitch_db', 'image': 'registry.osism.tech/kolla/openvswitch-db-server:2024.2', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'volumes': ['/etc/kolla/openvswitch-db-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', 'openvswitch_db:/var/lib/openvswitch/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovsdb-client list-dbs'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'openvswitch-db-server', 'value': {'container_name': 'openvswitch_db', 'image': 'registry.osism.tech/kolla/openvswitch-db-server:2024.2', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'volumes': ['/etc/kolla/openvswitch-db-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', 'openvswitch_db:/var/lib/openvswitch/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovsdb-client list-dbs'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'openvswitch-db-server', 'value': {'container_name': 'openvswitch_db', 'image': 'registry.osism.tech/kolla/openvswitch-db-server:2024.2', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'volumes': ['/etc/kolla/openvswitch-db-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', 'openvswitch_db:/var/lib/openvswitch/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovsdb-client list-dbs'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'openvswitch-vswitchd', 'value': {'container_name': 'openvswitch_vswitchd', 'image': 'registry.osism.tech/kolla/openvswitch-vswitchd:2024.2', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'privileged': True, 'volumes': ['/etc/kolla/openvswitch-vswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovs-appctl version'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'openvswitch-vswitchd', 'value': {'container_name': 'openvswitch_vswitchd', 'image': 'registry.osism.tech/kolla/openvswitch-vswitchd:2024.2', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'privileged': True, 'volumes': ['/etc/kolla/openvswitch-vswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovs-appctl version'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'openvswitch-vswitchd', 'value': {'container_name': 'openvswitch_vswitchd', 'image': 'registry.osism.tech/kolla/openvswitch-vswitchd:2024.2', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'privileged': True, 'volumes': ['/etc/kolla/openvswitch-vswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovs-appctl version'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'openvswitch-vswitchd', 'value': {'container_name': 'openvswitch_vswitchd', 'image': 'registry.osism.tech/kolla/openvswitch-vswitchd:2024.2', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'privileged': True, 'volumes': ['/etc/kolla/openvswitch-vswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovs-appctl version'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'openvswitch-vswitchd', 'value': {'container_name': 'openvswitch_vswitchd', 'image': 'registry.osism.tech/kolla/openvswitch-vswitchd:2024.2', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'privileged': True, 'volumes': ['/etc/kolla/openvswitch-vswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovs-appctl version'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'openvswitch-vswitchd', 'value': {'container_name': 'openvswitch_vswitchd', 'image': 'registry.osism.tech/kolla/openvswitch-vswitchd:2024.2', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'privileged': True, 'volumes': ['/etc/kolla/openvswitch-vswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovs-appctl version'], 'timeout': '30'}}})\u001b[0m\n\nTASK [openvswitch : Copying over ovs-vsctl wrapper] ****************************\nTuesday 24 February 2026  00:49:48 +0000 (0:00:02.791)       0:00:13.509 ****** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [openvswitch : Check openvswitch containers] ******************************\nTuesday 24 February 2026  00:49:49 +0000 (0:00:00.825)       0:00:14.335 ****** \n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'openvswitch-db-server', 'value': {'container_name': 'openvswitch_db', 'image': 'registry.osism.tech/kolla/openvswitch-db-server:2024.2', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'volumes': ['/etc/kolla/openvswitch-db-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', 'openvswitch_db:/var/lib/openvswitch/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovsdb-client list-dbs'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'openvswitch-db-server', 'value': {'container_name': 'openvswitch_db', 'image': 'registry.osism.tech/kolla/openvswitch-db-server:2024.2', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'volumes': ['/etc/kolla/openvswitch-db-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', 'openvswitch_db:/var/lib/openvswitch/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovsdb-client list-dbs'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'openvswitch-db-server', 'value': {'container_name': 'openvswitch_db', 'image': 'registry.osism.tech/kolla/openvswitch-db-server:2024.2', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'volumes': ['/etc/kolla/openvswitch-db-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', 'openvswitch_db:/var/lib/openvswitch/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovsdb-client list-dbs'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'openvswitch-db-server', 'value': {'container_name': 'openvswitch_db', 'image': 'registry.osism.tech/kolla/openvswitch-db-server:2024.2', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'volumes': ['/etc/kolla/openvswitch-db-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', 'openvswitch_db:/var/lib/openvswitch/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovsdb-client list-dbs'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'openvswitch-db-server', 'value': {'container_name': 'openvswitch_db', 'image': 'registry.osism.tech/kolla/openvswitch-db-server:2024.2', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'volumes': ['/etc/kolla/openvswitch-db-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', 'openvswitch_db:/var/lib/openvswitch/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovsdb-client list-dbs'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'openvswitch-db-server', 'value': {'container_name': 'openvswitch_db', 'image': 'registry.osism.tech/kolla/openvswitch-db-server:2024.2', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'volumes': ['/etc/kolla/openvswitch-db-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', 'openvswitch_db:/var/lib/openvswitch/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovsdb-client list-dbs'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'openvswitch-vswitchd', 'value': {'container_name': 'openvswitch_vswitchd', 'image': 'registry.osism.tech/kolla/openvswitch-vswitchd:2024.2', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'privileged': True, 'volumes': ['/etc/kolla/openvswitch-vswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovs-appctl version'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'openvswitch-vswitchd', 'value': {'container_name': 'openvswitch_vswitchd', 'image': 'registry.osism.tech/kolla/openvswitch-vswitchd:2024.2', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'privileged': True, 'volumes': ['/etc/kolla/openvswitch-vswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovs-appctl version'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'openvswitch-vswitchd', 'value': {'container_name': 'openvswitch_vswitchd', 'image': 'registry.osism.tech/kolla/openvswitch-vswitchd:2024.2', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'privileged': True, 'volumes': ['/etc/kolla/openvswitch-vswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovs-appctl version'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'openvswitch-vswitchd', 'value': {'container_name': 'openvswitch_vswitchd', 'image': 'registry.osism.tech/kolla/openvswitch-vswitchd:2024.2', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'privileged': True, 'volumes': ['/etc/kolla/openvswitch-vswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovs-appctl version'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'openvswitch-vswitchd', 'value': {'container_name': 'openvswitch_vswitchd', 'image': 'registry.osism.tech/kolla/openvswitch-vswitchd:2024.2', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'privileged': True, 'volumes': ['/etc/kolla/openvswitch-vswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovs-appctl version'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'openvswitch-vswitchd', 'value': {'container_name': 'openvswitch_vswitchd', 'image': 'registry.osism.tech/kolla/openvswitch-vswitchd:2024.2', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'privileged': True, 'volumes': ['/etc/kolla/openvswitch-vswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovs-appctl version'], 'timeout': '30'}}})\u001b[0m\n\nTASK [openvswitch : Flush Handlers] ********************************************\nTuesday 24 February 2026  00:49:51 +0000 (0:00:02.158)       0:00:16.494 ****** \n\nTASK [openvswitch : Flush Handlers] ********************************************\nTuesday 24 February 2026  00:49:52 +0000 (0:00:00.345)       0:00:16.840 ****** \n\nTASK [openvswitch : Flush Handlers] ********************************************\nTuesday 24 February 2026  00:49:52 +0000 (0:00:00.147)       0:00:16.987 ****** \n\nTASK [openvswitch : Flush Handlers] ********************************************\nTuesday 24 February 2026  00:49:52 +0000 (0:00:00.153)       0:00:17.140 ****** \n\nTASK [openvswitch : Flush Handlers] ********************************************\nTuesday 24 February 2026  00:49:52 +0000 (0:00:00.138)       0:00:17.279 ****** \n\nTASK [openvswitch : Flush Handlers] ********************************************\nTuesday 24 February 2026  00:49:52 +0000 (0:00:00.229)       0:00:17.509 ****** \n\nRUNNING HANDLER [openvswitch : Restart openvswitch-db-server container] ********\nTuesday 24 February 2026  00:49:52 +0000 (0:00:00.162)       0:00:17.671 ****** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\nRUNNING HANDLER [openvswitch : Waiting for openvswitch_db service to be ready] ***\nTuesday 24 February 2026  00:50:02 +0000 (0:00:10.090)       0:00:27.762 ****** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nRUNNING HANDLER [openvswitch : Restart openvswitch-vswitchd container] *********\nTuesday 24 February 2026  00:50:05 +0000 (0:00:02.400)       0:00:30.162 ****** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nTASK [openvswitch : Set system-id, hostname and hw-offload] ********************\nTuesday 24 February 2026  00:50:15 +0000 (0:00:10.030)       0:00:40.192 ****** \n\u001b[0;33mchanged: [testbed-node-2] => (item={'col': 'external_ids', 'name': 'system-id', 'value': 'testbed-node-2'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'col': 'external_ids', 'name': 'system-id', 'value': 'testbed-node-0'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'col': 'external_ids', 'name': 'system-id', 'value': 'testbed-node-4'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'col': 'external_ids', 'name': 'system-id', 'value': 'testbed-node-5'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'col': 'external_ids', 'name': 'system-id', 'value': 'testbed-node-3'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'col': 'external_ids', 'name': 'system-id', 'value': 'testbed-node-1'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'col': 'external_ids', 'name': 'hostname', 'value': 'testbed-node-4'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'col': 'external_ids', 'name': 'hostname', 'value': 'testbed-node-2'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'col': 'external_ids', 'name': 'hostname', 'value': 'testbed-node-0'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'col': 'external_ids', 'name': 'hostname', 'value': 'testbed-node-1'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'col': 'external_ids', 'name': 'hostname', 'value': 'testbed-node-3'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'col': 'external_ids', 'name': 'hostname', 'value': 'testbed-node-5'})\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item={'col': 'other_config', 'name': 'hw-offload', 'value': True, 'state': 'absent'})\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item={'col': 'other_config', 'name': 'hw-offload', 'value': True, 'state': 'absent'})\u001b[0m\n\u001b[0;32mok: [testbed-node-0] => (item={'col': 'other_config', 'name': 'hw-offload', 'value': True, 'state': 'absent'})\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item={'col': 'other_config', 'name': 'hw-offload', 'value': True, 'state': 'absent'})\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item={'col': 'other_config', 'name': 'hw-offload', 'value': True, 'state': 'absent'})\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item={'col': 'other_config', 'name': 'hw-offload', 'value': True, 'state': 'absent'})\u001b[0m\n\nTASK [openvswitch : Ensuring OVS bridge is properly setup] *********************\nTuesday 24 February 2026  00:50:22 +0000 (0:00:06.907)       0:00:47.100 ****** \n\u001b[0;36mskipping: [testbed-node-3] => (item=br-ex) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=br-ex) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=br-ex) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=br-ex)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=br-ex)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=br-ex)\u001b[0m\n\nTASK [openvswitch : Ensuring OVS ports are properly setup] *********************\nTuesday 24 February 2026  00:50:25 +0000 (0:00:03.009)       0:00:50.109 ****** \n\u001b[0;36mskipping: [testbed-node-3] => (item=['br-ex', 'vxlan0']) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=['br-ex', 'vxlan0']) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=['br-ex', 'vxlan0']) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=['br-ex', 'vxlan0'])\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=['br-ex', 'vxlan0'])\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=['br-ex', 'vxlan0'])\u001b[0m\n\nRUNNING HANDLER [openvswitch : Restart openvswitch-vswitchd container] *********\nTuesday 24 February 2026  00:50:29 +0000 (0:00:03.997)       0:00:54.107 ****** \n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=15  \u001b[0m \u001b[0;33mchanged=11  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=3   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=15  \u001b[0m \u001b[0;33mchanged=11  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=3   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=15  \u001b[0m \u001b[0;33mchanged=11  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=3   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-3\u001b[0m             : \u001b[0;32mok=13  \u001b[0m \u001b[0;33mchanged=9   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=5   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-4\u001b[0m             : \u001b[0;32mok=13  \u001b[0m \u001b[0;33mchanged=9   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=5   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-5\u001b[0m             : \u001b[0;32mok=13  \u001b[0m \u001b[0;33mchanged=9   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=5   \u001b[0m rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nTuesday 24 February 2026  00:50:38 +0000 (0:00:08.774)       0:01:02.881 ****** \n=============================================================================== \nopenvswitch : Restart openvswitch-vswitchd container ------------------- 18.81s\nopenvswitch : Restart openvswitch-db-server container ------------------ 10.09s\nopenvswitch : Set system-id, hostname and hw-offload -------------------- 6.91s\nopenvswitch : Ensuring OVS ports are properly setup --------------------- 4.00s\nopenvswitch : Ensuring OVS bridge is properly setup --------------------- 3.01s\nopenvswitch : Copying over config.json files for services --------------- 2.79s\nopenvswitch : Waiting for openvswitch_db service to be ready ------------ 2.40s\nopenvswitch : Check openvswitch containers ------------------------------ 2.16s\nmodule-load : Load modules ---------------------------------------------- 1.80s\nopenvswitch : Ensuring config directories exist ------------------------- 1.72s\nmodule-load : Persist modules via modules-load.d ------------------------ 1.66s\nopenvswitch : include_tasks --------------------------------------------- 1.46s\nopenvswitch : Flush Handlers -------------------------------------------- 1.18s\nmodule-load : Drop module persistence ----------------------------------- 1.10s\nGroup hosts based on enabled services ----------------------------------- 1.03s\nopenvswitch : Create /run/openvswitch directory on host ----------------- 0.91s\nopenvswitch : Copying over ovs-vsctl wrapper ---------------------------- 0.83s\nGroup hosts based on Kolla action --------------------------------------- 0.80s\n\u001b[32m2026-02-24 00:50:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 00:50:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ff4e45b7-3128-4ce1-95d9-909f33d92301 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:50:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e7b4207c-f95e-49ed-815c-ab81afbfeb3f is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:50:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ba7d4726-e54a-4b5b-b060-92175ab177a2 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:50:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:50:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 404270fa-50db-4a78-8c44-21df48acba39 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:50:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 00:50:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ff4e45b7-3128-4ce1-95d9-909f33d92301 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:50:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e7b4207c-f95e-49ed-815c-ab81afbfeb3f is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:50:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ba7d4726-e54a-4b5b-b060-92175ab177a2 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:50:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:50:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 404270fa-50db-4a78-8c44-21df48acba39 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:50:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 00:50:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ff4e45b7-3128-4ce1-95d9-909f33d92301 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:50:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e7b4207c-f95e-49ed-815c-ab81afbfeb3f is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:50:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ba7d4726-e54a-4b5b-b060-92175ab177a2 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:50:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:50:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 404270fa-50db-4a78-8c44-21df48acba39 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:50:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 00:50:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ff4e45b7-3128-4ce1-95d9-909f33d92301 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:50:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e7b4207c-f95e-49ed-815c-ab81afbfeb3f is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:50:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ba7d4726-e54a-4b5b-b060-92175ab177a2 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:50:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:50:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 404270fa-50db-4a78-8c44-21df48acba39 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:50:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 00:50:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ff4e45b7-3128-4ce1-95d9-909f33d92301 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:50:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e7b4207c-f95e-49ed-815c-ab81afbfeb3f is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:50:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ba7d4726-e54a-4b5b-b060-92175ab177a2 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:50:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:50:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 404270fa-50db-4a78-8c44-21df48acba39 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:50:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 00:50:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ff4e45b7-3128-4ce1-95d9-909f33d92301 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:50:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e7b4207c-f95e-49ed-815c-ab81afbfeb3f is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:50:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ba7d4726-e54a-4b5b-b060-92175ab177a2 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:50:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:50:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 404270fa-50db-4a78-8c44-21df48acba39 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:50:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 00:51:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ff4e45b7-3128-4ce1-95d9-909f33d92301 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:51:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e7b4207c-f95e-49ed-815c-ab81afbfeb3f is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:51:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ba7d4726-e54a-4b5b-b060-92175ab177a2 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:51:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:51:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 404270fa-50db-4a78-8c44-21df48acba39 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:51:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 00:51:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ff4e45b7-3128-4ce1-95d9-909f33d92301 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:51:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e7b4207c-f95e-49ed-815c-ab81afbfeb3f is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:51:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ba7d4726-e54a-4b5b-b060-92175ab177a2 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:51:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:51:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 404270fa-50db-4a78-8c44-21df48acba39 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:51:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 00:51:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ff4e45b7-3128-4ce1-95d9-909f33d92301 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:51:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e7b4207c-f95e-49ed-815c-ab81afbfeb3f is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:51:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ba7d4726-e54a-4b5b-b060-92175ab177a2 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:51:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:51:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 404270fa-50db-4a78-8c44-21df48acba39 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:51:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 00:51:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ff4e45b7-3128-4ce1-95d9-909f33d92301 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:51:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e7b4207c-f95e-49ed-815c-ab81afbfeb3f is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:51:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ba7d4726-e54a-4b5b-b060-92175ab177a2 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:51:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:51:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 404270fa-50db-4a78-8c44-21df48acba39 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:51:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 00:51:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ff4e45b7-3128-4ce1-95d9-909f33d92301 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:51:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e7b4207c-f95e-49ed-815c-ab81afbfeb3f is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:51:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ba7d4726-e54a-4b5b-b060-92175ab177a2 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:51:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:51:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 404270fa-50db-4a78-8c44-21df48acba39 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:51:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 00:51:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ff4e45b7-3128-4ce1-95d9-909f33d92301 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:51:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e7b4207c-f95e-49ed-815c-ab81afbfeb3f is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:51:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ba7d4726-e54a-4b5b-b060-92175ab177a2 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:51:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:51:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 404270fa-50db-4a78-8c44-21df48acba39 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:51:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 00:51:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ff4e45b7-3128-4ce1-95d9-909f33d92301 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:51:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e7b4207c-f95e-49ed-815c-ab81afbfeb3f is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:51:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ba7d4726-e54a-4b5b-b060-92175ab177a2 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:51:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:51:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 404270fa-50db-4a78-8c44-21df48acba39 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:51:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 00:51:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ff4e45b7-3128-4ce1-95d9-909f33d92301 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:51:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e7b4207c-f95e-49ed-815c-ab81afbfeb3f is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:51:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ba7d4726-e54a-4b5b-b060-92175ab177a2 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:51:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:51:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 404270fa-50db-4a78-8c44-21df48acba39 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:51:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 00:51:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ff4e45b7-3128-4ce1-95d9-909f33d92301 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:51:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e7b4207c-f95e-49ed-815c-ab81afbfeb3f is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:51:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ba7d4726-e54a-4b5b-b060-92175ab177a2 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:51:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:51:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 404270fa-50db-4a78-8c44-21df48acba39 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:51:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 00:51:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ff4e45b7-3128-4ce1-95d9-909f33d92301 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:51:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e7b4207c-f95e-49ed-815c-ab81afbfeb3f is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:51:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ba7d4726-e54a-4b5b-b060-92175ab177a2 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:51:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:51:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 404270fa-50db-4a78-8c44-21df48acba39 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:51:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 00:51:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ff4e45b7-3128-4ce1-95d9-909f33d92301 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:51:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e7b4207c-f95e-49ed-815c-ab81afbfeb3f is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:51:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ba7d4726-e54a-4b5b-b060-92175ab177a2 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:51:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:51:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 404270fa-50db-4a78-8c44-21df48acba39 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:51:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 00:51:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ff4e45b7-3128-4ce1-95d9-909f33d92301 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:51:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e7b4207c-f95e-49ed-815c-ab81afbfeb3f is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:51:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ba7d4726-e54a-4b5b-b060-92175ab177a2 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:51:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:51:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 404270fa-50db-4a78-8c44-21df48acba39 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:51:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 00:51:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ff4e45b7-3128-4ce1-95d9-909f33d92301 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:51:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e7b4207c-f95e-49ed-815c-ab81afbfeb3f is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:51:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ba7d4726-e54a-4b5b-b060-92175ab177a2 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:51:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:51:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 404270fa-50db-4a78-8c44-21df48acba39 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:51:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 00:51:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ff4e45b7-3128-4ce1-95d9-909f33d92301 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:51:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e7b4207c-f95e-49ed-815c-ab81afbfeb3f is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:51:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ba7d4726-e54a-4b5b-b060-92175ab177a2 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:51:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:51:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 404270fa-50db-4a78-8c44-21df48acba39 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:51:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 00:51:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ff4e45b7-3128-4ce1-95d9-909f33d92301 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:51:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e7b4207c-f95e-49ed-815c-ab81afbfeb3f is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:51:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ba7d4726-e54a-4b5b-b060-92175ab177a2 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:51:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:51:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 404270fa-50db-4a78-8c44-21df48acba39 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:51:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 00:51:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ff4e45b7-3128-4ce1-95d9-909f33d92301 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:51:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e7b4207c-f95e-49ed-815c-ab81afbfeb3f is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:51:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ba7d4726-e54a-4b5b-b060-92175ab177a2 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:51:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:51:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 404270fa-50db-4a78-8c44-21df48acba39 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:51:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 00:51:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ff4e45b7-3128-4ce1-95d9-909f33d92301 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:51:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e7b4207c-f95e-49ed-815c-ab81afbfeb3f is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:51:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ba7d4726-e54a-4b5b-b060-92175ab177a2 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:51:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:51:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 404270fa-50db-4a78-8c44-21df48acba39 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:51:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 00:51:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ff4e45b7-3128-4ce1-95d9-909f33d92301 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:51:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e7b4207c-f95e-49ed-815c-ab81afbfeb3f is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:51:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ba7d4726-e54a-4b5b-b060-92175ab177a2 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:51:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:51:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 404270fa-50db-4a78-8c44-21df48acba39 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:51:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 00:51:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ff4e45b7-3128-4ce1-95d9-909f33d92301 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:51:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e7b4207c-f95e-49ed-815c-ab81afbfeb3f is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:51:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ba7d4726-e54a-4b5b-b060-92175ab177a2 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:51:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:51:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 404270fa-50db-4a78-8c44-21df48acba39 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:51:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 00:51:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ff4e45b7-3128-4ce1-95d9-909f33d92301 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:51:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e7b4207c-f95e-49ed-815c-ab81afbfeb3f is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:51:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ba7d4726-e54a-4b5b-b060-92175ab177a2 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:51:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:51:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 404270fa-50db-4a78-8c44-21df48acba39 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:51:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 00:52:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ff4e45b7-3128-4ce1-95d9-909f33d92301 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:52:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e7b4207c-f95e-49ed-815c-ab81afbfeb3f is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:52:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ba7d4726-e54a-4b5b-b060-92175ab177a2 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:52:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:52:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 404270fa-50db-4a78-8c44-21df48acba39 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:52:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 00:52:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ff4e45b7-3128-4ce1-95d9-909f33d92301 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:52:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e7b4207c-f95e-49ed-815c-ab81afbfeb3f is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:52:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ba7d4726-e54a-4b5b-b060-92175ab177a2 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:52:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:52:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 404270fa-50db-4a78-8c44-21df48acba39 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:52:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 00:52:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ff4e45b7-3128-4ce1-95d9-909f33d92301 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:52:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e7b4207c-f95e-49ed-815c-ab81afbfeb3f is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:52:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ba7d4726-e54a-4b5b-b060-92175ab177a2 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:52:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:52:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 404270fa-50db-4a78-8c44-21df48acba39 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:52:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 00:52:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ff4e45b7-3128-4ce1-95d9-909f33d92301 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:52:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e7b4207c-f95e-49ed-815c-ab81afbfeb3f is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:52:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ba7d4726-e54a-4b5b-b060-92175ab177a2 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:52:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:52:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 404270fa-50db-4a78-8c44-21df48acba39 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:52:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\n\nPLAY [Set kolla_action_rabbitmq] ***********************************************\n\nTASK [Inform the user about the following task] ********************************\nTuesday 24 February 2026  00:49:57 +0000 (0:00:00.086)       0:00:00.086 ****** \n\u001b[0;32mok: [localhost] => {\u001b[0m\n\u001b[0;32m    \"msg\": \"The task 'Check RabbitMQ service' fails if the RabbitMQ service has not yet been deployed. This is fine.\"\u001b[0m\n\u001b[0;32m}\u001b[0m\n\nTASK [Check RabbitMQ service] **************************************************\nTuesday 24 February 2026  00:49:57 +0000 (0:00:00.040)       0:00:00.127 ****** \n\u001b[0;31mfatal: [localhost]: FAILED! => {\"changed\": false, \"elapsed\": 2, \"msg\": \"Timeout when waiting for search string RabbitMQ Management in 192.168.16.9:15672\"}\u001b[0m\n\u001b[0;36m...ignoring\u001b[0m\n\nTASK [Set kolla_action_rabbitmq = upgrade if RabbitMQ is already running] ******\nTuesday 24 February 2026  00:50:00 +0000 (0:00:02.801)       0:00:02.929 ****** \n\u001b[0;36mskipping: [localhost]\u001b[0m\n\nTASK [Set kolla_action_rabbitmq = kolla_action_ng] *****************************\nTuesday 24 February 2026  00:50:00 +0000 (0:00:00.038)       0:00:02.968 ****** \n\u001b[0;32mok: [localhost]\u001b[0m\n\nPLAY [Group hosts based on configuration] **************************************\n\nTASK [Group hosts based on Kolla action] ***************************************\nTuesday 24 February 2026  00:50:00 +0000 (0:00:00.169)       0:00:03.137 ****** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [Group hosts based on enabled services] ***********************************\nTuesday 24 February 2026  00:50:00 +0000 (0:00:00.233)       0:00:03.370 ****** \n\u001b[0;32mok: [testbed-node-0] => (item=enable_rabbitmq_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item=enable_rabbitmq_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item=enable_rabbitmq_True)\u001b[0m\n\nPLAY [Apply role rabbitmq] *****************************************************\n\nTASK [rabbitmq : include_tasks] ************************************************\nTuesday 24 February 2026  00:50:01 +0000 (0:00:00.603)       0:00:03.974 ****** \n\u001b[0;36mincluded: /ansible/roles/rabbitmq/tasks/deploy.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [rabbitmq : Get container facts] ******************************************\nTuesday 24 February 2026  00:50:01 +0000 (0:00:00.593)       0:00:04.567 ****** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [rabbitmq : Get current RabbitMQ version] *********************************\nTuesday 24 February 2026  00:50:03 +0000 (0:00:01.205)       0:00:05.773 ****** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [rabbitmq : Get new RabbitMQ version] *************************************\nTuesday 24 February 2026  00:50:04 +0000 (0:00:00.935)       0:00:06.708 ****** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [rabbitmq : Check if running RabbitMQ is at most one version behind] ******\nTuesday 24 February 2026  00:50:04 +0000 (0:00:00.713)       0:00:07.422 ****** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [rabbitmq : Catch when RabbitMQ is being downgraded] **********************\nTuesday 24 February 2026  00:50:05 +0000 (0:00:00.436)       0:00:07.858 ****** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [rabbitmq : include_tasks] ************************************************\nTuesday 24 February 2026  00:50:07 +0000 (0:00:02.218)       0:00:10.077 ****** \n\u001b[0;36mincluded: /ansible/roles/rabbitmq/tasks/remove-ha-all-policy.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [rabbitmq : Get container facts] ******************************************\nTuesday 24 February 2026  00:50:09 +0000 (0:00:01.657)       0:00:11.735 ****** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [rabbitmq : List RabbitMQ policies] ***************************************\nTuesday 24 February 2026  00:50:09 +0000 (0:00:00.705)       0:00:12.440 ****** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [rabbitmq : Remove ha-all policy from RabbitMQ] ***************************\nTuesday 24 February 2026  00:50:10 +0000 (0:00:00.360)       0:00:12.801 ****** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [rabbitmq : Ensuring config directories exist] ****************************\nTuesday 24 February 2026  00:50:10 +0000 (0:00:00.384)       0:00:13.185 ****** \n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'rabbitmq', 'value': {'container_name': 'rabbitmq', 'group': 'rabbitmq', 'enabled': True, 'image': 'registry.osism.tech/kolla/rabbitmq:2024.2', 'bootstrap_environment': {'KOLLA_BOOTSTRAP': None, 'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'volumes': ['/etc/kolla/rabbitmq/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'rabbitmq:/var/lib/rabbitmq/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_rabbitmq'], 'timeout': '30'}, 'haproxy': {'rabbitmq_management': {'enabled': 'yes', 'mode': 'http', 'port': '15672', 'host_group': 'rabbitmq'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'rabbitmq', 'value': {'container_name': 'rabbitmq', 'group': 'rabbitmq', 'enabled': True, 'image': 'registry.osism.tech/kolla/rabbitmq:2024.2', 'bootstrap_environment': {'KOLLA_BOOTSTRAP': None, 'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'volumes': ['/etc/kolla/rabbitmq/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'rabbitmq:/var/lib/rabbitmq/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_rabbitmq'], 'timeout': '30'}, 'haproxy': {'rabbitmq_management': {'enabled': 'yes', 'mode': 'http', 'port': '15672', 'host_group': 'rabbitmq'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'rabbitmq', 'value': {'container_name': 'rabbitmq', 'group': 'rabbitmq', 'enabled': True, 'image': 'registry.osism.tech/kolla/rabbitmq:2024.2', 'bootstrap_environment': {'KOLLA_BOOTSTRAP': None, 'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'volumes': ['/etc/kolla/rabbitmq/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'rabbitmq:/var/lib/rabbitmq/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_rabbitmq'], 'timeout': '30'}, 'haproxy': {'rabbitmq_management': {'enabled': 'yes', 'mode': 'http', 'port': '15672', 'host_group': 'rabbitmq'}}}})\u001b[0m\n\nTASK [rabbitmq : Copying over config.json files for services] ******************\nTuesday 24 February 2026  00:50:11 +0000 (0:00:01.189)       0:00:14.375 ****** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'rabbitmq', 'value': {'container_name': 'rabbitmq', 'group': 'rabbitmq', 'enabled': True, 'image': 'registry.osism.tech/kolla/rabbitmq:2024.2', 'bootstrap_environment': {'KOLLA_BOOTSTRAP': None, 'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'volumes': ['/etc/kolla/rabbitmq/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'rabbitmq:/var/lib/rabbitmq/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_rabbitmq'], 'timeout': '30'}, 'haproxy': {'rabbitmq_management': {'enabled': 'yes', 'mode': 'http', 'port': '15672', 'host_group': 'rabbitmq'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'rabbitmq', 'value': {'container_name': 'rabbitmq', 'group': 'rabbitmq', 'enabled': True, 'image': 'registry.osism.tech/kolla/rabbitmq:2024.2', 'bootstrap_environment': {'KOLLA_BOOTSTRAP': None, 'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'volumes': ['/etc/kolla/rabbitmq/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'rabbitmq:/var/lib/rabbitmq/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_rabbitmq'], 'timeout': '30'}, 'haproxy': {'rabbitmq_management': {'enabled': 'yes', 'mode': 'http', 'port': '15672', 'host_group': 'rabbitmq'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'rabbitmq', 'value': {'container_name': 'rabbitmq', 'group': 'rabbitmq', 'enabled': True, 'image': 'registry.osism.tech/kolla/rabbitmq:2024.2', 'bootstrap_environment': {'KOLLA_BOOTSTRAP': None, 'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'volumes': ['/etc/kolla/rabbitmq/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'rabbitmq:/var/lib/rabbitmq/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_rabbitmq'], 'timeout': '30'}, 'haproxy': {'rabbitmq_management': {'enabled': 'yes', 'mode': 'http', 'port': '15672', 'host_group': 'rabbitmq'}}}})\u001b[0m\n\nTASK [rabbitmq : Copying over rabbitmq-env.conf] *******************************\nTuesday 24 February 2026  00:50:14 +0000 (0:00:02.430)       0:00:16.806 ****** \n\u001b[0;33mchanged: [testbed-node-0] => (item=/ansible/roles/rabbitmq/templates/rabbitmq-env.conf.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=/ansible/roles/rabbitmq/templates/rabbitmq-env.conf.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=/ansible/roles/rabbitmq/templates/rabbitmq-env.conf.j2)\u001b[0m\n\nTASK [rabbitmq : Copying over rabbitmq.conf] ***********************************\nTuesday 24 February 2026  00:50:16 +0000 (0:00:01.812)       0:00:18.618 ****** \n\u001b[0;33mchanged: [testbed-node-0] => (item=/ansible/roles/rabbitmq/templates/rabbitmq.conf.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=/ansible/roles/rabbitmq/templates/rabbitmq.conf.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=/ansible/roles/rabbitmq/templates/rabbitmq.conf.j2)\u001b[0m\n\nTASK [rabbitmq : Copying over erl_inetrc] **************************************\nTuesday 24 February 2026  00:50:17 +0000 (0:00:01.843)       0:00:20.462 ****** \n\u001b[0;33mchanged: [testbed-node-0] => (item=/ansible/roles/rabbitmq/templates/erl_inetrc.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=/ansible/roles/rabbitmq/templates/erl_inetrc.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=/ansible/roles/rabbitmq/templates/erl_inetrc.j2)\u001b[0m\n\nTASK [rabbitmq : Copying over advanced.config] *********************************\nTuesday 24 February 2026  00:50:19 +0000 (0:00:01.294)       0:00:21.756 ****** \n\u001b[0;33mchanged: [testbed-node-0] => (item=/ansible/roles/rabbitmq/templates/advanced.config.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=/ansible/roles/rabbitmq/templates/advanced.config.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=/ansible/roles/rabbitmq/templates/advanced.config.j2)\u001b[0m\n\nTASK [rabbitmq : Copying over definitions.json] ********************************\nTuesday 24 February 2026  00:50:20 +0000 (0:00:01.697)       0:00:23.454 ****** \n\u001b[0;33mchanged: [testbed-node-0] => (item=/ansible/roles/rabbitmq/templates/definitions.json.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=/ansible/roles/rabbitmq/templates/definitions.json.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=/ansible/roles/rabbitmq/templates/definitions.json.j2)\u001b[0m\n\nTASK [rabbitmq : Copying over enabled_plugins] *********************************\nTuesday 24 February 2026  00:50:22 +0000 (0:00:01.489)       0:00:24.943 ****** \n\u001b[0;33mchanged: [testbed-node-0] => (item=/ansible/roles/rabbitmq/templates/enabled_plugins.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=/ansible/roles/rabbitmq/templates/enabled_plugins.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=/ansible/roles/rabbitmq/templates/enabled_plugins.j2)\u001b[0m\n\nTASK [rabbitmq : include_tasks] ************************************************\nTuesday 24 February 2026  00:50:23 +0000 (0:00:01.649)       0:00:26.592 ****** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [rabbitmq : Check rabbitmq containers] ************************************\nTuesday 24 February 2026  00:50:24 +0000 (0:00:00.427)       0:00:27.019 ****** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'rabbitmq', 'value': {'container_name': 'rabbitmq', 'group': 'rabbitmq', 'enabled': True, 'image': 'registry.osism.tech/kolla/rabbitmq:2024.2', 'bootstrap_environment': {'KOLLA_BOOTSTRAP': None, 'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'volumes': ['/etc/kolla/rabbitmq/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'rabbitmq:/var/lib/rabbitmq/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_rabbitmq'], 'timeout': '30'}, 'haproxy': {'rabbitmq_management': {'enabled': 'yes', 'mode': 'http', 'port': '15672', 'host_group': 'rabbitmq'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'rabbitmq', 'value': {'container_name': 'rabbitmq', 'group': 'rabbitmq', 'enabled': True, 'image': 'registry.osism.tech/kolla/rabbitmq:2024.2', 'bootstrap_environment': {'KOLLA_BOOTSTRAP': None, 'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'volumes': ['/etc/kolla/rabbitmq/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'rabbitmq:/var/lib/rabbitmq/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_rabbitmq'], 'timeout': '30'}, 'haproxy': {'rabbitmq_management': {'enabled': 'yes', 'mode': 'http', 'port': '15672', 'host_group': 'rabbitmq'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'rabbitmq', 'value': {'container_name': 'rabbitmq', 'group': 'rabbitmq', 'enabled': True, 'image': 'registry.osism.tech/kolla/rabbitmq:2024.2', 'bootstrap_environment': {'KOLLA_BOOTSTRAP': None, 'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'volumes': ['/etc/kolla/rabbitmq/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'rabbitmq:/var/lib/rabbitmq/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_rabbitmq'], 'timeout': '30'}, 'haproxy': {'rabbitmq_management': {'enabled': 'yes', 'mode': 'http', 'port': '15672', 'host_group': 'rabbitmq'}}}})\u001b[0m\n\nTASK [rabbitmq : Creating rabbitmq volume] *************************************\nTuesday 24 February 2026  00:50:26 +0000 (0:00:02.066)       0:00:29.086 ****** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [rabbitmq : Running RabbitMQ bootstrap container] *************************\nTuesday 24 February 2026  00:50:27 +0000 (0:00:01.174)       0:00:30.260 ****** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\nRUNNING HANDLER [rabbitmq : Restart rabbitmq container] ************************\nTuesday 24 February 2026  00:50:35 +0000 (0:00:07.674)       0:00:37.934 ****** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nPLAY [Restart rabbitmq services] ***********************************************\n\nTASK [rabbitmq : Get info on RabbitMQ container] *******************************\nTuesday 24 February 2026  00:50:35 +0000 (0:00:00.282)       0:00:38.217 ****** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [rabbitmq : Put RabbitMQ node into maintenance mode] **********************\nTuesday 24 February 2026  00:50:36 +0000 (0:00:00.614)       0:00:38.831 ****** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [rabbitmq : Restart rabbitmq container] ***********************************\nTuesday 24 February 2026  00:50:36 +0000 (0:00:00.213)       0:00:39.044 ****** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [rabbitmq : Waiting for rabbitmq to start] ********************************\nTuesday 24 February 2026  00:50:38 +0000 (0:00:02.185)       0:00:41.230 ****** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nPLAY [Restart rabbitmq services] ***********************************************\n\nTASK [rabbitmq : Get info on RabbitMQ container] *******************************\nTuesday 24 February 2026  00:51:34 +0000 (0:00:55.674)       0:01:36.905 ****** \n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\nTASK [rabbitmq : Put RabbitMQ node into maintenance mode] **********************\nTuesday 24 February 2026  00:51:35 +0000 (0:00:00.758)       0:01:37.664 ****** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [rabbitmq : Restart rabbitmq container] ***********************************\nTuesday 24 February 2026  00:51:35 +0000 (0:00:00.237)       0:01:37.901 ****** \n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\nTASK [rabbitmq : Waiting for rabbitmq to start] ********************************\nTuesday 24 February 2026  00:51:37 +0000 (0:00:02.552)       0:01:40.454 ****** \n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\nPLAY [Restart rabbitmq services] ***********************************************\n\nTASK [rabbitmq : Get info on RabbitMQ container] *******************************\nTuesday 24 February 2026  00:51:52 +0000 (0:00:14.407)       0:01:54.862 ****** \n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [rabbitmq : Put RabbitMQ node into maintenance mode] **********************\nTuesday 24 February 2026  00:51:52 +0000 (0:00:00.728)       0:01:55.590 ****** \n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [rabbitmq : Restart rabbitmq container] ***********************************\nTuesday 24 February 2026  00:51:53 +0000 (0:00:00.351)       0:01:55.941 ****** \n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [rabbitmq : Waiting for rabbitmq to start] ********************************\nTuesday 24 February 2026  00:52:00 +0000 (0:00:06.886)       0:02:02.828 ****** \n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nPLAY [Apply rabbitmq post-configuration] ***************************************\n\nTASK [Include rabbitmq post-deploy.yml] ****************************************\nTuesday 24 February 2026  00:52:10 +0000 (0:00:10.477)       0:02:13.306 ****** \n\u001b[0;36mincluded: rabbitmq for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [rabbitmq : Enable all stable feature flags] ******************************\nTuesday 24 February 2026  00:52:11 +0000 (0:00:00.731)       0:02:14.037 ****** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[1;35m[WARNING]: Could not match supplied host pattern, ignoring:\u001b[0m\n\u001b[1;35menable_outward_rabbitmq_True\u001b[0m\n\nPLAY [Apply role rabbitmq (outward)] *******************************************\n\u001b[0;36mskipping: no hosts matched\u001b[0m\n\u001b[1;35m[WARNING]: Could not match supplied host pattern, ignoring:\u001b[0m\n\u001b[1;35moutward_rabbitmq_restart\u001b[0m\n\nPLAY [Restart rabbitmq (outward) services] *************************************\n\u001b[0;36mskipping: no hosts matched\u001b[0m\n\nPLAY [Apply rabbitmq (outward) post-configuration] *****************************\n\u001b[0;36mskipping: no hosts matched\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;32mlocalhost\u001b[0m                  : \u001b[0;32mok=3   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=1   \u001b[0m rescued=0    \u001b[1;35mignored=1   \u001b[0m\n\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=23  \u001b[0m \u001b[0;33mchanged=14  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=8   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=21  \u001b[0m \u001b[0;33mchanged=14  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=21  \u001b[0m \u001b[0;33mchanged=14  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nTuesday 24 February 2026  00:52:13 +0000 (0:00:02.397)       0:02:16.434 ****** \n=============================================================================== \nrabbitmq : Waiting for rabbitmq to start ------------------------------- 80.56s\nrabbitmq : Restart rabbitmq container ---------------------------------- 11.63s\nrabbitmq : Running RabbitMQ bootstrap container ------------------------- 7.67s\nCheck RabbitMQ service -------------------------------------------------- 2.80s\nrabbitmq : Copying over config.json files for services ------------------ 2.43s\nrabbitmq : Enable all stable feature flags ------------------------------ 2.40s\nrabbitmq : Catch when RabbitMQ is being downgraded ---------------------- 2.22s\nrabbitmq : Get info on RabbitMQ container ------------------------------- 2.10s\nrabbitmq : Check rabbitmq containers ------------------------------------ 2.07s\nrabbitmq : Copying over rabbitmq.conf ----------------------------------- 1.84s\nrabbitmq : Copying over rabbitmq-env.conf ------------------------------- 1.81s\nrabbitmq : Copying over advanced.config --------------------------------- 1.70s\nrabbitmq : include_tasks ------------------------------------------------ 1.66s\nrabbitmq : Copying over enabled_plugins --------------------------------- 1.65s\nrabbitmq : Copying over definitions.json -------------------------------- 1.49s\nrabbitmq : Copying over erl_inetrc -------------------------------------- 1.29s\nrabbitmq : Get container facts ------------------------------------------ 1.21s\nrabbitmq : Ensuring config directories exist ---------------------------- 1.19s\nrabbitmq : Creating rabbitmq volume ------------------------------------- 1.17s\nrabbitmq : Get current RabbitMQ version --------------------------------- 0.94s\n\u001b[32m2026-02-24 00:52:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ff4e45b7-3128-4ce1-95d9-909f33d92301 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:52:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e7b4207c-f95e-49ed-815c-ab81afbfeb3f is in state SUCCESS\u001b[0m\n\u001b[32m2026-02-24 00:52:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ba7d4726-e54a-4b5b-b060-92175ab177a2 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:52:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:52:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 404270fa-50db-4a78-8c44-21df48acba39 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:52:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\n\u001b[32m2026-02-24 00:52:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ff4e45b7-3128-4ce1-95d9-909f33d92301 is in state SUCCESS\u001b[0m\n\nPLAY [Prepare all k3s nodes] ***************************************************\n\nTASK [k3s_prereq : Validating arguments against arg spec 'main' - Prerequisites] ***\nTuesday 24 February 2026  00:47:14 +0000 (0:00:00.200)       0:00:00.200 ****** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [k3s_prereq : Set same timezone on every Server] **************************\nTuesday 24 February 2026  00:47:15 +0000 (0:00:00.745)       0:00:00.946 ****** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [k3s_prereq : Set SELinux to disabled state] ******************************\nTuesday 24 February 2026  00:47:16 +0000 (0:00:00.757)       0:00:01.703 ****** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [k3s_prereq : Enable IPv4 forwarding] *************************************\nTuesday 24 February 2026  00:47:16 +0000 (0:00:00.778)       0:00:02.482 ****** \n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [k3s_prereq : Enable IPv6 forwarding] *************************************\nTuesday 24 February 2026  00:47:19 +0000 (0:00:02.159)       0:00:04.642 ****** \n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [k3s_prereq : Enable IPv6 router advertisements] **************************\nTuesday 24 February 2026  00:47:20 +0000 (0:00:01.258)       0:00:05.900 ****** \n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\nTASK [k3s_prereq : Add br_netfilter to /etc/modules-load.d/] *******************\nTuesday 24 February 2026  00:47:22 +0000 (0:00:01.954)       0:00:07.854 ****** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [k3s_prereq : Load br_netfilter] ******************************************\nTuesday 24 February 2026  00:47:23 +0000 (0:00:00.839)       0:00:08.694 ****** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [k3s_prereq : Set bridge-nf-call-iptables (just to be sure)] **************\nTuesday 24 February 2026  00:47:24 +0000 (0:00:00.837)       0:00:09.532 ****** \n\u001b[0;36mskipping: [testbed-node-3] => (item=net.bridge.bridge-nf-call-iptables) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=net.bridge.bridge-nf-call-ip6tables) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=net.bridge.bridge-nf-call-iptables) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=net.bridge.bridge-nf-call-ip6tables) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=net.bridge.bridge-nf-call-iptables) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=net.bridge.bridge-nf-call-ip6tables) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=net.bridge.bridge-nf-call-iptables) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=net.bridge.bridge-nf-call-ip6tables) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=net.bridge.bridge-nf-call-iptables) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=net.bridge.bridge-nf-call-ip6tables) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=net.bridge.bridge-nf-call-iptables) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=net.bridge.bridge-nf-call-ip6tables) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [k3s_prereq : Add /usr/local/bin to sudo secure_path] *********************\nTuesday 24 February 2026  00:47:24 +0000 (0:00:00.889)       0:00:10.422 ****** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [k3s_download : Validating arguments against arg spec 'main' - Manage the downloading of K3S binaries] ***\nTuesday 24 February 2026  00:47:26 +0000 (0:00:01.499)       0:00:11.921 ****** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [k3s_download : Download k3s binary x64] **********************************\nTuesday 24 February 2026  00:47:27 +0000 (0:00:01.145)       0:00:13.067 ****** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\nTASK [k3s_download : Download k3s binary arm64] ********************************\nTuesday 24 February 2026  00:47:33 +0000 (0:00:05.707)       0:00:18.774 ****** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [k3s_download : Download k3s binary armhf] ********************************\nTuesday 24 February 2026  00:47:35 +0000 (0:00:01.833)       0:00:20.608 ****** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [k3s_custom_registries : Validating arguments against arg spec 'main' - Configure the use of a custom container registry] ***\nTuesday 24 February 2026  00:47:37 +0000 (0:00:02.094)       0:00:22.702 ****** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [k3s_custom_registries : Create directory /etc/rancher/k3s] ***************\nTuesday 24 February 2026  00:47:38 +0000 (0:00:01.638)       0:00:24.340 ****** \n\u001b[0;36mskipping: [testbed-node-3] => (item=rancher) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=rancher/k3s) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=rancher) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=rancher/k3s) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=rancher) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=rancher/k3s) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=rancher) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=rancher/k3s) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=rancher) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=rancher/k3s) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=rancher) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=rancher/k3s) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [k3s_custom_registries : Insert registries into /etc/rancher/k3s/registries.yaml] ***\nTuesday 24 February 2026  00:47:40 +0000 (0:00:01.815)       0:00:26.155 ****** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [k3s_custom_registries : Remove /etc/rancher/k3s/registries.yaml when no registries configured] ***\nTuesday 24 February 2026  00:47:41 +0000 (0:00:00.953)       0:00:27.109 ****** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nPLAY [Deploy k3s master nodes] *************************************************\n\nTASK [k3s_server : Validating arguments against arg spec 'main' - Setup k3s servers] ***\nTuesday 24 February 2026  00:47:43 +0000 (0:00:01.457)       0:00:28.567 ****** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [k3s_server : Stop k3s-init] **********************************************\nTuesday 24 February 2026  00:47:44 +0000 (0:00:01.351)       0:00:29.919 ****** \n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [k3s_server : Stop k3s] ***************************************************\nTuesday 24 February 2026  00:47:45 +0000 (0:00:01.528)       0:00:31.447 ****** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [k3s_server : Clean previous runs of k3s-init] ****************************\nTuesday 24 February 2026  00:47:47 +0000 (0:00:01.110)       0:00:32.557 ****** \n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [k3s_server : Deploy K3s http_proxy conf] *********************************\nTuesday 24 February 2026  00:47:48 +0000 (0:00:01.019)       0:00:33.576 ****** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [k3s_server : Create /etc/rancher/k3s directory] **************************\nTuesday 24 February 2026  00:47:48 +0000 (0:00:00.498)       0:00:34.075 ****** \n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [k3s_server : Create custom resolv.conf for k3s] **************************\nTuesday 24 February 2026  00:47:49 +0000 (0:00:01.037)       0:00:35.113 ****** \n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [k3s_server : Deploy vip manifest] ****************************************\nTuesday 24 February 2026  00:47:51 +0000 (0:00:01.960)       0:00:37.073 ****** \n\u001b[0;36mincluded: /ansible/roles/k3s_server/tasks/vip.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [k3s_server : Set _kube_vip_bgp_peers fact] *******************************\nTuesday 24 February 2026  00:47:52 +0000 (0:00:00.618)       0:00:37.692 ****** \n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\nTASK [k3s_server : Create manifests directory on first master] *****************\nTuesday 24 February 2026  00:47:55 +0000 (0:00:02.833)       0:00:40.525 ****** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [k3s_server : Download vip rbac manifest to first master] *****************\nTuesday 24 February 2026  00:47:56 +0000 (0:00:01.085)       0:00:41.611 ****** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [k3s_server : Copy vip manifest to first master] **************************\nTuesday 24 February 2026  00:47:57 +0000 (0:00:01.741)       0:00:43.353 ****** \n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [k3s_server : Deploy metallb manifest] ************************************\nTuesday 24 February 2026  00:47:59 +0000 (0:00:01.809)       0:00:45.162 ****** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [k3s_server : Deploy kube-vip manifest] ***********************************\nTuesday 24 February 2026  00:48:00 +0000 (0:00:00.689)       0:00:45.852 ****** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [k3s_server : Init cluster inside the transient k3s-init service] *********\nTuesday 24 February 2026  00:48:00 +0000 (0:00:00.420)       0:00:46.272 ****** \n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [k3s_server : Detect Kubernetes version for label compatibility] **********\nTuesday 24 February 2026  00:48:03 +0000 (0:00:02.332)       0:00:48.604 ****** \n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [k3s_server : Set node role label selector based on Kubernetes version] ***\nTuesday 24 February 2026  00:48:05 +0000 (0:00:02.802)       0:00:51.407 ****** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [k3s_server : Verify that all nodes actually joined (check k3s-init.service if this fails)] ***\nTuesday 24 February 2026  00:48:06 +0000 (0:00:00.862)       0:00:52.270 ****** \n\u001b[1;30mFAILED - RETRYING: [testbed-node-1]: Verify that all nodes actually joined (check k3s-init.service if this fails) (20 retries left).\u001b[0m\n\u001b[1;30mFAILED - RETRYING: [testbed-node-0]: Verify that all nodes actually joined (check k3s-init.service if this fails) (20 retries left).\u001b[0m\n\u001b[1;30mFAILED - RETRYING: [testbed-node-2]: Verify that all nodes actually joined (check k3s-init.service if this fails) (20 retries left).\u001b[0m\n\u001b[1;30mFAILED - RETRYING: [testbed-node-0]: Verify that all nodes actually joined (check k3s-init.service if this fails) (19 retries left).\u001b[0m\n\u001b[1;30mFAILED - RETRYING: [testbed-node-1]: Verify that all nodes actually joined (check k3s-init.service if this fails) (19 retries left).\u001b[0m\n\u001b[1;30mFAILED - RETRYING: [testbed-node-2]: Verify that all nodes actually joined (check k3s-init.service if this fails) (19 retries left).\u001b[0m\n\u001b[1;30mFAILED - RETRYING: [testbed-node-0]: Verify that all nodes actually joined (check k3s-init.service if this fails) (18 retries left).\u001b[0m\n\u001b[1;30mFAILED - RETRYING: [testbed-node-1]: Verify that all nodes actually joined (check k3s-init.service if this fails) (18 retries left).\u001b[0m\n\u001b[1;30mFAILED - RETRYING: [testbed-node-2]: Verify that all nodes actually joined (check k3s-init.service if this fails) (18 retries left).\u001b[0m\n\u001b[1;30mFAILED - RETRYING: [testbed-node-0]: Verify that all nodes actually joined (check k3s-init.service if this fails) (17 retries left).\u001b[0m\n\u001b[1;30mFAILED - RETRYING: [testbed-node-1]: Verify that all nodes actually joined (check k3s-init.service if this fails) (17 retries left).\u001b[0m\n\u001b[1;30mFAILED - RETRYING: [testbed-node-2]: Verify that all nodes actually joined (check k3s-init.service if this fails) (17 retries left).\u001b[0m\n\u001b[1;30mFAILED - RETRYING: [testbed-node-0]: Verify that all nodes actually joined (check k3s-init.service if this fails) (16 retries left).\u001b[0m\n\u001b[1;30mFAILED - RETRYING: [testbed-node-1]: Verify that all nodes actually joined (check k3s-init.service if this fails) (16 retries left).\u001b[0m\n\u001b[1;30mFAILED - RETRYING: [testbed-node-2]: Verify that all nodes actually joined (check k3s-init.service if this fails) (16 retries left).\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\nTASK [k3s_server : Save logs of k3s-init.service] ******************************\nTuesday 24 February 2026  00:49:00 +0000 (0:00:54.020)       0:01:46.290 ****** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [k3s_server : Kill the temporary service used for initialization] *********\nTuesday 24 February 2026  00:49:01 +0000 (0:00:00.623)       0:01:46.913 ****** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [k3s_server : Copy K3s service file] **************************************\nTuesday 24 February 2026  00:49:02 +0000 (0:00:01.184)       0:01:48.099 ****** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [k3s_server : Enable and check K3s service] *******************************\nTuesday 24 February 2026  00:49:03 +0000 (0:00:01.320)       0:01:49.419 ****** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\nTASK [k3s_server : Wait for node-token] ****************************************\nTuesday 24 February 2026  00:49:57 +0000 (0:00:53.123)       0:02:42.542 ****** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\nTASK [k3s_server : Register node-token file access mode] ***********************\nTuesday 24 February 2026  00:49:57 +0000 (0:00:00.633)       0:02:43.176 ****** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\nTASK [k3s_server : Change file access node-token] ******************************\nTuesday 24 February 2026  00:49:58 +0000 (0:00:00.632)       0:02:43.809 ****** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [k3s_server : Read node-token from master] ********************************\nTuesday 24 February 2026  00:49:58 +0000 (0:00:00.534)       0:02:44.343 ****** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [k3s_server : Store Master node-token] ************************************\nTuesday 24 February 2026  00:49:59 +0000 (0:00:00.760)       0:02:45.103 ****** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [k3s_server : Restore node-token file access] *****************************\nTuesday 24 February 2026  00:49:59 +0000 (0:00:00.340)       0:02:45.443 ****** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [k3s_server : Create directory .kube] *************************************\nTuesday 24 February 2026  00:50:00 +0000 (0:00:00.636)       0:02:46.080 ****** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [k3s_server : Copy config file to user home directory] ********************\nTuesday 24 February 2026  00:50:01 +0000 (0:00:00.563)       0:02:46.644 ****** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [k3s_server : Configure kubectl cluster to https://192.168.16.8:6443] *****\nTuesday 24 February 2026  00:50:01 +0000 (0:00:00.852)       0:02:47.497 ****** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [k3s_server : Create kubectl symlink] *************************************\nTuesday 24 February 2026  00:50:02 +0000 (0:00:00.665)       0:02:48.163 ****** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [k3s_server : Create crictl symlink] **************************************\nTuesday 24 February 2026  00:50:02 +0000 (0:00:00.326)       0:02:48.490 ****** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [k3s_server : Get contents of manifests folder] ***************************\nTuesday 24 February 2026  00:50:03 +0000 (0:00:00.391)       0:02:48.881 ****** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [k3s_server : Get sub dirs of manifests folder] ***************************\nTuesday 24 February 2026  00:50:04 +0000 (0:00:01.028)       0:02:49.910 ****** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [k3s_server : Remove manifests and folders that are only needed for bootstrapping cluster so k3s doesn't auto apply on start] ***\nTuesday 24 February 2026  00:50:05 +0000 (0:00:00.678)       0:02:50.588 ****** \n\u001b[0;33mchanged: [testbed-node-0] => (item=/var/lib/rancher/k3s/server/manifests/rolebindings.yaml)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=/var/lib/rancher/k3s/server/manifests/rolebindings.yaml)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=/var/lib/rancher/k3s/server/manifests/rolebindings.yaml)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=/var/lib/rancher/k3s/server/manifests/local-storage.yaml)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=/var/lib/rancher/k3s/server/manifests/local-storage.yaml)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=/var/lib/rancher/k3s/server/manifests/local-storage.yaml)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=/var/lib/rancher/k3s/server/manifests/coredns.yaml)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=/var/lib/rancher/k3s/server/manifests/coredns.yaml)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=/var/lib/rancher/k3s/server/manifests/vip.yaml)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=/var/lib/rancher/k3s/server/manifests/coredns.yaml)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=/var/lib/rancher/k3s/server/manifests/runtimes.yaml)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=/var/lib/rancher/k3s/server/manifests/vip-rbac.yaml)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=/var/lib/rancher/k3s/server/manifests/runtimes.yaml)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=/var/lib/rancher/k3s/server/manifests/ccm.yaml)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=/var/lib/rancher/k3s/server/manifests/runtimes.yaml)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=/var/lib/rancher/k3s/server/manifests/ccm.yaml)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=/var/lib/rancher/k3s/server/manifests/ccm.yaml)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=/var/lib/rancher/k3s/server/manifests/metrics-server)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=/var/lib/rancher/k3s/server/manifests/metrics-server)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=/var/lib/rancher/k3s/server/manifests/metrics-server)\u001b[0m\n\nPLAY [Deploy k3s worker nodes] *************************************************\n\nTASK [k3s_agent : Validating arguments against arg spec 'main' - Setup k3s agents] ***\nTuesday 24 February 2026  00:50:08 +0000 (0:00:03.029)       0:02:53.617 ****** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [k3s_agent : Check if system is PXE-booted] *******************************\nTuesday 24 February 2026  00:50:08 +0000 (0:00:00.489)       0:02:54.107 ****** \n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\nTASK [k3s_agent : Set fact for PXE-booted system] ******************************\nTuesday 24 February 2026  00:50:10 +0000 (0:00:01.680)       0:02:55.787 ****** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [k3s_agent : Include http_proxy configuration tasks] **********************\nTuesday 24 February 2026  00:50:10 +0000 (0:00:00.397)       0:02:56.185 ****** \n\u001b[0;36mincluded: /ansible/roles/k3s_agent/tasks/http_proxy.yml for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m\n\nTASK [k3s_agent : Create k3s-node.service.d directory] *************************\nTuesday 24 February 2026  00:50:11 +0000 (0:00:00.700)       0:02:56.885 ****** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [k3s_agent : Copy K3s http_proxy conf file] *******************************\nTuesday 24 February 2026  00:50:11 +0000 (0:00:00.228)       0:02:57.113 ****** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [k3s_agent : Deploy K3s http_proxy conf] **********************************\nTuesday 24 February 2026  00:50:11 +0000 (0:00:00.286)       0:02:57.400 ****** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [k3s_agent : Create /etc/rancher/k3s directory] ***************************\nTuesday 24 February 2026  00:50:12 +0000 (0:00:00.279)       0:02:57.680 ****** \n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nTASK [k3s_agent : Create custom resolv.conf for k3s] ***************************\nTuesday 24 February 2026  00:50:13 +0000 (0:00:00.861)       0:02:58.541 ****** \n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nTASK [k3s_agent : Configure the k3s service] ***********************************\nTuesday 24 February 2026  00:50:14 +0000 (0:00:01.366)       0:02:59.908 ****** \n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nTASK [k3s_agent : Manage k3s service] ******************************************\nTuesday 24 February 2026  00:50:16 +0000 (0:00:01.848)       0:03:01.757 ****** \n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\nPLAY [Prepare kubeconfig file] *************************************************\n\nTASK [Get home directory of operator user] *************************************\nTuesday 24 February 2026  00:50:26 +0000 (0:00:10.203)       0:03:11.961 ****** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\nTASK [Create .kube directory] **************************************************\nTuesday 24 February 2026  00:50:27 +0000 (0:00:00.753)       0:03:12.714 ****** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nTASK [Get kubeconfig file] *****************************************************\nTuesday 24 February 2026  00:50:27 +0000 (0:00:00.397)       0:03:13.112 ****** \n\u001b[0;32mok: [testbed-manager -> testbed-node-0(192.168.16.10)]\u001b[0m\n\nTASK [Write kubeconfig file] ***************************************************\nTuesday 24 February 2026  00:50:28 +0000 (0:00:00.543)       0:03:13.656 ****** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nTASK [Change server address in the kubeconfig] *********************************\nTuesday 24 February 2026  00:50:28 +0000 (0:00:00.749)       0:03:14.406 ****** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nTASK [Make kubeconfig available for use inside the manager service] ************\nTuesday 24 February 2026  00:50:29 +0000 (0:00:00.519)       0:03:14.926 ****** \n\u001b[0;33mchanged: [testbed-manager -> localhost]\u001b[0m\n\nTASK [Change server address in the kubeconfig inside the manager service] ******\nTuesday 24 February 2026  00:50:31 +0000 (0:00:01.615)       0:03:16.541 ****** \n\u001b[0;33mchanged: [testbed-manager -> localhost]\u001b[0m\n\nTASK [Set KUBECONFIG environment variable] *************************************\nTuesday 24 February 2026  00:50:31 +0000 (0:00:00.829)       0:03:17.371 ****** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nTASK [Enable kubectl command line completion] **********************************\nTuesday 24 February 2026  00:50:32 +0000 (0:00:00.492)       0:03:17.863 ****** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nPLAY [Apply role kubectl] ******************************************************\n\nTASK [kubectl : Gather variables for each operating system] ********************\nTuesday 24 February 2026  00:50:32 +0000 (0:00:00.420)       0:03:18.283 ****** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\nTASK [kubectl : Include distribution specific install tasks] *******************\nTuesday 24 February 2026  00:50:32 +0000 (0:00:00.129)       0:03:18.412 ****** \n\u001b[0;36mincluded: /ansible/roles/kubectl/tasks/install-Debian-family.yml for testbed-manager\u001b[0m\n\nTASK [kubectl : Remove old architecture-dependent repository] ******************\nTuesday 24 February 2026  00:50:33 +0000 (0:00:00.229)       0:03:18.642 ****** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\nTASK [kubectl : Install apt-transport-https package] ***************************\nTuesday 24 February 2026  00:50:33 +0000 (0:00:00.838)       0:03:19.481 ****** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\nTASK [kubectl : Add repository gpg key] ****************************************\nTuesday 24 February 2026  00:50:35 +0000 (0:00:01.468)       0:03:20.949 ****** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nTASK [kubectl : Set permissions of gpg key] ************************************\nTuesday 24 February 2026  00:50:36 +0000 (0:00:00.718)       0:03:21.667 ****** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\nTASK [kubectl : Add repository Debian] *****************************************\nTuesday 24 February 2026  00:50:36 +0000 (0:00:00.423)       0:03:22.091 ****** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nTASK [kubectl : Install required packages] *************************************\nTuesday 24 February 2026  00:50:43 +0000 (0:00:06.676)       0:03:28.768 ****** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nTASK [kubectl : Remove kubectl symlink] ****************************************\nTuesday 24 February 2026  00:50:55 +0000 (0:00:12.522)       0:03:41.291 ****** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\nPLAY [Run post actions on master nodes] ****************************************\n\nTASK [k3s_server_post : Validating arguments against arg spec 'main' - Configure k3s cluster] ***\nTuesday 24 February 2026  00:50:56 +0000 (0:00:00.520)       0:03:41.811 ****** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [k3s_server_post : Deploy calico] *****************************************\nTuesday 24 February 2026  00:50:56 +0000 (0:00:00.306)       0:03:42.118 ****** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [k3s_server_post : Deploy cilium] *****************************************\nTuesday 24 February 2026  00:50:57 +0000 (0:00:00.616)       0:03:42.735 ****** \n\u001b[0;36mincluded: /ansible/roles/k3s_server_post/tasks/cilium.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [k3s_server_post : Create tmp directory on first master] ******************\nTuesday 24 February 2026  00:50:57 +0000 (0:00:00.763)       0:03:43.498 ****** \n\u001b[0;33mchanged: [testbed-node-0 -> localhost]\u001b[0m\n\nTASK [k3s_server_post : Wait for connectivity to kube VIP] *********************\nTuesday 24 February 2026  00:50:59 +0000 (0:00:01.044)       0:03:44.543 ****** \n\u001b[0;32mok: [testbed-node-0 -> localhost]\u001b[0m\n\nTASK [k3s_server_post : Fail if kube VIP not reachable] ************************\nTuesday 24 February 2026  00:50:59 +0000 (0:00:00.873)       0:03:45.416 ****** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [k3s_server_post : Test for existing Cilium install] **********************\nTuesday 24 February 2026  00:51:00 +0000 (0:00:00.121)       0:03:45.538 ****** \n\u001b[0;32mok: [testbed-node-0 -> localhost]\u001b[0m\n\nTASK [k3s_server_post : Check Cilium version] **********************************\nTuesday 24 February 2026  00:51:01 +0000 (0:00:00.997)       0:03:46.535 ****** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [k3s_server_post : Parse installed Cilium version] ************************\nTuesday 24 February 2026  00:51:01 +0000 (0:00:00.124)       0:03:46.660 ****** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [k3s_server_post : Determine if Cilium needs update] **********************\nTuesday 24 February 2026  00:51:01 +0000 (0:00:00.114)       0:03:46.774 ****** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [k3s_server_post : Log result] ********************************************\nTuesday 24 February 2026  00:51:01 +0000 (0:00:00.121)       0:03:46.896 ****** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [k3s_server_post : Install Cilium] ****************************************\nTuesday 24 February 2026  00:51:01 +0000 (0:00:00.115)       0:03:47.011 ****** \n\u001b[0;33mchanged: [testbed-node-0 -> localhost]\u001b[0m\n\nTASK [k3s_server_post : Wait for Cilium resources] *****************************\nTuesday 24 February 2026  00:51:06 +0000 (0:00:05.132)       0:03:52.144 ****** \n\u001b[0;32mok: [testbed-node-0 -> localhost] => (item=deployment/cilium-operator)\u001b[0m\n\u001b[1;30mFAILED - RETRYING: [testbed-node-0 -> localhost]: Wait for Cilium resources (30 retries left).\u001b[0m\n\u001b[0;32mok: [testbed-node-0 -> localhost] => (item=daemonset/cilium)\u001b[0m\n\u001b[0;32mok: [testbed-node-0 -> localhost] => (item=deployment/hubble-relay)\u001b[0m\n\u001b[0;32mok: [testbed-node-0 -> localhost] => (item=deployment/hubble-ui)\u001b[0m\n\nTASK [k3s_server_post : Set _cilium_bgp_neighbors fact] ************************\nTuesday 24 February 2026  00:51:48 +0000 (0:00:42.155)       0:04:34.300 ****** \n\u001b[0;32mok: [testbed-node-0 -> localhost]\u001b[0m\n\nTASK [k3s_server_post : Copy BGP manifests to first master] ********************\nTuesday 24 February 2026  00:51:50 +0000 (0:00:01.220)       0:04:35.520 ****** \n\u001b[0;33mchanged: [testbed-node-0 -> localhost]\u001b[0m\n\nTASK [k3s_server_post : Apply BGP manifests] ***********************************\nTuesday 24 February 2026  00:51:51 +0000 (0:00:01.540)       0:04:37.060 ****** \n\u001b[0;33mchanged: [testbed-node-0 -> localhost]\u001b[0m\n\nTASK [k3s_server_post : Print error message if BGP manifests application fails] ***\nTuesday 24 February 2026  00:51:52 +0000 (0:00:00.842)       0:04:37.903 ****** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [k3s_server_post : Test for BGP config resources] *************************\nTuesday 24 February 2026  00:51:52 +0000 (0:00:00.104)       0:04:38.008 ****** \n\u001b[0;32mok: [testbed-node-0 -> localhost] => (item=kubectl get CiliumBGPPeeringPolicy.cilium.io)\u001b[0m\n\u001b[0;32mok: [testbed-node-0 -> localhost] => (item=kubectl get CiliumLoadBalancerIPPool.cilium.io)\u001b[0m\n\nTASK [k3s_server_post : Deploy metallb pool] ***********************************\nTuesday 24 February 2026  00:51:54 +0000 (0:00:01.645)       0:04:39.654 ****** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [k3s_server_post : Remove tmp directory used for manifests] ***************\nTuesday 24 February 2026  00:51:54 +0000 (0:00:00.258)       0:04:39.913 ****** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nPLAY [Apply role k9s] **********************************************************\n\nTASK [k9s : Gather variables for each operating system] ************************\nTuesday 24 February 2026  00:51:55 +0000 (0:00:00.974)       0:04:40.887 ****** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\nTASK [k9s : Include distribution specific install tasks] ***********************\nTuesday 24 February 2026  00:51:55 +0000 (0:00:00.133)       0:04:41.021 ****** \n\u001b[0;36mincluded: /ansible/roles/k9s/tasks/install-Debian-family.yml for testbed-manager\u001b[0m\n\nTASK [k9s : Install k9s packages] **********************************************\nTuesday 24 February 2026  00:51:55 +0000 (0:00:00.192)       0:04:41.214 ****** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nPLAY [Manage labels, annotations, and taints on all k3s nodes] *****************\n\nTASK [Merge labels, annotations, and taints] ***********************************\nTuesday 24 February 2026  00:52:01 +0000 (0:00:05.433)       0:04:46.648 ****** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [Manage labels] ***********************************************************\nTuesday 24 February 2026  00:52:01 +0000 (0:00:00.673)       0:04:47.321 ****** \n\u001b[0;32mok: [testbed-node-3 -> localhost] => (item=node-role.osism.tech/compute-plane=true)\u001b[0m\n\u001b[0;32mok: [testbed-node-4 -> localhost] => (item=node-role.osism.tech/compute-plane=true)\u001b[0m\n\u001b[0;32mok: [testbed-node-5 -> localhost] => (item=node-role.osism.tech/compute-plane=true)\u001b[0m\n\u001b[0;32mok: [testbed-node-0 -> localhost] => (item=node-role.osism.tech/control-plane=true)\u001b[0m\n\u001b[0;32mok: [testbed-node-1 -> localhost] => (item=node-role.osism.tech/control-plane=true)\u001b[0m\n\u001b[0;32mok: [testbed-node-5 -> localhost] => (item=node-role.kubernetes.io/worker=worker)\u001b[0m\n\u001b[0;32mok: [testbed-node-2 -> localhost] => (item=node-role.osism.tech/control-plane=true)\u001b[0m\n\u001b[0;32mok: [testbed-node-3 -> localhost] => (item=node-role.kubernetes.io/worker=worker)\u001b[0m\n\u001b[0;32mok: [testbed-node-4 -> localhost] => (item=node-role.kubernetes.io/worker=worker)\u001b[0m\n\u001b[0;32mok: [testbed-node-0 -> localhost] => (item=openstack-control-plane=enabled)\u001b[0m\n\u001b[0;32mok: [testbed-node-1 -> localhost] => (item=openstack-control-plane=enabled)\u001b[0m\n\u001b[0;32mok: [testbed-node-5 -> localhost] => (item=node-role.osism.tech/rook-osd=true)\u001b[0m\n\u001b[0;32mok: [testbed-node-2 -> localhost] => (item=openstack-control-plane=enabled)\u001b[0m\n\u001b[0;32mok: [testbed-node-4 -> localhost] => (item=node-role.osism.tech/rook-osd=true)\u001b[0m\n\u001b[0;32mok: [testbed-node-1 -> localhost] => (item=node-role.osism.tech/network-plane=true)\u001b[0m\n\u001b[0;32mok: [testbed-node-3 -> localhost] => (item=node-role.osism.tech/rook-osd=true)\u001b[0m\n\u001b[0;32mok: [testbed-node-0 -> localhost] => (item=node-role.osism.tech/network-plane=true)\u001b[0m\n\u001b[0;32mok: [testbed-node-2 -> localhost] => (item=node-role.osism.tech/network-plane=true)\u001b[0m\n\u001b[0;32mok: [testbed-node-1 -> localhost] => (item=node-role.osism.tech/rook-mds=true)\u001b[0m\n\u001b[0;32mok: [testbed-node-0 -> localhost] => (item=node-role.osism.tech/rook-mds=true)\u001b[0m\n\u001b[0;32mok: [testbed-node-1 -> localhost] => (item=node-role.osism.tech/rook-mgr=true)\u001b[0m\n\u001b[0;32mok: [testbed-node-2 -> localhost] => (item=node-role.osism.tech/rook-mds=true)\u001b[0m\n\u001b[0;32mok: [testbed-node-2 -> localhost] => (item=node-role.osism.tech/rook-mgr=true)\u001b[0m\n\u001b[0;32mok: [testbed-node-1 -> localhost] => (item=node-role.osism.tech/rook-mon=true)\u001b[0m\n\u001b[0;32mok: [testbed-node-0 -> localhost] => (item=node-role.osism.tech/rook-mgr=true)\u001b[0m\n\u001b[0;32mok: [testbed-node-2 -> localhost] => (item=node-role.osism.tech/rook-mon=true)\u001b[0m\n\u001b[0;32mok: [testbed-node-1 -> localhost] => (item=node-role.osism.tech/rook-rgw=true)\u001b[0m\n\u001b[0;32mok: [testbed-node-0 -> localhost] => (item=node-role.osism.tech/rook-mon=true)\u001b[0m\n\u001b[0;32mok: [testbed-node-0 -> localhost] => (item=node-role.osism.tech/rook-rgw=true)\u001b[0m\n\u001b[0;32mok: [testbed-node-2 -> localhost] => (item=node-role.osism.tech/rook-rgw=true)\u001b[0m\n\nTASK [Manage annotations] ******************************************************\nTuesday 24 February 2026  00:52:15 +0000 (0:00:13.202)       0:05:00.523 ****** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [Manage taints] ***********************************************************\nTuesday 24 February 2026  00:52:15 +0000 (0:00:00.533)       0:05:01.057 ****** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;33mtestbed-manager\u001b[0m            : \u001b[0;32mok=21  \u001b[0m \u001b[0;33mchanged=11  \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=50  \u001b[0m \u001b[0;33mchanged=23  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=28  \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=38  \u001b[0m \u001b[0;33mchanged=16  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=25  \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=38  \u001b[0m \u001b[0;33mchanged=16  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=25  \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-3\u001b[0m             : \u001b[0;32mok=16  \u001b[0m \u001b[0;33mchanged=8   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=17  \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-4\u001b[0m             : \u001b[0;32mok=16  \u001b[0m \u001b[0;33mchanged=8   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=17  \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-5\u001b[0m             : \u001b[0;32mok=16  \u001b[0m \u001b[0;33mchanged=8   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=17  \u001b[0m rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nTuesday 24 February 2026  00:52:15 +0000 (0:00:00.386)       0:05:01.443 ****** \n=============================================================================== \nk3s_server : Verify that all nodes actually joined (check k3s-init.service if this fails) -- 54.02s\nk3s_server : Enable and check K3s service ------------------------------ 53.12s\nk3s_server_post : Wait for Cilium resources ---------------------------- 42.16s\nManage labels ---------------------------------------------------------- 13.20s\nkubectl : Install required packages ------------------------------------ 12.52s\nk3s_agent : Manage k3s service ----------------------------------------- 10.20s\nkubectl : Add repository Debian ----------------------------------------- 6.68s\nk3s_download : Download k3s binary x64 ---------------------------------- 5.71s\nk9s : Install k9s packages ---------------------------------------------- 5.43s\nk3s_server_post : Install Cilium ---------------------------------------- 5.13s\nk3s_server : Remove manifests and folders that are only needed for bootstrapping cluster so k3s doesn't auto apply on start --- 3.03s\nk3s_server : Set _kube_vip_bgp_peers fact ------------------------------- 2.83s\nk3s_server : Detect Kubernetes version for label compatibility ---------- 2.80s\nk3s_server : Init cluster inside the transient k3s-init service --------- 2.33s\nk3s_prereq : Enable IPv4 forwarding ------------------------------------- 2.16s\nk3s_download : Download k3s binary armhf -------------------------------- 2.09s\nk3s_server : Create custom resolv.conf for k3s -------------------------- 1.96s\nk3s_prereq : Enable IPv6 router advertisements -------------------------- 1.95s\nk3s_agent : Configure the k3s service ----------------------------------- 1.85s\nk3s_download : Download k3s binary arm64 -------------------------------- 1.83s\n\u001b[32m2026-02-24 00:52:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask feb6adf1-2c00-4649-9a4f-39e67ac78a53 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:52:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ba7d4726-e54a-4b5b-b060-92175ab177a2 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:52:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6bfc963b-db63-4b91-bc2e-bbd0ebbc84e3 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:52:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:52:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 404270fa-50db-4a78-8c44-21df48acba39 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:52:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 00:52:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask feb6adf1-2c00-4649-9a4f-39e67ac78a53 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:52:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ba7d4726-e54a-4b5b-b060-92175ab177a2 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:52:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6bfc963b-db63-4b91-bc2e-bbd0ebbc84e3 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:52:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:52:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 404270fa-50db-4a78-8c44-21df48acba39 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:52:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 00:52:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask feb6adf1-2c00-4649-9a4f-39e67ac78a53 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:52:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ba7d4726-e54a-4b5b-b060-92175ab177a2 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:52:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6bfc963b-db63-4b91-bc2e-bbd0ebbc84e3 is in state SUCCESS\u001b[0m\n\u001b[32m2026-02-24 00:52:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:52:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 404270fa-50db-4a78-8c44-21df48acba39 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:52:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 00:52:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask feb6adf1-2c00-4649-9a4f-39e67ac78a53 is in state SUCCESS\u001b[0m\n\u001b[32m2026-02-24 00:52:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ba7d4726-e54a-4b5b-b060-92175ab177a2 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:52:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:52:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 404270fa-50db-4a78-8c44-21df48acba39 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:52:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 00:52:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ba7d4726-e54a-4b5b-b060-92175ab177a2 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:52:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:52:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 404270fa-50db-4a78-8c44-21df48acba39 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:52:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 00:52:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ba7d4726-e54a-4b5b-b060-92175ab177a2 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:52:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:52:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 404270fa-50db-4a78-8c44-21df48acba39 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:52:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 00:52:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ba7d4726-e54a-4b5b-b060-92175ab177a2 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:52:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:52:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 404270fa-50db-4a78-8c44-21df48acba39 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:52:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 00:52:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ba7d4726-e54a-4b5b-b060-92175ab177a2 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:52:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:52:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 404270fa-50db-4a78-8c44-21df48acba39 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:52:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 00:52:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ba7d4726-e54a-4b5b-b060-92175ab177a2 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:52:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:52:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 404270fa-50db-4a78-8c44-21df48acba39 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:52:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 00:52:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ba7d4726-e54a-4b5b-b060-92175ab177a2 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:52:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:52:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 404270fa-50db-4a78-8c44-21df48acba39 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:52:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 00:52:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ba7d4726-e54a-4b5b-b060-92175ab177a2 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:52:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:52:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 404270fa-50db-4a78-8c44-21df48acba39 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:52:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 00:52:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ba7d4726-e54a-4b5b-b060-92175ab177a2 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:52:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:52:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 404270fa-50db-4a78-8c44-21df48acba39 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:52:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 00:52:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ba7d4726-e54a-4b5b-b060-92175ab177a2 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:52:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:52:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 404270fa-50db-4a78-8c44-21df48acba39 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:52:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 00:52:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ba7d4726-e54a-4b5b-b060-92175ab177a2 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:52:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:52:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 404270fa-50db-4a78-8c44-21df48acba39 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:52:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 00:53:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ba7d4726-e54a-4b5b-b060-92175ab177a2 is in state SUCCESS\u001b[0m\n\n\nPLAY [Copy kubeconfig to the configuration repository] *************************\n\nTASK [Get kubeconfig file] *****************************************************\nTuesday 24 February 2026  00:52:20 +0000 (0:00:00.142)       0:00:00.142 ****** \n\u001b[0;32mok: [testbed-manager -> testbed-node-0(192.168.16.10)]\u001b[0m\n\nTASK [Write kubeconfig file] ***************************************************\nTuesday 24 February 2026  00:52:21 +0000 (0:00:00.665)       0:00:00.808 ****** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nTASK [Change server address in the kubeconfig file] ****************************\nTuesday 24 February 2026  00:52:22 +0000 (0:00:00.956)       0:00:01.764 ****** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;33mtestbed-manager\u001b[0m            : \u001b[0;32mok=3   \u001b[0m \u001b[0;33mchanged=2   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nTuesday 24 February 2026  00:52:22 +0000 (0:00:00.368)       0:00:02.133 ****** \n=============================================================================== \nWrite kubeconfig file --------------------------------------------------- 0.96s\nGet kubeconfig file ----------------------------------------------------- 0.67s\nChange server address in the kubeconfig file ---------------------------- 0.37s\n\n\nPLAY [Prepare kubeconfig file] *************************************************\n\nTASK [Get home directory of operator user] *************************************\nTuesday 24 February 2026  00:52:19 +0000 (0:00:00.119)       0:00:00.119 ****** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\nTASK [Create .kube directory] **************************************************\nTuesday 24 February 2026  00:52:20 +0000 (0:00:00.474)       0:00:00.594 ****** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\nTASK [Get kubeconfig file] *****************************************************\nTuesday 24 February 2026  00:52:20 +0000 (0:00:00.530)       0:00:01.125 ****** \n\u001b[0;32mok: [testbed-manager -> testbed-node-0(192.168.16.10)]\u001b[0m\n\nTASK [Write kubeconfig file] ***************************************************\nTuesday 24 February 2026  00:52:21 +0000 (0:00:00.646)       0:00:01.772 ****** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nTASK [Change server address in the kubeconfig] *********************************\nTuesday 24 February 2026  00:52:22 +0000 (0:00:01.278)       0:00:03.050 ****** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nTASK [Make kubeconfig available for use inside the manager service] ************\nTuesday 24 February 2026  00:52:23 +0000 (0:00:00.454)       0:00:03.505 ****** \n\u001b[0;33mchanged: [testbed-manager -> localhost]\u001b[0m\n\nTASK [Change server address in the kubeconfig inside the manager service] ******\nTuesday 24 February 2026  00:52:24 +0000 (0:00:01.527)       0:00:05.032 ****** \n\u001b[0;33mchanged: [testbed-manager -> localhost]\u001b[0m\n\nTASK [Set KUBECONFIG environment variable] *************************************\nTuesday 24 February 2026  00:52:25 +0000 (0:00:00.782)       0:00:05.815 ****** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\nTASK [Enable kubectl command line completion] **********************************\nTuesday 24 February 2026  00:52:25 +0000 (0:00:00.359)       0:00:06.174 ****** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;33mtestbed-manager\u001b[0m            : \u001b[0;32mok=9   \u001b[0m \u001b[0;33mchanged=4   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nTuesday 24 February 2026  00:52:26 +0000 (0:00:00.340)       0:00:06.515 ****** \n=============================================================================== \nMake kubeconfig available for use inside the manager service ------------ 1.53s\nWrite kubeconfig file --------------------------------------------------- 1.28s\nChange server address in the kubeconfig inside the manager service ------ 0.78s\nGet kubeconfig file ----------------------------------------------------- 0.65s\nCreate .kube directory -------------------------------------------------- 0.53s\nGet home directory of operator user ------------------------------------- 0.47s\nChange server address in the kubeconfig --------------------------------- 0.45s\nSet KUBECONFIG environment variable ------------------------------------- 0.36s\nEnable kubectl command line completion ---------------------------------- 0.34s\n\n\nPLAY [Group hosts based on configuration] **************************************\n\nTASK [Group hosts based on Kolla action] ***************************************\nTuesday 24 February 2026  00:50:43 +0000 (0:00:00.140)       0:00:00.140 ****** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [Group hosts based on enabled services] ***********************************\nTuesday 24 February 2026  00:50:43 +0000 (0:00:00.849)       0:00:00.989 ****** \n\u001b[0;32mok: [testbed-node-3] => (item=enable_ovn_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item=enable_ovn_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item=enable_ovn_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-0] => (item=enable_ovn_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item=enable_ovn_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item=enable_ovn_True)\u001b[0m\n\nPLAY [Apply role ovn-controller] ***********************************************\n\nTASK [ovn-controller : include_tasks] ******************************************\nTuesday 24 February 2026  00:50:44 +0000 (0:00:01.006)       0:00:01.995 ****** \n\u001b[0;36mincluded: /ansible/roles/ovn-controller/tasks/deploy.yml for testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [ovn-controller : Ensuring config directories exist] **********************\nTuesday 24 February 2026  00:50:46 +0000 (0:00:01.030)       0:00:03.026 ****** \n\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-controller:2024.2', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-controller:2024.2', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-controller:2024.2', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-controller:2024.2', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-controller:2024.2', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-controller:2024.2', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\nTASK [ovn-controller : Copying over config.json files for services] ************\nTuesday 24 February 2026  00:50:47 +0000 (0:00:01.504)       0:00:04.530 ****** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-controller:2024.2', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-controller:2024.2', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-controller:2024.2', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-controller:2024.2', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-controller:2024.2', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-controller:2024.2', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\nTASK [ovn-controller : Ensuring systemd override directory exists] *************\nTuesday 24 February 2026  00:50:49 +0000 (0:00:01.684)       0:00:06.215 ****** \n\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-controller:2024.2', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-controller:2024.2', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-controller:2024.2', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-controller:2024.2', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-controller:2024.2', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-controller:2024.2', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\nTASK [ovn-controller : Copying over systemd override] **************************\nTuesday 24 February 2026  00:50:50 +0000 (0:00:01.372)       0:00:07.587 ****** \n\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-controller:2024.2', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-controller:2024.2', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-controller:2024.2', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-controller:2024.2', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-controller:2024.2', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-controller:2024.2', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\nTASK [ovn-controller : Check ovn-controller containers] ************************\nTuesday 24 February 2026  00:50:52 +0000 (0:00:01.445)       0:00:09.033 ****** \n\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-controller:2024.2', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-controller:2024.2', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-controller:2024.2', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-controller:2024.2', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-controller:2024.2', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-controller:2024.2', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\nTASK [ovn-controller : Create br-int bridge on OpenvSwitch] ********************\nTuesday 24 February 2026  00:50:53 +0000 (0:00:01.520)       0:00:10.553 ****** \n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [ovn-controller : Configure OVN in OVSDB] *********************************\nTuesday 24 February 2026  00:50:56 +0000 (0:00:02.732)       0:00:13.285 ****** \n\u001b[0;33mchanged: [testbed-node-3] => (item={'name': 'ovn-encap-ip', 'value': '192.168.16.13'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'name': 'ovn-encap-ip', 'value': '192.168.16.14'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'name': 'ovn-encap-ip', 'value': '192.168.16.12'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'name': 'ovn-encap-ip', 'value': '192.168.16.15'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'name': 'ovn-encap-ip', 'value': '192.168.16.11'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'name': 'ovn-encap-ip', 'value': '192.168.16.10'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'name': 'ovn-encap-type', 'value': 'geneve'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'name': 'ovn-encap-type', 'value': 'geneve'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'name': 'ovn-encap-type', 'value': 'geneve'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'name': 'ovn-encap-type', 'value': 'geneve'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'name': 'ovn-encap-type', 'value': 'geneve'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'name': 'ovn-encap-type', 'value': 'geneve'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'name': 'ovn-remote', 'value': 'tcp:192.168.16.10:6642,tcp:192.168.16.11:6642,tcp:192.168.16.12:6642'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'name': 'ovn-remote', 'value': 'tcp:192.168.16.10:6642,tcp:192.168.16.11:6642,tcp:192.168.16.12:6642'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'name': 'ovn-remote', 'value': 'tcp:192.168.16.10:6642,tcp:192.168.16.11:6642,tcp:192.168.16.12:6642'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'name': 'ovn-remote', 'value': 'tcp:192.168.16.10:6642,tcp:192.168.16.11:6642,tcp:192.168.16.12:6642'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'name': 'ovn-remote', 'value': 'tcp:192.168.16.10:6642,tcp:192.168.16.11:6642,tcp:192.168.16.12:6642'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'name': 'ovn-remote', 'value': 'tcp:192.168.16.10:6642,tcp:192.168.16.11:6642,tcp:192.168.16.12:6642'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'name': 'ovn-remote-probe-interval', 'value': '60000'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'name': 'ovn-remote-probe-interval', 'value': '60000'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'name': 'ovn-remote-probe-interval', 'value': '60000'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'name': 'ovn-remote-probe-interval', 'value': '60000'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'name': 'ovn-remote-probe-interval', 'value': '60000'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'name': 'ovn-remote-probe-interval', 'value': '60000'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'name': 'ovn-openflow-probe-interval', 'value': '60'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'name': 'ovn-openflow-probe-interval', 'value': '60'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'name': 'ovn-openflow-probe-interval', 'value': '60'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'name': 'ovn-openflow-probe-interval', 'value': '60'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'name': 'ovn-openflow-probe-interval', 'value': '60'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'name': 'ovn-openflow-probe-interval', 'value': '60'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'name': 'ovn-monitor-all', 'value': False})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'name': 'ovn-monitor-all', 'value': False})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'name': 'ovn-monitor-all', 'value': False})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'name': 'ovn-monitor-all', 'value': False})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'name': 'ovn-monitor-all', 'value': False})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'name': 'ovn-monitor-all', 'value': False})\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item={'name': 'ovn-bridge-mappings', 'value': 'physnet1:br-ex', 'state': 'absent'})\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item={'name': 'ovn-bridge-mappings', 'value': 'physnet1:br-ex', 'state': 'absent'})\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item={'name': 'ovn-bridge-mappings', 'value': 'physnet1:br-ex', 'state': 'absent'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'name': 'ovn-bridge-mappings', 'value': 'physnet1:br-ex', 'state': 'present'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'name': 'ovn-bridge-mappings', 'value': 'physnet1:br-ex', 'state': 'present'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'name': 'ovn-bridge-mappings', 'value': 'physnet1:br-ex', 'state': 'present'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'name': 'ovn-chassis-mac-mappings', 'value': 'physnet1:52:54:00:2f:fa:44', 'state': 'present'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'name': 'ovn-chassis-mac-mappings', 'value': 'physnet1:52:54:00:71:3a:c3', 'state': 'present'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'name': 'ovn-chassis-mac-mappings', 'value': 'physnet1:52:54:00:89:18:56', 'state': 'present'})\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item={'name': 'ovn-chassis-mac-mappings', 'value': 'physnet1:52:54:00:29:4a:9b', 'state': 'absent'})\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item={'name': 'ovn-chassis-mac-mappings', 'value': 'physnet1:52:54:00:33:12:50', 'state': 'absent'})\u001b[0m\n\u001b[0;32mok: [testbed-node-0] => (item={'name': 'ovn-chassis-mac-mappings', 'value': 'physnet1:52:54:00:52:c1:40', 'state': 'absent'})\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item={'name': 'ovn-cms-options', 'value': '', 'state': 'absent'})\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item={'name': 'ovn-cms-options', 'value': '', 'state': 'absent'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'name': 'ovn-cms-options', 'value': 'enable-chassis-as-gw,availability-zones=nova', 'state': 'present'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'name': 'ovn-cms-options', 'value': 'enable-chassis-as-gw,availability-zones=nova', 'state': 'present'})\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item={'name': 'ovn-cms-options', 'value': '', 'state': 'absent'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'name': 'ovn-cms-options', 'value': 'enable-chassis-as-gw,availability-zones=nova', 'state': 'present'})\u001b[0m\n\nTASK [ovn-controller : Flush handlers] *****************************************\nTuesday 24 February 2026  00:51:18 +0000 (0:00:22.571)       0:00:35.857 ****** \n\nTASK [ovn-controller : Flush handlers] *****************************************\nTuesday 24 February 2026  00:51:18 +0000 (0:00:00.063)       0:00:35.921 ****** \n\nTASK [ovn-controller : Flush handlers] *****************************************\nTuesday 24 February 2026  00:51:18 +0000 (0:00:00.063)       0:00:35.984 ****** \n\nTASK [ovn-controller : Flush handlers] *****************************************\nTuesday 24 February 2026  00:51:19 +0000 (0:00:00.060)       0:00:36.045 ****** \n\nTASK [ovn-controller : Flush handlers] *****************************************\nTuesday 24 February 2026  00:51:19 +0000 (0:00:00.065)       0:00:36.111 ****** \n\nTASK [ovn-controller : Flush handlers] *****************************************\nTuesday 24 February 2026  00:51:19 +0000 (0:00:00.062)       0:00:36.173 ****** \n\nRUNNING HANDLER [ovn-controller : Reload systemd config] ***********************\nTuesday 24 February 2026  00:51:19 +0000 (0:00:00.062)       0:00:36.236 ****** \n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [ovn-controller : Restart ovn-controller container] ************\nTuesday 24 February 2026  00:51:20 +0000 (0:00:01.561)       0:00:37.798 ****** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nPLAY [Apply role ovn-db] *******************************************************\n\nTASK [ovn-db : include_tasks] **************************************************\nTuesday 24 February 2026  00:51:46 +0000 (0:00:25.870)       0:01:03.668 ****** \n\u001b[0;36mincluded: /ansible/roles/ovn-db/tasks/deploy.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [ovn-db : include_tasks] **************************************************\nTuesday 24 February 2026  00:51:47 +0000 (0:00:00.583)       0:01:04.251 ****** \n\u001b[0;36mincluded: /ansible/roles/ovn-db/tasks/lookup_cluster.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [ovn-db : Checking for any existing OVN DB container volumes] *************\nTuesday 24 February 2026  00:51:47 +0000 (0:00:00.696)       0:01:04.947 ****** \n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ovn-db : Divide hosts by their OVN NB volume availability] ***************\nTuesday 24 February 2026  00:51:48 +0000 (0:00:00.975)       0:01:05.923 ****** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ovn-db : Divide hosts by their OVN SB volume availability] ***************\nTuesday 24 February 2026  00:51:49 +0000 (0:00:00.462)       0:01:06.385 ****** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ovn-db : Establish whether the OVN NB cluster has already existed] *******\nTuesday 24 February 2026  00:51:49 +0000 (0:00:00.419)       0:01:06.805 ****** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ovn-db : Establish whether the OVN SB cluster has already existed] *******\nTuesday 24 February 2026  00:51:50 +0000 (0:00:00.597)       0:01:07.403 ****** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ovn-db : Check if running on all OVN NB DB hosts] ************************\nTuesday 24 February 2026  00:51:51 +0000 (0:00:00.667)       0:01:08.070 ****** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ovn-db : Check OVN NB service port liveness] *****************************\nTuesday 24 February 2026  00:51:51 +0000 (0:00:00.298)       0:01:08.369 ****** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ovn-db : Divide hosts by their OVN NB service port liveness] *************\nTuesday 24 February 2026  00:51:51 +0000 (0:00:00.270)       0:01:08.640 ****** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ovn-db : Get OVN NB database information] ********************************\nTuesday 24 February 2026  00:51:51 +0000 (0:00:00.259)       0:01:08.899 ****** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ovn-db : Divide hosts by their OVN NB leader/follower role] **************\nTuesday 24 February 2026  00:51:52 +0000 (0:00:00.417)       0:01:09.317 ****** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ovn-db : Fail on existing OVN NB cluster with no leader] *****************\nTuesday 24 February 2026  00:51:52 +0000 (0:00:00.300)       0:01:09.617 ****** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ovn-db : Check if running on all OVN SB DB hosts] ************************\nTuesday 24 February 2026  00:51:52 +0000 (0:00:00.262)       0:01:09.879 ****** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ovn-db : Check OVN SB service port liveness] *****************************\nTuesday 24 February 2026  00:51:53 +0000 (0:00:00.306)       0:01:10.185 ****** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ovn-db : Divide hosts by their OVN SB service port liveness] *************\nTuesday 24 February 2026  00:51:53 +0000 (0:00:00.442)       0:01:10.628 ****** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ovn-db : Get OVN SB database information] ********************************\nTuesday 24 February 2026  00:51:53 +0000 (0:00:00.244)       0:01:10.872 ****** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ovn-db : Divide hosts by their OVN SB leader/follower role] **************\nTuesday 24 February 2026  00:51:54 +0000 (0:00:00.269)       0:01:11.142 ****** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ovn-db : Fail on existing OVN SB cluster with no leader] *****************\nTuesday 24 February 2026  00:51:54 +0000 (0:00:00.269)       0:01:11.411 ****** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ovn-db : include_tasks] **************************************************\nTuesday 24 February 2026  00:51:54 +0000 (0:00:00.260)       0:01:11.672 ****** \n\u001b[0;36mincluded: /ansible/roles/ovn-db/tasks/bootstrap-initial.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [ovn-db : Set bootstrap args fact for NB (new cluster)] *******************\nTuesday 24 February 2026  00:51:55 +0000 (0:00:00.688)       0:01:12.360 ****** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ovn-db : Set bootstrap args fact for SB (new cluster)] *******************\nTuesday 24 February 2026  00:51:55 +0000 (0:00:00.382)       0:01:12.743 ****** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ovn-db : Check NB cluster status] ****************************************\nTuesday 24 February 2026  00:51:56 +0000 (0:00:00.361)       0:01:13.104 ****** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ovn-db : Check SB cluster status] ****************************************\nTuesday 24 February 2026  00:51:56 +0000 (0:00:00.467)       0:01:13.571 ****** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ovn-db : Remove an old node with the same ip address as the new node in NB DB] ***\nTuesday 24 February 2026  00:51:56 +0000 (0:00:00.340)       0:01:13.911 ****** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ovn-db : Remove an old node with the same ip address as the new node in SB DB] ***\nTuesday 24 February 2026  00:51:57 +0000 (0:00:00.339)       0:01:14.250 ****** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ovn-db : Set bootstrap args fact for NB (new member)] ********************\nTuesday 24 February 2026  00:51:57 +0000 (0:00:00.275)       0:01:14.526 ****** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ovn-db : Set bootstrap args fact for SB (new member)] ********************\nTuesday 24 February 2026  00:51:57 +0000 (0:00:00.463)       0:01:14.989 ****** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ovn-db : Ensuring config directories exist] ******************************\nTuesday 24 February 2026  00:51:58 +0000 (0:00:00.346)       0:01:15.336 ****** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ovn-northd', 'value': {'container_name': 'ovn_northd', 'group': 'ovn-northd', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-northd:2024.2', 'volumes': ['/etc/kolla/ovn-northd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ovn-northd', 'value': {'container_name': 'ovn_northd', 'group': 'ovn-northd', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-northd:2024.2', 'volumes': ['/etc/kolla/ovn-northd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ovn-northd', 'value': {'container_name': 'ovn_northd', 'group': 'ovn-northd', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-northd:2024.2', 'volumes': ['/etc/kolla/ovn-northd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ovn-nb-db', 'value': {'container_name': 'ovn_nb_db', 'group': 'ovn-nb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-nb-db-server:2024.2', 'volumes': ['/etc/kolla/ovn-nb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_nb_db:/var/lib/openvswitch/ovn-nb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ovn-nb-db', 'value': {'container_name': 'ovn_nb_db', 'group': 'ovn-nb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-nb-db-server:2024.2', 'volumes': ['/etc/kolla/ovn-nb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_nb_db:/var/lib/openvswitch/ovn-nb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ovn-nb-db', 'value': {'container_name': 'ovn_nb_db', 'group': 'ovn-nb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-nb-db-server:2024.2', 'volumes': ['/etc/kolla/ovn-nb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_nb_db:/var/lib/openvswitch/ovn-nb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ovn-sb-db', 'value': {'container_name': 'ovn_sb_db', 'group': 'ovn-sb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-sb-db-server:2024.2', 'volumes': ['/etc/kolla/ovn-sb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_sb_db:/var/lib/openvswitch/ovn-sb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ovn-sb-db', 'value': {'container_name': 'ovn_sb_db', 'group': 'ovn-sb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-sb-db-server:2024.2', 'volumes': ['/etc/kolla/ovn-sb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_sb_db:/var/lib/openvswitch/ovn-sb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ovn-sb-db', 'value': {'container_name': 'ovn_sb_db', 'group': 'ovn-sb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-sb-db-server:2024.2', 'volumes': ['/etc/kolla/ovn-sb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_sb_db:/var/lib/openvswitch/ovn-sb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\nTASK [ovn-db : Copying over config.json files for services] ********************\nTuesday 24 February 2026  00:52:00 +0000 (0:00:01.830)       0:01:17.167 ****** \n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ovn-northd', 'value': {'container_name': 'ovn_northd', 'group': 'ovn-northd', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-northd:2024.2', 'volumes': ['/etc/kolla/ovn-northd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ovn-northd', 'value': {'container_name': 'ovn_northd', 'group': 'ovn-northd', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-northd:2024.2', 'volumes': ['/etc/kolla/ovn-northd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ovn-northd', 'value': {'container_name': 'ovn_northd', 'group': 'ovn-northd', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-northd:2024.2', 'volumes': ['/etc/kolla/ovn-northd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ovn-nb-db', 'value': {'container_name': 'ovn_nb_db', 'group': 'ovn-nb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-nb-db-server:2024.2', 'volumes': ['/etc/kolla/ovn-nb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_nb_db:/var/lib/openvswitch/ovn-nb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ovn-nb-db', 'value': {'container_name': 'ovn_nb_db', 'group': 'ovn-nb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-nb-db-server:2024.2', 'volumes': ['/etc/kolla/ovn-nb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_nb_db:/var/lib/openvswitch/ovn-nb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ovn-nb-db', 'value': {'container_name': 'ovn_nb_db', 'group': 'ovn-nb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-nb-db-server:2024.2', 'volumes': ['/etc/kolla/ovn-nb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_nb_db:/var/lib/openvswitch/ovn-nb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ovn-sb-db', 'value': {'container_name': 'ovn_sb_db', 'group': 'ovn-sb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-sb-db-server:2024.2', 'volumes': ['/etc/kolla/ovn-sb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_sb_db:/var/lib/openvswitch/ovn-sb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ovn-sb-db', 'value': {'container_name': 'ovn_sb_db', 'group': 'ovn-sb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-sb-db-server:2024.2', 'volumes': ['/etc/kolla/ovn-sb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_sb_db:/var/lib/openvswitch/ovn-sb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ovn-sb-db', 'value': {'container_name': 'ovn_sb_db', 'group': 'ovn-sb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-sb-db-server:2024.2', 'volumes': ['/etc/kolla/ovn-sb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_sb_db:/var/lib/openvswitch/ovn-sb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\nTASK [ovn-db : Check ovn containers] *******************************************\nTuesday 24 February 2026  00:52:05 +0000 (0:00:05.458)       0:01:22.625 ****** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ovn-northd', 'value': {'container_name': 'ovn_northd', 'group': 'ovn-northd', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-northd:2024.2', 'volumes': ['/etc/kolla/ovn-northd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ovn-northd', 'value': {'container_name': 'ovn_northd', 'group': 'ovn-northd', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-northd:2024.2', 'volumes': ['/etc/kolla/ovn-northd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ovn-northd', 'value': {'container_name': 'ovn_northd', 'group': 'ovn-northd', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-northd:2024.2', 'volumes': ['/etc/kolla/ovn-northd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ovn-nb-db', 'value': {'container_name': 'ovn_nb_db', 'group': 'ovn-nb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-nb-db-server:2024.2', 'volumes': ['/etc/kolla/ovn-nb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_nb_db:/var/lib/openvswitch/ovn-nb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ovn-nb-db', 'value': {'container_name': 'ovn_nb_db', 'group': 'ovn-nb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-nb-db-server:2024.2', 'volumes': ['/etc/kolla/ovn-nb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_nb_db:/var/lib/openvswitch/ovn-nb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ovn-nb-db', 'value': {'container_name': 'ovn_nb_db', 'group': 'ovn-nb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-nb-db-server:2024.2', 'volumes': ['/etc/kolla/ovn-nb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_nb_db:/var/lib/openvswitch/ovn-nb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ovn-sb-db', 'value': {'container_name': 'ovn_sb_db', 'group': 'ovn-sb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-sb-db-server:2024.2', 'volumes': ['/etc/kolla/ovn-sb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_sb_db:/var/lib/openvswitch/ovn-sb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ovn-sb-db', 'value': {'container_name': 'ovn_sb_db', 'group': 'ovn-sb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-sb-db-server:2024.2', 'volumes': ['/etc/kolla/ovn-sb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_sb_db:/var/lib/openvswitch/ovn-sb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ovn-sb-db', 'value': {'container_name': 'ovn_sb_db', 'group': 'ovn-sb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-sb-db-server:2024.2', 'volumes': ['/etc/kolla/ovn-sb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_sb_db:/var/lib/openvswitch/ovn-sb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\nTASK [ovn-db : Flush handlers] *************************************************\nTuesday 24 February 2026  00:52:08 +0000 (0:00:02.544)       0:01:25.169 ****** \n\nTASK [ovn-db : Flush handlers] *************************************************\nTuesday 24 February 2026  00:52:08 +0000 (0:00:00.089)       0:01:25.259 ****** \n\nTASK [ovn-db : Flush handlers] *************************************************\nTuesday 24 February 2026  00:52:08 +0000 (0:00:00.081)       0:01:25.340 ****** \n\nRUNNING HANDLER [ovn-db : Restart ovn-nb-db container] *************************\nTuesday 24 February 2026  00:52:08 +0000 (0:00:00.092)       0:01:25.433 ****** \n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nRUNNING HANDLER [ovn-db : Restart ovn-sb-db container] *************************\nTuesday 24 February 2026  00:52:15 +0000 (0:00:07.022)       0:01:32.455 ****** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [ovn-db : Restart ovn-northd container] ************************\nTuesday 24 February 2026  00:52:18 +0000 (0:00:03.035)       0:01:35.490 ****** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [ovn-db : Wait for leader election] ***************************************\nTuesday 24 February 2026  00:52:21 +0000 (0:00:02.988)       0:01:38.478 ****** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [ovn-db : Get OVN_Northbound cluster leader] ******************************\nTuesday 24 February 2026  00:52:21 +0000 (0:00:00.101)       0:01:38.580 ****** \n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\nTASK [ovn-db : Configure OVN NB connection settings] ***************************\nTuesday 24 February 2026  00:52:22 +0000 (0:00:00.753)       0:01:39.334 ****** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [ovn-db : Get OVN_Southbound cluster leader] ******************************\nTuesday 24 February 2026  00:52:22 +0000 (0:00:00.651)       0:01:39.985 ****** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ovn-db : Configure OVN SB connection settings] ***************************\nTuesday 24 February 2026  00:52:23 +0000 (0:00:00.889)       0:01:40.874 ****** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [ovn-db : Wait for ovn-nb-db] *********************************************\nTuesday 24 February 2026  00:52:24 +0000 (0:00:00.734)       0:01:41.609 ****** \n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ovn-db : Wait for ovn-sb-db] *********************************************\nTuesday 24 February 2026  00:52:25 +0000 (0:00:00.919)       0:01:42.529 ****** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ovn-db : Unset bootstrap args fact] **************************************\nTuesday 24 February 2026  00:52:26 +0000 (0:00:00.720)       0:01:43.249 ****** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ovn-db : Ensuring config directories exist] ******************************\nTuesday 24 February 2026  00:52:26 +0000 (0:00:00.280)       0:01:43.530 ****** \n\u001b[0;32mok: [testbed-node-0] => (item={'key': 'ovn-northd', 'value': {'container_name': 'ovn_northd', 'group': 'ovn-northd', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-northd:2024.2', 'volumes': ['/etc/kolla/ovn-northd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item={'key': 'ovn-northd', 'value': {'container_name': 'ovn_northd', 'group': 'ovn-northd', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-northd:2024.2', 'volumes': ['/etc/kolla/ovn-northd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item={'key': 'ovn-northd', 'value': {'container_name': 'ovn_northd', 'group': 'ovn-northd', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-northd:2024.2', 'volumes': ['/etc/kolla/ovn-northd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;32mok: [testbed-node-0] => (item={'key': 'ovn-nb-db', 'value': {'container_name': 'ovn_nb_db', 'group': 'ovn-nb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-nb-db-server:2024.2', 'volumes': ['/etc/kolla/ovn-nb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_nb_db:/var/lib/openvswitch/ovn-nb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item={'key': 'ovn-nb-db', 'value': {'container_name': 'ovn_nb_db', 'group': 'ovn-nb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-nb-db-server:2024.2', 'volumes': ['/etc/kolla/ovn-nb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_nb_db:/var/lib/openvswitch/ovn-nb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item={'key': 'ovn-nb-db', 'value': {'container_name': 'ovn_nb_db', 'group': 'ovn-nb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-nb-db-server:2024.2', 'volumes': ['/etc/kolla/ovn-nb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_nb_db:/var/lib/openvswitch/ovn-nb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;32mok: [testbed-node-0] => (item={'key': 'ovn-sb-db', 'value': {'container_name': 'ovn_sb_db', 'group': 'ovn-sb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-sb-db-server:2024.2', 'volumes': ['/etc/kolla/ovn-sb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_sb_db:/var/lib/openvswitch/ovn-sb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item={'key': 'ovn-sb-db', 'value': {'container_name': 'ovn_sb_db', 'group': 'ovn-sb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-sb-db-server:2024.2', 'volumes': ['/etc/kolla/ovn-sb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_sb_db:/var/lib/openvswitch/ovn-sb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item={'key': 'ovn-sb-db', 'value': {'container_name': 'ovn_sb_db', 'group': 'ovn-sb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-sb-db-server:2024.2', 'volumes': ['/etc/kolla/ovn-sb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_sb_db:/var/lib/openvswitch/ovn-sb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\nTASK [ovn-db : Copying over config.json files for services] ********************\nTuesday 24 February 2026  00:52:27 +0000 (0:00:01.430)       0:01:44.960 ****** \n\u001b[0;32mok: [testbed-node-0] => (item={'key': 'ovn-northd', 'value': {'container_name': 'ovn_northd', 'group': 'ovn-northd', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-northd:2024.2', 'volumes': ['/etc/kolla/ovn-northd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item={'key': 'ovn-northd', 'value': {'container_name': 'ovn_northd', 'group': 'ovn-northd', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-northd:2024.2', 'volumes': ['/etc/kolla/ovn-northd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item={'key': 'ovn-northd', 'value': {'container_name': 'ovn_northd', 'group': 'ovn-northd', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-northd:2024.2', 'volumes': ['/etc/kolla/ovn-northd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;32mok: [testbed-node-0] => (item={'key': 'ovn-nb-db', 'value': {'container_name': 'ovn_nb_db', 'group': 'ovn-nb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-nb-db-server:2024.2', 'volumes': ['/etc/kolla/ovn-nb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_nb_db:/var/lib/openvswitch/ovn-nb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ovn-nb-db', 'value': {'container_name': 'ovn_nb_db', 'group': 'ovn-nb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-nb-db-server:2024.2', 'volumes': ['/etc/kolla/ovn-nb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_nb_db:/var/lib/openvswitch/ovn-nb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ovn-nb-db', 'value': {'container_name': 'ovn_nb_db', 'group': 'ovn-nb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-nb-db-server:2024.2', 'volumes': ['/etc/kolla/ovn-nb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_nb_db:/var/lib/openvswitch/ovn-nb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;32mok: [testbed-node-0] => (item={'key': 'ovn-sb-db', 'value': {'container_name': 'ovn_sb_db', 'group': 'ovn-sb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-sb-db-server:2024.2', 'volumes': ['/etc/kolla/ovn-sb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_sb_db:/var/lib/openvswitch/ovn-sb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ovn-sb-db', 'value': {'container_name': 'ovn_sb_db', 'group': 'ovn-sb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-sb-db-server:2024.2', 'volumes': ['/etc/kolla/ovn-sb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_sb_db:/var/lib/openvswitch/ovn-sb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ovn-sb-db', 'value': {'container_name': 'ovn_sb_db', 'group': 'ovn-sb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-sb-db-server:2024.2', 'volumes': ['/etc/kolla/ovn-sb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_sb_db:/var/lib/openvswitch/ovn-sb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\nTASK [ovn-db : Check ovn containers] *******************************************\nTuesday 24 February 2026  00:52:31 +0000 (0:00:03.749)       0:01:48.710 ****** \n\u001b[0;32mok: [testbed-node-0] => (item={'key': 'ovn-northd', 'value': {'container_name': 'ovn_northd', 'group': 'ovn-northd', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-northd:2024.2', 'volumes': ['/etc/kolla/ovn-northd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item={'key': 'ovn-northd', 'value': {'container_name': 'ovn_northd', 'group': 'ovn-northd', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-northd:2024.2', 'volumes': ['/etc/kolla/ovn-northd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item={'key': 'ovn-northd', 'value': {'container_name': 'ovn_northd', 'group': 'ovn-northd', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-northd:2024.2', 'volumes': ['/etc/kolla/ovn-northd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;32mok: [testbed-node-0] => (item={'key': 'ovn-nb-db', 'value': {'container_name': 'ovn_nb_db', 'group': 'ovn-nb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-nb-db-server:2024.2', 'volumes': ['/etc/kolla/ovn-nb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_nb_db:/var/lib/openvswitch/ovn-nb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ovn-nb-db', 'value': {'container_name': 'ovn_nb_db', 'group': 'ovn-nb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-nb-db-server:2024.2', 'volumes': ['/etc/kolla/ovn-nb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_nb_db:/var/lib/openvswitch/ovn-nb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ovn-nb-db', 'value': {'container_name': 'ovn_nb_db', 'group': 'ovn-nb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-nb-db-server:2024.2', 'volumes': ['/etc/kolla/ovn-nb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_nb_db:/var/lib/openvswitch/ovn-nb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;32mok: [testbed-node-0] => (item={'key': 'ovn-sb-db', 'value': {'container_name': 'ovn_sb_db', 'group': 'ovn-sb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-sb-db-server:2024.2', 'volumes': ['/etc/kolla/ovn-sb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_sb_db:/var/lib/openvswitch/ovn-sb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ovn-sb-db', 'value': {'container_name': 'ovn_sb_db', 'group': 'ovn-sb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-sb-db-server:2024.2', 'volumes': ['/etc/kolla/ovn-sb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_sb_db:/var/lib/openvswitch/ovn-sb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ovn-sb-db', 'value': {'container_name': 'ovn_sb_db', 'group': 'ovn-sb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-sb-db-server:2024.2', 'volumes': ['/etc/kolla/ovn-sb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_sb_db:/var/lib/openvswitch/ovn-sb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\nTASK [ovn-db : Flush handlers] *************************************************\nTuesday 24 February 2026  00:52:34 +0000 (0:00:03.027)       0:01:51.737 ****** \n\nTASK [ovn-db : Flush handlers] *************************************************\nTuesday 24 February 2026  00:52:34 +0000 (0:00:00.105)       0:01:51.843 ****** \n\nTASK [ovn-db : Flush handlers] *************************************************\nTuesday 24 February 2026  00:52:34 +0000 (0:00:00.069)       0:01:51.912 ****** \n\nRUNNING HANDLER [ovn-db : Restart ovn-nb-db container] *************************\nTuesday 24 February 2026  00:52:34 +0000 (0:00:00.066)       0:01:51.979 ****** \n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [ovn-db : Restart ovn-sb-db container] *************************\nTuesday 24 February 2026  00:52:41 +0000 (0:00:06.527)       0:01:58.508 ****** \n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [ovn-db : Restart ovn-northd container] ************************\nTuesday 24 February 2026  00:52:47 +0000 (0:00:06.167)       0:02:04.675 ****** \n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [ovn-db : Wait for leader election] ***************************************\nTuesday 24 February 2026  00:52:54 +0000 (0:00:06.811)       0:02:11.486 ****** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [ovn-db : Get OVN_Northbound cluster leader] ******************************\nTuesday 24 February 2026  00:52:54 +0000 (0:00:00.124)       0:02:11.611 ****** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\nTASK [ovn-db : Configure OVN NB connection settings] ***************************\nTuesday 24 February 2026  00:52:55 +0000 (0:00:00.849)       0:02:12.461 ****** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [ovn-db : Get OVN_Southbound cluster leader] ******************************\nTuesday 24 February 2026  00:52:56 +0000 (0:00:00.804)       0:02:13.265 ****** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ovn-db : Configure OVN SB connection settings] ***************************\nTuesday 24 February 2026  00:52:57 +0000 (0:00:00.881)       0:02:14.147 ****** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [ovn-db : Wait for ovn-nb-db] *********************************************\nTuesday 24 February 2026  00:52:57 +0000 (0:00:00.666)       0:02:14.813 ****** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\nTASK [ovn-db : Wait for ovn-sb-db] *********************************************\nTuesday 24 February 2026  00:52:58 +0000 (0:00:00.807)       0:02:15.621 ****** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=44  \u001b[0m \u001b[0;33mchanged=18  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=20  \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=43  \u001b[0m \u001b[0;33mchanged=19  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=22  \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=43  \u001b[0m \u001b[0;33mchanged=19  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=22  \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-3\u001b[0m             : \u001b[0;32mok=12  \u001b[0m \u001b[0;33mchanged=8   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-4\u001b[0m             : \u001b[0;32mok=12  \u001b[0m \u001b[0;33mchanged=8   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-5\u001b[0m             : \u001b[0;32mok=12  \u001b[0m \u001b[0;33mchanged=8   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nTuesday 24 February 2026  00:52:59 +0000 (0:00:01.001)       0:02:16.623 ****** \n=============================================================================== \novn-controller : Restart ovn-controller container ---------------------- 25.87s\novn-controller : Configure OVN in OVSDB -------------------------------- 22.57s\novn-db : Restart ovn-nb-db container ----------------------------------- 13.55s\novn-db : Restart ovn-northd container ----------------------------------- 9.80s\novn-db : Restart ovn-sb-db container ------------------------------------ 9.20s\novn-db : Copying over config.json files for services -------------------- 5.46s\novn-db : Copying over config.json files for services -------------------- 3.75s\novn-db : Check ovn containers ------------------------------------------- 3.03s\novn-controller : Create br-int bridge on OpenvSwitch -------------------- 2.73s\novn-db : Check ovn containers ------------------------------------------- 2.54s\novn-db : Ensuring config directories exist ------------------------------ 1.83s\novn-controller : Copying over config.json files for services ------------ 1.68s\novn-controller : Reload systemd config ---------------------------------- 1.56s\novn-controller : Check ovn-controller containers ------------------------ 1.52s\novn-controller : Ensuring config directories exist ---------------------- 1.50s\novn-controller : Copying over systemd override -------------------------- 1.45s\novn-db : Ensuring config directories exist ------------------------------ 1.43s\novn-controller : Ensuring systemd override directory exists ------------- 1.37s\novn-controller : include_tasks ------------------------------------------ 1.03s\nGroup hosts based on enabled services ----------------------------------- 1.01s\n\u001b[32m2026-02-24 00:53:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:53:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 404270fa-50db-4a78-8c44-21df48acba39 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:53:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 00:53:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:53:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 404270fa-50db-4a78-8c44-21df48acba39 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:53:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 00:53:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:53:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 404270fa-50db-4a78-8c44-21df48acba39 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:53:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 00:53:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:53:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 404270fa-50db-4a78-8c44-21df48acba39 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:53:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 00:53:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:53:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 404270fa-50db-4a78-8c44-21df48acba39 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:53:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 00:53:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:53:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 404270fa-50db-4a78-8c44-21df48acba39 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:53:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 00:53:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:53:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 404270fa-50db-4a78-8c44-21df48acba39 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:53:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 00:53:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:53:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 404270fa-50db-4a78-8c44-21df48acba39 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:53:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 00:53:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:53:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 404270fa-50db-4a78-8c44-21df48acba39 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:53:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 00:53:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:53:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 404270fa-50db-4a78-8c44-21df48acba39 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:53:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 00:53:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:53:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 404270fa-50db-4a78-8c44-21df48acba39 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:53:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 00:53:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:53:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 404270fa-50db-4a78-8c44-21df48acba39 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:53:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 00:53:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:53:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 404270fa-50db-4a78-8c44-21df48acba39 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:53:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 00:53:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:53:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 404270fa-50db-4a78-8c44-21df48acba39 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:53:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 00:53:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:53:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 404270fa-50db-4a78-8c44-21df48acba39 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:53:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 00:53:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:53:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 404270fa-50db-4a78-8c44-21df48acba39 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:53:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 00:53:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:53:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 404270fa-50db-4a78-8c44-21df48acba39 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:53:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 00:53:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:53:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 404270fa-50db-4a78-8c44-21df48acba39 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:53:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 00:53:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:53:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 404270fa-50db-4a78-8c44-21df48acba39 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:53:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 00:53:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:53:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 404270fa-50db-4a78-8c44-21df48acba39 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:53:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 00:54:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:54:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 404270fa-50db-4a78-8c44-21df48acba39 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:54:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 00:54:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:54:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 404270fa-50db-4a78-8c44-21df48acba39 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:54:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 00:54:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:54:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 404270fa-50db-4a78-8c44-21df48acba39 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:54:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 00:54:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:54:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 404270fa-50db-4a78-8c44-21df48acba39 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:54:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 00:54:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:54:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 404270fa-50db-4a78-8c44-21df48acba39 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:54:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 00:54:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:54:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 404270fa-50db-4a78-8c44-21df48acba39 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:54:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 00:54:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:54:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 404270fa-50db-4a78-8c44-21df48acba39 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:54:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 00:54:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:54:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 404270fa-50db-4a78-8c44-21df48acba39 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:54:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 00:54:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:54:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 404270fa-50db-4a78-8c44-21df48acba39 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:54:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 00:54:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:54:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 404270fa-50db-4a78-8c44-21df48acba39 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:54:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 00:54:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:54:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 404270fa-50db-4a78-8c44-21df48acba39 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:54:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 00:54:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:54:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 404270fa-50db-4a78-8c44-21df48acba39 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:54:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 00:54:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:54:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 404270fa-50db-4a78-8c44-21df48acba39 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:54:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 00:54:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:54:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 404270fa-50db-4a78-8c44-21df48acba39 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:54:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 00:54:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:54:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 404270fa-50db-4a78-8c44-21df48acba39 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:54:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 00:54:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:54:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 404270fa-50db-4a78-8c44-21df48acba39 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:54:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 00:54:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:54:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 404270fa-50db-4a78-8c44-21df48acba39 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:54:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 00:54:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:54:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 404270fa-50db-4a78-8c44-21df48acba39 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:54:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 00:54:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:54:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 404270fa-50db-4a78-8c44-21df48acba39 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:54:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 00:54:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:54:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 404270fa-50db-4a78-8c44-21df48acba39 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:54:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 00:55:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:55:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 404270fa-50db-4a78-8c44-21df48acba39 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:55:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 00:55:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:55:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 404270fa-50db-4a78-8c44-21df48acba39 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:55:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 00:55:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:55:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 404270fa-50db-4a78-8c44-21df48acba39 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:55:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 00:55:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:55:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 404270fa-50db-4a78-8c44-21df48acba39 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:55:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 00:55:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:55:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 404270fa-50db-4a78-8c44-21df48acba39 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:55:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 00:55:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:55:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 404270fa-50db-4a78-8c44-21df48acba39 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:55:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 00:55:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:55:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 404270fa-50db-4a78-8c44-21df48acba39 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:55:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 00:55:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:55:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 404270fa-50db-4a78-8c44-21df48acba39 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:55:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 00:55:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:55:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 404270fa-50db-4a78-8c44-21df48acba39 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:55:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 00:55:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:55:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 404270fa-50db-4a78-8c44-21df48acba39 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:55:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 00:55:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:55:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 404270fa-50db-4a78-8c44-21df48acba39 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:55:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 00:55:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:55:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 404270fa-50db-4a78-8c44-21df48acba39 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:55:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 00:55:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:55:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 404270fa-50db-4a78-8c44-21df48acba39 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:55:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 00:55:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:55:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 404270fa-50db-4a78-8c44-21df48acba39 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:55:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 00:55:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:55:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 404270fa-50db-4a78-8c44-21df48acba39 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:55:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 00:55:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:55:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 404270fa-50db-4a78-8c44-21df48acba39 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:55:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 00:55:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:55:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 404270fa-50db-4a78-8c44-21df48acba39 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:55:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 00:55:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:55:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 404270fa-50db-4a78-8c44-21df48acba39 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:55:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 00:55:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:55:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 404270fa-50db-4a78-8c44-21df48acba39 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:55:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 00:56:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:56:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 404270fa-50db-4a78-8c44-21df48acba39 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:56:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 00:56:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:56:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 404270fa-50db-4a78-8c44-21df48acba39 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:56:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 00:56:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:56:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 404270fa-50db-4a78-8c44-21df48acba39 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:56:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 00:56:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:56:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 404270fa-50db-4a78-8c44-21df48acba39 is in state SUCCESS\u001b[0m\n\n\nPLAY [Group hosts based on configuration] **************************************\n\nTASK [Group hosts based on Kolla action] ***************************************\nTuesday 24 February 2026  00:49:35 +0000 (0:00:00.368)       0:00:00.368 ****** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [Group hosts based on enabled services] ***********************************\nTuesday 24 February 2026  00:49:35 +0000 (0:00:00.452)       0:00:00.820 ****** \n\u001b[0;32mok: [testbed-node-0] => (item=enable_loadbalancer_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item=enable_loadbalancer_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item=enable_loadbalancer_True)\u001b[0m\n\nPLAY [Apply role loadbalancer] *************************************************\n\nTASK [loadbalancer : include_tasks] ********************************************\nTuesday 24 February 2026  00:49:36 +0000 (0:00:00.426)       0:00:01.247 ****** \n\u001b[0;36mincluded: /ansible/roles/loadbalancer/tasks/deploy.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [loadbalancer : Check IPv6 support] ***************************************\nTuesday 24 February 2026  00:49:37 +0000 (0:00:00.988)       0:00:02.235 ****** \n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [Setting sysctl values] ***************************************************\nTuesday 24 February 2026  00:49:38 +0000 (0:00:01.009)       0:00:03.244 ****** \n\u001b[0;36mincluded: sysctl for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [sysctl : Check IPv6 support] *********************************************\nTuesday 24 February 2026  00:49:39 +0000 (0:00:01.173)       0:00:04.418 ****** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [sysctl : Setting sysctl values] ******************************************\nTuesday 24 February 2026  00:49:40 +0000 (0:00:00.666)       0:00:05.084 ****** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'name': 'net.ipv6.ip_nonlocal_bind', 'value': 1})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'name': 'net.ipv6.ip_nonlocal_bind', 'value': 1})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'name': 'net.ipv4.ip_nonlocal_bind', 'value': 1})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'name': 'net.ipv4.ip_nonlocal_bind', 'value': 1})\u001b[0m\n\u001b[0;32mok: [testbed-node-0] => (item={'name': 'net.ipv4.tcp_retries2', 'value': 'KOLLA_UNSET'})\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item={'name': 'net.ipv4.tcp_retries2', 'value': 'KOLLA_UNSET'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'name': 'net.ipv6.ip_nonlocal_bind', 'value': 1})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'name': 'net.unix.max_dgram_qlen', 'value': 128})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'name': 'net.unix.max_dgram_qlen', 'value': 128})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'name': 'net.ipv4.ip_nonlocal_bind', 'value': 1})\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item={'name': 'net.ipv4.tcp_retries2', 'value': 'KOLLA_UNSET'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'name': 'net.unix.max_dgram_qlen', 'value': 128})\u001b[0m\n\nTASK [module-load : Load modules] **********************************************\nTuesday 24 February 2026  00:49:43 +0000 (0:00:03.747)       0:00:08.832 ****** \n\u001b[0;33mchanged: [testbed-node-0] => (item=ip_vs)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=ip_vs)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=ip_vs)\u001b[0m\n\nTASK [module-load : Persist modules via modules-load.d] ************************\nTuesday 24 February 2026  00:49:44 +0000 (0:00:00.832)       0:00:09.665 ****** \n\u001b[0;33mchanged: [testbed-node-0] => (item=ip_vs)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=ip_vs)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=ip_vs)\u001b[0m\n\nTASK [module-load : Drop module persistence] ***********************************\nTuesday 24 February 2026  00:49:46 +0000 (0:00:01.418)       0:00:11.083 ****** \n\u001b[0;36mskipping: [testbed-node-0] => (item=ip_vs) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=ip_vs) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=ip_vs) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [loadbalancer : Ensuring config directories exist] ************************\nTuesday 24 February 2026  00:49:47 +0000 (0:00:01.199)       0:00:12.282 ****** \n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:61313'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:61313'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:61313'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2024.2', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2024.2', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2024.2', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}})\u001b[0m\n\nTASK [loadbalancer : Ensuring haproxy service config subdir exists] ************\nTuesday 24 February 2026  00:49:49 +0000 (0:00:02.010)       0:00:14.292 ****** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\nTASK [loadbalancer : Ensuring proxysql service config subdirectories exist] ****\nTuesday 24 February 2026  00:49:50 +0000 (0:00:01.202)       0:00:15.495 ****** \n\u001b[0;33mchanged: [testbed-node-0] => (item=users)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=users)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=users)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=rules)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=rules)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=rules)\u001b[0m\n\nTASK [loadbalancer : Ensuring keepalived checks subdir exists] *****************\nTuesday 24 February 2026  00:49:52 +0000 (0:00:01.637)       0:00:17.133 ****** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [loadbalancer : Remove mariadb.cfg if proxysql enabled] *******************\nTuesday 24 February 2026  00:49:53 +0000 (0:00:01.733)       0:00:18.866 ****** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [loadbalancer : Removing checks for services which are disabled] **********\nTuesday 24 February 2026  00:49:56 +0000 (0:00:02.364)       0:00:21.231 ****** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:61313'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2024.2', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'haproxy-ssh', 'value': {'container_name': 'haproxy_ssh', 'group': 'loadbalancer', 'enabled': False, 'image': 'registry.osism.tech/kolla/haproxy-ssh:2024.2', 'volumes': ['/etc/kolla/haproxy-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', '__omit_place_holder__852810438d06097b6927d4e594558f870baad287', '__omit_place_holder__852810438d06097b6927d4e594558f870baad287'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 2985'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:61313'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:61313'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2024.2', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2024.2', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'haproxy-ssh', 'value': {'container_name': 'haproxy_ssh', 'group': 'loadbalancer', 'enabled': False, 'image': 'registry.osism.tech/kolla/haproxy-ssh:2024.2', 'volumes': ['/etc/kolla/haproxy-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', '__omit_place_holder__852810438d06097b6927d4e594558f870baad287', '__omit_place_holder__852810438d06097b6927d4e594558f870baad287'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 2985'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'haproxy-ssh', 'value': {'container_name': 'haproxy_ssh', 'group': 'loadbalancer', 'enabled': False, 'image': 'registry.osism.tech/kolla/haproxy-ssh:2024.2', 'volumes': ['/etc/kolla/haproxy-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', '__omit_place_holder__852810438d06097b6927d4e594558f870baad287', '__omit_place_holder__852810438d06097b6927d4e594558f870baad287'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 2985'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [loadbalancer : Copying checks for services which are enabled] ************\nTuesday 24 February 2026  00:49:57 +0000 (0:00:00.748)       0:00:21.979 ****** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:61313'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:61313'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:61313'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2024.2', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'haproxy-ssh', 'value': {'container_name': 'haproxy_ssh', 'group': 'loadbalancer', 'enabled': False, 'image': 'registry.osism.tech/kolla/haproxy-ssh:2024.2', 'volumes': ['/etc/kolla/haproxy-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', '__omit_place_holder__852810438d06097b6927d4e594558f870baad287', '__omit_place_holder__852810438d06097b6927d4e594558f870baad287'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 2985'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2024.2', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'haproxy-ssh', 'value': {'container_name': 'haproxy_ssh', 'group': 'loadbalancer', 'enabled': False, 'image': 'registry.osism.tech/kolla/haproxy-ssh:2024.2', 'volumes': ['/etc/kolla/haproxy-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', '__omit_place_holder__852810438d06097b6927d4e594558f870baad287', '__omit_place_holder__852810438d06097b6927d4e594558f870baad287'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 2985'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2024.2', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'haproxy-ssh', 'value': {'container_name': 'haproxy_ssh', 'group': 'loadbalancer', 'enabled': False, 'image': 'registry.osism.tech/kolla/haproxy-ssh:2024.2', 'volumes': ['/etc/kolla/haproxy-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', '__omit_place_holder__852810438d06097b6927d4e594558f870baad287', '__omit_place_holder__852810438d06097b6927d4e594558f870baad287'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 2985'], 'timeout': '30'}}}) \u001b[0m\n\nTASK [loadbalancer : Copying over config.json files for services] **************\nTuesday 24 February 2026  00:49:59 +0000 (0:00:02.501)       0:00:24.481 ****** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:61313'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:61313'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:61313'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2024.2', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2024.2', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2024.2', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}})\u001b[0m\n\nTASK [loadbalancer : Copying over haproxy.cfg] *********************************\nTuesday 24 February 2026  00:50:02 +0000 (0:00:02.886)       0:00:27.367 ****** \n\u001b[0;33mchanged: [testbed-node-2] => (item=/ansible/roles/loadbalancer/templates/haproxy/haproxy_main.cfg.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=/ansible/roles/loadbalancer/templates/haproxy/haproxy_main.cfg.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=/ansible/roles/loadbalancer/templates/haproxy/haproxy_main.cfg.j2)\u001b[0m\n\nTASK [loadbalancer : Copying over proxysql config] *****************************\nTuesday 24 February 2026  00:50:05 +0000 (0:00:03.147)       0:00:30.514 ****** \n\u001b[0;33mchanged: [testbed-node-0] => (item=/ansible/roles/loadbalancer/templates/proxysql/proxysql.yaml.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=/ansible/roles/loadbalancer/templates/proxysql/proxysql.yaml.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=/ansible/roles/loadbalancer/templates/proxysql/proxysql.yaml.j2)\u001b[0m\n\nTASK [loadbalancer : Copying over haproxy single external frontend config] *****\nTuesday 24 February 2026  00:50:10 +0000 (0:00:04.951)       0:00:35.466 ****** \n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [loadbalancer : Copying over custom haproxy services configuration] *******\nTuesday 24 February 2026  00:50:11 +0000 (0:00:00.930)       0:00:36.396 ****** \n\u001b[0;33mchanged: [testbed-node-0] => (item=/opt/configuration/environments/kolla/files/overlays/haproxy/services.d/haproxy.cfg)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=/opt/configuration/environments/kolla/files/overlays/haproxy/services.d/haproxy.cfg)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=/opt/configuration/environments/kolla/files/overlays/haproxy/services.d/haproxy.cfg)\u001b[0m\n\nTASK [loadbalancer : Copying over keepalived.conf] *****************************\nTuesday 24 February 2026  00:50:14 +0000 (0:00:02.905)       0:00:39.302 ****** \n\u001b[0;33mchanged: [testbed-node-0] => (item=/ansible/roles/loadbalancer/templates/keepalived/keepalived.conf.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=/ansible/roles/loadbalancer/templates/keepalived/keepalived.conf.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=/ansible/roles/loadbalancer/templates/keepalived/keepalived.conf.j2)\u001b[0m\n\nTASK [loadbalancer : Copying over haproxy.pem] *********************************\nTuesday 24 February 2026  00:50:17 +0000 (0:00:02.861)       0:00:42.163 ****** \n\u001b[0;33mchanged: [testbed-node-0] => (item=haproxy.pem)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=haproxy.pem)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=haproxy.pem)\u001b[0m\n\nTASK [loadbalancer : Copying over haproxy-internal.pem] ************************\nTuesday 24 February 2026  00:50:18 +0000 (0:00:01.532)       0:00:43.696 ****** \n\u001b[0;33mchanged: [testbed-node-0] => (item=haproxy-internal.pem)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=haproxy-internal.pem)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=haproxy-internal.pem)\u001b[0m\n\nTASK [loadbalancer : include_tasks] ********************************************\nTuesday 24 February 2026  00:50:20 +0000 (0:00:01.388)       0:00:45.085 ****** \n\u001b[0;36mincluded: /ansible/roles/loadbalancer/tasks/copy-certs.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [service-cert-copy : loadbalancer | Copying over extra CA certificates] ***\nTuesday 24 February 2026  00:50:20 +0000 (0:00:00.805)       0:00:45.891 ****** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:61313'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:61313'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:61313'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2024.2', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2024.2', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2024.2', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}})\u001b[0m\n\nTASK [service-cert-copy : loadbalancer | Copying over backend internal TLS certificate] ***\nTuesday 24 February 2026  00:50:24 +0000 (0:00:03.628)       0:00:49.519 ****** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:61313'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:61313'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2024.2', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2024.2', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:61313'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2024.2', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [service-cert-copy : loadbalancer | Copying over backend internal TLS key] ***\nTuesday 24 February 2026  00:50:25 +0000 (0:00:01.014)       0:00:50.533 ****** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:61313'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2024.2', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:61313'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2024.2', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:61313'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2024.2', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [service-cert-copy : mariadb | Copying over extra CA certificates] ********\nTuesday 24 February 2026  00:50:26 +0000 (0:00:00.845)       0:00:51.379 ****** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:61313'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2024.2', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:61313'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2024.2', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:61313'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2024.2', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [service-cert-copy : mariadb | Copying over backend internal TLS certificate] ***\nTuesday 24 February 2026  00:50:27 +0000 (0:00:00.848)       0:00:52.227 ****** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:61313'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2024.2', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:61313'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2024.2', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:61313'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2024.2', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [service-cert-copy : mariadb | Copying over backend internal TLS key] *****\nTuesday 24 February 2026  00:50:28 +0000 (0:00:00.909)       0:00:53.137 ****** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:61313'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2024.2', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:61313'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2024.2', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:61313'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2024.2', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [service-cert-copy : proxysql | Copying over extra CA certificates] *******\nTuesday 24 February 2026  00:50:29 +0000 (0:00:01.005)       0:00:54.142 ****** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:61313'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2024.2', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:61313'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2024.2', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:61313'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2024.2', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [service-cert-copy : proxysql | Copying over backend internal TLS certificate] ***\nTuesday 24 February 2026  00:50:31 +0000 (0:00:02.049)       0:00:56.192 ****** \n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:61313'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2024.2', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:61313'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2024.2', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:61313'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2024.2', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [service-cert-copy : proxysql | Copying over backend internal TLS key] ****\nTuesday 24 February 2026  00:50:32 +0000 (0:00:00.950)       0:00:57.143 ****** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:61313'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2024.2', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:61313'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2024.2', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:61313'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2024.2', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [loadbalancer : Copying over haproxy start script] ************************\nTuesday 24 February 2026  00:50:33 +0000 (0:00:00.774)       0:00:57.917 ****** \n\u001b[0;33mchanged: [testbed-node-0] => (item=/ansible/roles/loadbalancer/templates/haproxy/haproxy_run.sh.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=/ansible/roles/loadbalancer/templates/haproxy/haproxy_run.sh.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=/ansible/roles/loadbalancer/templates/haproxy/haproxy_run.sh.j2)\u001b[0m\n\nTASK [loadbalancer : Copying over proxysql start script] ***********************\nTuesday 24 February 2026  00:50:34 +0000 (0:00:01.797)       0:00:59.714 ****** \n\u001b[0;33mchanged: [testbed-node-0] => (item=/ansible/roles/loadbalancer/templates/proxysql/proxysql_run.sh.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=/ansible/roles/loadbalancer/templates/proxysql/proxysql_run.sh.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=/ansible/roles/loadbalancer/templates/proxysql/proxysql_run.sh.j2)\u001b[0m\n\nTASK [loadbalancer : Copying files for haproxy-ssh] ****************************\nTuesday 24 February 2026  00:50:36 +0000 (0:00:01.728)       0:01:01.443 ****** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'src': 'haproxy-ssh/sshd_config.j2', 'dest': 'sshd_config'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'src': 'haproxy-ssh/sshd_config.j2', 'dest': 'sshd_config'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'src': 'haproxy-ssh/sshd_config.j2', 'dest': 'sshd_config'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'src': 'haproxy-ssh/id_rsa.pub', 'dest': 'id_rsa.pub'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'src': 'haproxy-ssh/id_rsa.pub', 'dest': 'id_rsa.pub'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'src': 'haproxy-ssh/id_rsa.pub', 'dest': 'id_rsa.pub'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [loadbalancer : Check loadbalancer containers] ****************************\nTuesday 24 February 2026  00:50:37 +0000 (0:00:01.067)       0:01:02.511 ****** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:61313'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:61313'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:61313'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2024.2', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2024.2', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2024.2', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}})\u001b[0m\n\nTASK [include_role : aodh] *****************************************************\nTuesday 24 February 2026  00:50:40 +0000 (0:00:02.955)       0:01:05.467 ****** \n\u001b[0;36mincluded: aodh for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [haproxy-config : Copying over aodh haproxy config] ***********************\nTuesday 24 February 2026  00:50:41 +0000 (0:00:00.551)       0:01:06.019 ****** \n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'aodh-api', 'value': {'container_name': 'aodh_api', 'group': 'aodh-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/aodh-api:2024.2', 'volumes': ['/etc/kolla/aodh-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'aodh:/var/lib/aodh/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8042'], 'timeout': '30'}, 'haproxy': {'aodh_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8042', 'listen_port': '8042'}, 'aodh_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8042', 'listen_port': '8042'}}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'aodh-evaluator', 'value': {'container_name': 'aodh_evaluator', 'group': 'aodh-evaluator', 'enabled': True, 'image': 'registry.osism.tech/kolla/aodh-evaluator:2024.2', 'volumes': ['/etc/kolla/aodh-evaluator/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-evaluator 3306'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'aodh-listener', 'value': {'container_name': 'aodh_listener', 'group': 'aodh-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/aodh-listener:2024.2', 'volumes': ['/etc/kolla/aodh-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-listener 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'aodh-notifier', 'value': {'container_name': 'aodh_notifier', 'group': 'aodh-notifier', 'enabled': True, 'image': 'registry.osism.tech/kolla/aodh-notifier:2024.2', 'volumes': ['/etc/kolla/aodh-notifier/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-notifier 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'aodh-api', 'value': {'container_name': 'aodh_api', 'group': 'aodh-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/aodh-api:2024.2', 'volumes': ['/etc/kolla/aodh-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'aodh:/var/lib/aodh/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8042'], 'timeout': '30'}, 'haproxy': {'aodh_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8042', 'listen_port': '8042'}, 'aodh_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8042', 'listen_port': '8042'}}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'aodh-evaluator', 'value': {'container_name': 'aodh_evaluator', 'group': 'aodh-evaluator', 'enabled': True, 'image': 'registry.osism.tech/kolla/aodh-evaluator:2024.2', 'volumes': ['/etc/kolla/aodh-evaluator/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-evaluator 3306'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'aodh-listener', 'value': {'container_name': 'aodh_listener', 'group': 'aodh-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/aodh-listener:2024.2', 'volumes': ['/etc/kolla/aodh-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-listener 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'aodh-notifier', 'value': {'container_name': 'aodh_notifier', 'group': 'aodh-notifier', 'enabled': True, 'image': 'registry.osism.tech/kolla/aodh-notifier:2024.2', 'volumes': ['/etc/kolla/aodh-notifier/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-notifier 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'aodh-api', 'value': {'container_name': 'aodh_api', 'group': 'aodh-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/aodh-api:2024.2', 'volumes': ['/etc/kolla/aodh-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'aodh:/var/lib/aodh/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8042'], 'timeout': '30'}, 'haproxy': {'aodh_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8042', 'listen_port': '8042'}, 'aodh_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8042', 'listen_port': '8042'}}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'aodh-evaluator', 'value': {'container_name': 'aodh_evaluator', 'group': 'aodh-evaluator', 'enabled': True, 'image': 'registry.osism.tech/kolla/aodh-evaluator:2024.2', 'volumes': ['/etc/kolla/aodh-evaluator/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-evaluator 3306'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'aodh-listener', 'value': {'container_name': 'aodh_listener', 'group': 'aodh-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/aodh-listener:2024.2', 'volumes': ['/etc/kolla/aodh-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-listener 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'aodh-notifier', 'value': {'container_name': 'aodh_notifier', 'group': 'aodh-notifier', 'enabled': True, 'image': 'registry.osism.tech/kolla/aodh-notifier:2024.2', 'volumes': ['/etc/kolla/aodh-notifier/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-notifier 5672'], 'timeout': '30'}}}) \u001b[0m\n\nTASK [haproxy-config : Add configuration for aodh when using single external frontend] ***\nTuesday 24 February 2026  00:50:45 +0000 (0:00:04.644)       0:01:10.663 ****** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'aodh-api', 'value': {'container_name': 'aodh_api', 'group': 'aodh-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/aodh-api:2024.2', 'volumes': ['/etc/kolla/aodh-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'aodh:/var/lib/aodh/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8042'], 'timeout': '30'}, 'haproxy': {'aodh_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8042', 'listen_port': '8042'}, 'aodh_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8042', 'listen_port': '8042'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'aodh-evaluator', 'value': {'container_name': 'aodh_evaluator', 'group': 'aodh-evaluator', 'enabled': True, 'image': 'registry.osism.tech/kolla/aodh-evaluator:2024.2', 'volumes': ['/etc/kolla/aodh-evaluator/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-evaluator 3306'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'aodh-listener', 'value': {'container_name': 'aodh_listener', 'group': 'aodh-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/aodh-listener:2024.2', 'volumes': ['/etc/kolla/aodh-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-listener 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'aodh-notifier', 'value': {'container_name': 'aodh_notifier', 'group': 'aodh-notifier', 'enabled': True, 'image': 'registry.osism.tech/kolla/aodh-notifier:2024.2', 'volumes': ['/etc/kolla/aodh-notifier/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-notifier 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'aodh-api', 'value': {'container_name': 'aodh_api', 'group': 'aodh-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/aodh-api:2024.2', 'volumes': ['/etc/kolla/aodh-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'aodh:/var/lib/aodh/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8042'], 'timeout': '30'}, 'haproxy': {'aodh_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8042', 'listen_port': '8042'}, 'aodh_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8042', 'listen_port': '8042'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'aodh-evaluator', 'value': {'container_name': 'aodh_evaluator', 'group': 'aodh-evaluator', 'enabled': True, 'image': 'registry.osism.tech/kolla/aodh-evaluator:2024.2', 'volumes': ['/etc/kolla/aodh-evaluator/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-evaluator 3306'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'aodh-listener', 'value': {'container_name': 'aodh_listener', 'group': 'aodh-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/aodh-listener:2024.2', 'volumes': ['/etc/kolla/aodh-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-listener 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'aodh-notifier', 'value': {'container_name': 'aodh_notifier', 'group': 'aodh-notifier', 'enabled': True, 'image': 'registry.osism.tech/kolla/aodh-notifier:2024.2', 'volumes': ['/etc/kolla/aodh-notifier/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-notifier 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'aodh-api', 'value': {'container_name': 'aodh_api', 'group': 'aodh-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/aodh-api:2024.2', 'volumes': ['/etc/kolla/aodh-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'aodh:/var/lib/aodh/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8042'], 'timeout': '30'}, 'haproxy': {'aodh_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8042', 'listen_port': '8042'}, 'aodh_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8042', 'listen_port': '8042'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'aodh-evaluator', 'value': {'container_name': 'aodh_evaluator', 'group': 'aodh-evaluator', 'enabled': True, 'image': 'registry.osism.tech/kolla/aodh-evaluator:2024.2', 'volumes': ['/etc/kolla/aodh-evaluator/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-evaluator 3306'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'aodh-listener', 'value': {'container_name': 'aodh_listener', 'group': 'aodh-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/aodh-listener:2024.2', 'volumes': ['/etc/kolla/aodh-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-listener 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'aodh-notifier', 'value': {'container_name': 'aodh_notifier', 'group': 'aodh-notifier', 'enabled': True, 'image': 'registry.osism.tech/kolla/aodh-notifier:2024.2', 'volumes': ['/etc/kolla/aodh-notifier/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-notifier 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [haproxy-config : Configuring firewall for aodh] **************************\nTuesday 24 February 2026  00:50:47 +0000 (0:00:01.439)       0:01:12.103 ****** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'aodh_api', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8042', 'listen_port': '8042'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'aodh_api_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8042', 'listen_port': '8042'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'aodh_api', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8042', 'listen_port': '8042'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'aodh_api_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8042', 'listen_port': '8042'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'aodh_api', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8042', 'listen_port': '8042'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'aodh_api_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8042', 'listen_port': '8042'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [proxysql-config : Copying over aodh ProxySQL users config] ***************\nTuesday 24 February 2026  00:50:48 +0000 (0:00:00.937)       0:01:13.041 ****** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [proxysql-config : Copying over aodh ProxySQL rules config] ***************\nTuesday 24 February 2026  00:50:49 +0000 (0:00:01.461)       0:01:14.502 ****** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [include_role : barbican] *************************************************\nTuesday 24 February 2026  00:50:51 +0000 (0:00:01.959)       0:01:16.462 ****** \n\u001b[0;36mincluded: barbican for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [haproxy-config : Copying over barbican haproxy config] *******************\nTuesday 24 February 2026  00:50:52 +0000 (0:00:00.744)       0:01:17.206 ****** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-api:2024.2', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-keystone-listener:2024.2', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-worker:2024.2', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-api:2024.2', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-keystone-listener:2024.2', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-worker:2024.2', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-api:2024.2', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-keystone-listener:2024.2', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-worker:2024.2', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}}) \u001b[0m\n\nTASK [haproxy-config : Add configuration for barbican when using single external frontend] ***\nTuesday 24 February 2026  00:50:56 +0000 (0:00:03.789)       0:01:20.995 ****** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-api:2024.2', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-keystone-listener:2024.2', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-worker:2024.2', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-api:2024.2', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-keystone-listener:2024.2', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-worker:2024.2', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-api:2024.2', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-keystone-listener:2024.2', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-worker:2024.2', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [haproxy-config : Configuring firewall for barbican] **********************\nTuesday 24 February 2026  00:50:57 +0000 (0:00:01.468)       0:01:22.464 ****** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'barbican_api', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'barbican_api_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'barbican_api', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'barbican_api_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'barbican_api', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'barbican_api_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [proxysql-config : Copying over barbican ProxySQL users config] ***********\nTuesday 24 February 2026  00:50:59 +0000 (0:00:02.179)       0:01:24.644 ****** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [proxysql-config : Copying over barbican ProxySQL rules config] ***********\nTuesday 24 February 2026  00:51:01 +0000 (0:00:01.323)       0:01:25.967 ****** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [include_role : blazar] ***************************************************\nTuesday 24 February 2026  00:51:02 +0000 (0:00:01.888)       0:01:27.856 ****** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [include_role : ceph-rgw] *************************************************\nTuesday 24 February 2026  00:51:03 +0000 (0:00:00.287)       0:01:28.143 ****** \n\u001b[0;36mincluded: ceph-rgw for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [haproxy-config : Copying over ceph-rgw haproxy config] *******************\nTuesday 24 February 2026  00:51:03 +0000 (0:00:00.706)       0:01:28.849 ****** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ceph-rgw', 'value': {'group': 'all', 'enabled': True, 'haproxy': {'radosgw': {'enabled': True, 'mode': 'http', 'external': False, 'port': '6780', 'custom_member_list': ['server testbed-node-3 192.168.16.13:7480 check inter 2000 rise 2 fall 5', 'server testbed-node-4 192.168.16.14:7480 check inter 2000 rise 2 fall 5', 'server testbed-node-5 192.168.16.15:7480 check inter 2000 rise 2 fall 5']}, 'radosgw_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6780', 'custom_member_list': ['server testbed-node-3 192.168.16.13:7480 check inter 2000 rise 2 fall 5', 'server testbed-node-4 192.168.16.14:7480 check inter 2000 rise 2 fall 5', 'server testbed-node-5 192.168.16.15:7480 check inter 2000 rise 2 fall 5']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ceph-rgw', 'value': {'group': 'all', 'enabled': True, 'haproxy': {'radosgw': {'enabled': True, 'mode': 'http', 'external': False, 'port': '6780', 'custom_member_list': ['server testbed-node-3 192.168.16.13:7480 check inter 2000 rise 2 fall 5', 'server testbed-node-4 192.168.16.14:7480 check inter 2000 rise 2 fall 5', 'server testbed-node-5 192.168.16.15:7480 check inter 2000 rise 2 fall 5']}, 'radosgw_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6780', 'custom_member_list': ['server testbed-node-3 192.168.16.13:7480 check inter 2000 rise 2 fall 5', 'server testbed-node-4 192.168.16.14:7480 check inter 2000 rise 2 fall 5', 'server testbed-node-5 192.168.16.15:7480 check inter 2000 rise 2 fall 5']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ceph-rgw', 'value': {'group': 'all', 'enabled': True, 'haproxy': {'radosgw': {'enabled': True, 'mode': 'http', 'external': False, 'port': '6780', 'custom_member_list': ['server testbed-node-3 192.168.16.13:7480 check inter 2000 rise 2 fall 5', 'server testbed-node-4 192.168.16.14:7480 check inter 2000 rise 2 fall 5', 'server testbed-node-5 192.168.16.15:7480 check inter 2000 rise 2 fall 5']}, 'radosgw_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6780', 'custom_member_list': ['server testbed-node-3 192.168.16.13:7480 check inter 2000 rise 2 fall 5', 'server testbed-node-4 192.168.16.14:7480 check inter 2000 rise 2 fall 5', 'server testbed-node-5 192.168.16.15:7480 check inter 2000 rise 2 fall 5']}}}})\u001b[0m\n\nTASK [haproxy-config : Add configuration for ceph-rgw when using single external frontend] ***\nTuesday 24 February 2026  00:51:06 +0000 (0:00:02.656)       0:01:31.506 ****** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'ceph-rgw', 'value': {'group': 'all', 'enabled': True, 'haproxy': {'radosgw': {'enabled': True, 'mode': 'http', 'external': False, 'port': '6780', 'custom_member_list': ['server testbed-node-3 192.168.16.13:7480 check inter 2000 rise 2 fall 5', 'server testbed-node-4 192.168.16.14:7480 check inter 2000 rise 2 fall 5', 'server testbed-node-5 192.168.16.15:7480 check inter 2000 rise 2 fall 5']}, 'radosgw_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6780', 'custom_member_list': ['server testbed-node-3 192.168.16.13:7480 check inter 2000 rise 2 fall 5', 'server testbed-node-4 192.168.16.14:7480 check inter 2000 rise 2 fall 5', 'server testbed-node-5 192.168.16.15:7480 check inter 2000 rise 2 fall 5']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'ceph-rgw', 'value': {'group': 'all', 'enabled': True, 'haproxy': {'radosgw': {'enabled': True, 'mode': 'http', 'external': False, 'port': '6780', 'custom_member_list': ['server testbed-node-3 192.168.16.13:7480 check inter 2000 rise 2 fall 5', 'server testbed-node-4 192.168.16.14:7480 check inter 2000 rise 2 fall 5', 'server testbed-node-5 192.168.16.15:7480 check inter 2000 rise 2 fall 5']}, 'radosgw_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6780', 'custom_member_list': ['server testbed-node-3 192.168.16.13:7480 check inter 2000 rise 2 fall 5', 'server testbed-node-4 192.168.16.14:7480 check inter 2000 rise 2 fall 5', 'server testbed-node-5 192.168.16.15:7480 check inter 2000 rise 2 fall 5']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'ceph-rgw', 'value': {'group': 'all', 'enabled': True, 'haproxy': {'radosgw': {'enabled': True, 'mode': 'http', 'external': False, 'port': '6780', 'custom_member_list': ['server testbed-node-3 192.168.16.13:7480 check inter 2000 rise 2 fall 5', 'server testbed-node-4 192.168.16.14:7480 check inter 2000 rise 2 fall 5', 'server testbed-node-5 192.168.16.15:7480 check inter 2000 rise 2 fall 5']}, 'radosgw_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6780', 'custom_member_list': ['server testbed-node-3 192.168.16.13:7480 check inter 2000 rise 2 fall 5', 'server testbed-node-4 192.168.16.14:7480 check inter 2000 rise 2 fall 5', 'server testbed-node-5 192.168.16.15:7480 check inter 2000 rise 2 fall 5']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [haproxy-config : Configuring firewall for ceph-rgw] **********************\nTuesday 24 February 2026  00:51:08 +0000 (0:00:01.986)       0:01:33.493 ****** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'radosgw', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '6780', 'custom_member_list': ['server testbed-node-3 192.168.16.13:7480 check inter 2000 rise 2 fall 5', 'server testbed-node-4 192.168.16.14:7480 check inter 2000 rise 2 fall 5', 'server testbed-node-5 192.168.16.15:7480 check inter 2000 rise 2 fall 5']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'radosgw_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6780', 'custom_member_list': ['server testbed-node-3 192.168.16.13:7480 check inter 2000 rise 2 fall 5', 'server testbed-node-4 192.168.16.14:7480 check inter 2000 rise 2 fall 5', 'server testbed-node-5 192.168.16.15:7480 check inter 2000 rise 2 fall 5']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'radosgw', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '6780', 'custom_member_list': ['server testbed-node-3 192.168.16.13:7480 check inter 2000 rise 2 fall 5', 'server testbed-node-4 192.168.16.14:7480 check inter 2000 rise 2 fall 5', 'server testbed-node-5 192.168.16.15:7480 check inter 2000 rise 2 fall 5']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'radosgw_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6780', 'custom_member_list': ['server testbed-node-3 192.168.16.13:7480 check inter 2000 rise 2 fall 5', 'server testbed-node-4 192.168.16.14:7480 check inter 2000 rise 2 fall 5', 'server testbed-node-5 192.168.16.15:7480 check inter 2000 rise 2 fall 5']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'radosgw', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '6780', 'custom_member_list': ['server testbed-node-3 192.168.16.13:7480 check inter 2000 rise 2 fall 5', 'server testbed-node-4 192.168.16.14:7480 check inter 2000 rise 2 fall 5', 'server testbed-node-5 192.168.16.15:7480 check inter 2000 rise 2 fall 5']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'radosgw_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6780', 'custom_member_list': ['server testbed-node-3 192.168.16.13:7480 check inter 2000 rise 2 fall 5', 'server testbed-node-4 192.168.16.14:7480 check inter 2000 rise 2 fall 5', 'server testbed-node-5 192.168.16.15:7480 check inter 2000 rise 2 fall 5']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [proxysql-config : Copying over ceph-rgw ProxySQL users config] ***********\nTuesday 24 February 2026  00:51:10 +0000 (0:00:01.615)       0:01:35.109 ****** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [proxysql-config : Copying over ceph-rgw ProxySQL rules config] ***********\nTuesday 24 February 2026  00:51:10 +0000 (0:00:00.375)       0:01:35.484 ****** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [include_role : cinder] ***************************************************\nTuesday 24 February 2026  00:51:11 +0000 (0:00:01.057)       0:01:36.542 ****** \n\u001b[0;36mincluded: cinder for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [haproxy-config : Copying over cinder haproxy config] *********************\nTuesday 24 February 2026  00:51:12 +0000 (0:00:00.664)       0:01:37.206 ****** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-api:2024.2', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8776'], 'timeout': '30'}, 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-scheduler:2024.2', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-volume:2024.2', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-api:2024.2', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8776'], 'timeout': '30'}, 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-backup:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-scheduler:2024.2', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-volume:2024.2', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-api:2024.2', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8776'], 'timeout': '30'}, 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-backup:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-scheduler:2024.2', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-volume:2024.2', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-backup:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}}) \u001b[0m\n\nTASK [haproxy-config : Add configuration for cinder when using single external frontend] ***\nTuesday 24 February 2026  00:51:15 +0000 (0:00:03.380)       0:01:40.587 ****** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-api:2024.2', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8776'], 'timeout': '30'}, 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-scheduler:2024.2', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-volume:2024.2', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-backup:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-api:2024.2', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8776'], 'timeout': '30'}, 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-scheduler:2024.2', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-volume:2024.2', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-backup:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-api:2024.2', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8776'], 'timeout': '30'}, 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-scheduler:2024.2', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-volume:2024.2', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-backup:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [haproxy-config : Configuring firewall for cinder] ************************\nTuesday 24 February 2026  00:51:16 +0000 (0:00:00.775)       0:01:41.363 ****** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'cinder_api', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'cinder_api_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'cinder_api', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'cinder_api_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'cinder_api', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'cinder_api_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [proxysql-config : Copying over cinder ProxySQL users config] *************\nTuesday 24 February 2026  00:51:17 +0000 (0:00:00.841)       0:01:42.204 ****** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [proxysql-config : Copying over cinder ProxySQL rules config] *************\nTuesday 24 February 2026  00:51:18 +0000 (0:00:01.245)       0:01:43.449 ****** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [include_role : cloudkitty] ***********************************************\nTuesday 24 February 2026  00:51:20 +0000 (0:00:01.903)       0:01:45.353 ****** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [include_role : cyborg] ***************************************************\nTuesday 24 February 2026  00:51:20 +0000 (0:00:00.513)       0:01:45.867 ****** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [include_role : designate] ************************************************\nTuesday 24 February 2026  00:51:21 +0000 (0:00:00.288)       0:01:46.155 ****** \n\u001b[0;36mincluded: designate for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [haproxy-config : Copying over designate haproxy config] ******************\nTuesday 24 February 2026  00:51:22 +0000 (0:00:00.792)       0:01:46.947 ****** \n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-api:2024.2', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-api:2024.2', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-backend-bind9:2024.2', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-backend-bind9:2024.2', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-central:2024.2', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-mdns:2024.2', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-producer:2024.2', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-central:2024.2', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-worker:2024.2', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-sink', 'value': {'container_name': 'designate_sink', 'group': 'designate-sink', 'enabled': False, 'image': 'registry.osism.tech/kolla/designate-sink:2024.2', 'volumes': ['/etc/kolla/designate-sink/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-sink 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-mdns:2024.2', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-producer:2024.2', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-worker:2024.2', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-sink', 'value': {'container_name': 'designate_sink', 'group': 'designate-sink', 'enabled': False, 'image': 'registry.osism.tech/kolla/designate-sink:2024.2', 'volumes': ['/etc/kolla/designate-sink/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-sink 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-api:2024.2', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-backend-bind9:2024.2', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-central:2024.2', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-mdns:2024.2', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-producer:2024.2', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-worker:2024.2', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-sink', 'value': {'container_name': 'designate_sink', 'group': 'designate-sink', 'enabled': False, 'image': 'registry.osism.tech/kolla/designate-sink:2024.2', 'volumes': ['/etc/kolla/designate-sink/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-sink 5672'], 'timeout': '30'}}}) \u001b[0m\n\nTASK [haproxy-config : Add configuration for designate when using single external frontend] ***\nTuesday 24 February 2026  00:51:26 +0000 (0:00:04.593)       0:01:51.541 ****** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-api:2024.2', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-backend-bind9:2024.2', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-central:2024.2', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-mdns:2024.2', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-producer:2024.2', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-worker:2024.2', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-sink', 'value': {'container_name': 'designate_sink', 'group': 'designate-sink', 'enabled': False, 'image': 'registry.osism.tech/kolla/designate-sink:2024.2', 'volumes': ['/etc/kolla/designate-sink/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-sink 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-api:2024.2', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-backend-bind9:2024.2', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-central:2024.2', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-mdns:2024.2', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-producer:2024.2', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-worker:2024.2', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-sink', 'value': {'container_name': 'designate_sink', 'group': 'designate-sink', 'enabled': False, 'image': 'registry.osism.tech/kolla/designate-sink:2024.2', 'volumes': ['/etc/kolla/designate-sink/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-sink 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-api:2024.2', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-backend-bind9:2024.2', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-central:2024.2', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-mdns:2024.2', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-producer:2024.2', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-worker:2024.2', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-sink', 'value': {'container_name': 'designate_sink', 'group': 'designate-sink', 'enabled': False, 'image': 'registry.osism.tech/kolla/designate-sink:2024.2', 'volumes': ['/etc/kolla/designate-sink/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-sink 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [haproxy-config : Configuring firewall for designate] *********************\nTuesday 24 February 2026  00:51:27 +0000 (0:00:01.064)       0:01:52.606 ****** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate_api', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate_api_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate_api', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate_api_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate_api', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate_api_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [proxysql-config : Copying over designate ProxySQL users config] **********\nTuesday 24 February 2026  00:51:28 +0000 (0:00:00.950)       0:01:53.556 ****** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [proxysql-config : Copying over designate ProxySQL rules config] **********\nTuesday 24 February 2026  00:51:30 +0000 (0:00:01.571)       0:01:55.128 ****** \n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [include_role : etcd] *****************************************************\nTuesday 24 February 2026  00:51:32 +0000 (0:00:01.903)       0:01:57.032 ****** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [include_role : glance] ***************************************************\nTuesday 24 February 2026  00:51:32 +0000 (0:00:00.290)       0:01:57.323 ****** \n\u001b[0;36mincluded: glance for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [haproxy-config : Copying over glance haproxy config] *********************\nTuesday 24 February 2026  00:51:33 +0000 (0:00:00.799)       0:01:58.122 ****** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'glance-api', 'value': {'container_name': 'glance_api', 'group': 'glance-api', 'host_in_groups': True, 'enabled': True, 'image': 'registry.osism.tech/kolla/glance-api:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.10,192.168.16.9'}, 'privileged': True, 'volumes': ['/etc/kolla/glance-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'glance:/var/lib/glance/', '', 'kolla_logs:/var/log/kolla/', '', '', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9292'], 'timeout': '30'}, 'haproxy': {'glance_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}, 'glance_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'glance-tls-proxy', 'value': {'container_name': 'glance_tls_proxy', 'group': 'glance-api', 'host_in_groups': True, 'enabled': 'no', 'image': 'registry.osism.tech/kolla/glance-tls-proxy:2024.2', 'volumes': ['/etc/kolla/glance-tls-proxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl -u openstack:password 192.168.16.10:9293'], 'timeout': '30'}, 'haproxy': {'glance_tls_proxy': {'enabled': False, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', ''], 'tls_backend': 'yes'}, 'glance_tls_proxy_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', ''], 'tls_backend': 'yes'}}}}) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'glance-api', 'value': {'container_name': 'glance_api', 'group': 'glance-api', 'host_in_groups': True, 'enabled': True, 'image': 'registry.osism.tech/kolla/glance-api:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.12,192.168.16.9'}, 'privileged': True, 'volumes': ['/etc/kolla/glance-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'glance:/var/lib/glance/', '', 'kolla_logs:/var/log/kolla/', '', '', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9292'], 'timeout': '30'}, 'haproxy': {'glance_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}, 'glance_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'glance-tls-proxy', 'value': {'container_name': 'glance_tls_proxy', 'group': 'glance-api', 'host_in_groups': True, 'enabled': 'no', 'image': 'registry.osism.tech/kolla/glance-tls-proxy:2024.2', 'volumes': ['/etc/kolla/glance-tls-proxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl -u openstack:password 192.168.16.12:9293'], 'timeout': '30'}, 'haproxy': {'glance_tls_proxy': {'enabled': False, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', ''], 'tls_backend': 'yes'}, 'glance_tls_proxy_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', ''], 'tls_backend': 'yes'}}}}) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'glance-api', 'value': {'container_name': 'glance_api', 'group': 'glance-api', 'host_in_groups': True, 'enabled': True, 'image': 'registry.osism.tech/kolla/glance-api:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.11,192.168.16.9'}, 'privileged': True, 'volumes': ['/etc/kolla/glance-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'glance:/var/lib/glance/', '', 'kolla_logs:/var/log/kolla/', '', '', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9292'], 'timeout': '30'}, 'haproxy': {'glance_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}, 'glance_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'glance-tls-proxy', 'value': {'container_name': 'glance_tls_proxy', 'group': 'glance-api', 'host_in_groups': True, 'enabled': 'no', 'image': 'registry.osism.tech/kolla/glance-tls-proxy:2024.2', 'volumes': ['/etc/kolla/glance-tls-proxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl -u openstack:password 192.168.16.11:9293'], 'timeout': '30'}, 'haproxy': {'glance_tls_proxy': {'enabled': False, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', ''], 'tls_backend': 'yes'}, 'glance_tls_proxy_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', ''], 'tls_backend': 'yes'}}}}) \u001b[0m\n\nTASK [haproxy-config : Add configuration for glance when using single external frontend] ***\nTuesday 24 February 2026  00:51:37 +0000 (0:00:04.542)       0:02:02.664 ****** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'glance-api', 'value': {'container_name': 'glance_api', 'group': 'glance-api', 'host_in_groups': True, 'enabled': True, 'image': 'registry.osism.tech/kolla/glance-api:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.10,192.168.16.9'}, 'privileged': True, 'volumes': ['/etc/kolla/glance-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'glance:/var/lib/glance/', '', 'kolla_logs:/var/log/kolla/', '', '', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9292'], 'timeout': '30'}, 'haproxy': {'glance_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}, 'glance_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'glance-tls-proxy', 'value': {'container_name': 'glance_tls_proxy', 'group': 'glance-api', 'host_in_groups': True, 'enabled': 'no', 'image': 'registry.osism.tech/kolla/glance-tls-proxy:2024.2', 'volumes': ['/etc/kolla/glance-tls-proxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl -u openstack:password 192.168.16.10:9293'], 'timeout': '30'}, 'haproxy': {'glance_tls_proxy': {'enabled': False, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', ''], 'tls_backend': 'yes'}, 'glance_tls_proxy_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', ''], 'tls_backend': 'yes'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'glance-api', 'value': {'container_name': 'glance_api', 'group': 'glance-api', 'host_in_groups': True, 'enabled': True, 'image': 'registry.osism.tech/kolla/glance-api:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.12,192.168.16.9'}, 'privileged': True, 'volumes': ['/etc/kolla/glance-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'glance:/var/lib/glance/', '', 'kolla_logs:/var/log/kolla/', '', '', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9292'], 'timeout': '30'}, 'haproxy': {'glance_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}, 'glance_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'glance-tls-proxy', 'value': {'container_name': 'glance_tls_proxy', 'group': 'glance-api', 'host_in_groups': True, 'enabled': 'no', 'image': 'registry.osism.tech/kolla/glance-tls-proxy:2024.2', 'volumes': ['/etc/kolla/glance-tls-proxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl -u openstack:password 192.168.16.12:9293'], 'timeout': '30'}, 'haproxy': {'glance_tls_proxy': {'enabled': False, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', ''], 'tls_backend': 'yes'}, 'glance_tls_proxy_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', ''], 'tls_backend': 'yes'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'glance-api', 'value': {'container_name': 'glance_api', 'group': 'glance-api', 'host_in_groups': True, 'enabled': True, 'image': 'registry.osism.tech/kolla/glance-api:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.11,192.168.16.9'}, 'privileged': True, 'volumes': ['/etc/kolla/glance-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'glance:/var/lib/glance/', '', 'kolla_logs:/var/log/kolla/', '', '', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9292'], 'timeout': '30'}, 'haproxy': {'glance_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}, 'glance_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'glance-tls-proxy', 'value': {'container_name': 'glance_tls_proxy', 'group': 'glance-api', 'host_in_groups': True, 'enabled': 'no', 'image': 'registry.osism.tech/kolla/glance-tls-proxy:2024.2', 'volumes': ['/etc/kolla/glance-tls-proxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl -u openstack:password 192.168.16.11:9293'], 'timeout': '30'}, 'haproxy': {'glance_tls_proxy': {'enabled': False, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', ''], 'tls_backend': 'yes'}, 'glance_tls_proxy_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', ''], 'tls_backend': 'yes'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [haproxy-config : Configuring firewall for glance] ************************\nTuesday 24 February 2026  00:51:40 +0000 (0:00:02.663)       0:02:05.327 ****** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'glance_api', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'glance_api_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'glance_api', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'glance_api_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'glance_api', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'glance_api_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [proxysql-config : Copying over glance ProxySQL users config] *************\nTuesday 24 February 2026  00:51:43 +0000 (0:00:02.677)       0:02:08.005 ****** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [proxysql-config : Copying over glance ProxySQL rules config] *************\nTuesday 24 February 2026  00:51:44 +0000 (0:00:01.305)       0:02:09.311 ****** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [include_role : gnocchi] **************************************************\nTuesday 24 February 2026  00:51:46 +0000 (0:00:01.919)       0:02:11.230 ****** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [include_role : grafana] **************************************************\nTuesday 24 February 2026  00:51:46 +0000 (0:00:00.431)       0:02:11.661 ****** \n\u001b[0;36mincluded: grafana for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [haproxy-config : Copying over grafana haproxy config] ********************\nTuesday 24 February 2026  00:51:47 +0000 (0:00:00.834)       0:02:12.496 ****** \n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'grafana', 'value': {'container_name': 'grafana', 'group': 'grafana', 'enabled': True, 'image': 'registry.osism.tech/kolla/grafana:2024.2', 'volumes': ['/etc/kolla/grafana/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'grafana_server': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000'}, 'grafana_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'grafana', 'value': {'container_name': 'grafana', 'group': 'grafana', 'enabled': True, 'image': 'registry.osism.tech/kolla/grafana:2024.2', 'volumes': ['/etc/kolla/grafana/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'grafana_server': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000'}, 'grafana_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'grafana', 'value': {'container_name': 'grafana', 'group': 'grafana', 'enabled': True, 'image': 'registry.osism.tech/kolla/grafana:2024.2', 'volumes': ['/etc/kolla/grafana/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'grafana_server': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000'}, 'grafana_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000'}}}})\u001b[0m\n\nTASK [haproxy-config : Add configuration for grafana when using single external frontend] ***\nTuesday 24 February 2026  00:51:51 +0000 (0:00:04.209)       0:02:16.705 ****** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'grafana', 'value': {'container_name': 'grafana', 'group': 'grafana', 'enabled': True, 'image': 'registry.osism.tech/kolla/grafana:2024.2', 'volumes': ['/etc/kolla/grafana/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'grafana_server': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000'}, 'grafana_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'grafana', 'value': {'container_name': 'grafana', 'group': 'grafana', 'enabled': True, 'image': 'registry.osism.tech/kolla/grafana:2024.2', 'volumes': ['/etc/kolla/grafana/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'grafana_server': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000'}, 'grafana_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'grafana', 'value': {'container_name': 'grafana', 'group': 'grafana', 'enabled': True, 'image': 'registry.osism.tech/kolla/grafana:2024.2', 'volumes': ['/etc/kolla/grafana/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'grafana_server': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000'}, 'grafana_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [haproxy-config : Configuring firewall for grafana] ***********************\nTuesday 24 February 2026  00:51:52 +0000 (0:00:00.588)       0:02:17.294 ****** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'grafana_server', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'grafana_server_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'grafana_server', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'grafana_server_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'grafana_server', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'grafana_server_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [proxysql-config : Copying over grafana ProxySQL users config] ************\nTuesday 24 February 2026  00:51:52 +0000 (0:00:00.578)       0:02:17.873 ****** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [proxysql-config : Copying over grafana ProxySQL rules config] ************\nTuesday 24 February 2026  00:51:54 +0000 (0:00:01.509)       0:02:19.382 ****** \n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [include_role : heat] *****************************************************\nTuesday 24 February 2026  00:51:56 +0000 (0:00:01.975)       0:02:21.358 ****** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [include_role : horizon] **************************************************\nTuesday 24 February 2026  00:51:56 +0000 (0:00:00.511)       0:02:21.869 ****** \n\u001b[0;36mincluded: horizon for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [haproxy-config : Copying over horizon haproxy config] ********************\nTuesday 24 February 2026  00:51:57 +0000 (0:00:00.886)       0:02:22.756 ****** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'horizon', 'value': {'container_name': 'horizon', 'group': 'horizon', 'enabled': True, 'image': 'registry.osism.tech/kolla/horizon:2024.2', 'environment': {'ENABLE_BLAZAR': 'no', 'ENABLE_CLOUDKITTY': 'no', 'ENABLE_DESIGNATE': 'yes', 'ENABLE_FWAAS': 'no', 'ENABLE_HEAT': 'no', 'ENABLE_IRONIC': 'no', 'ENABLE_MAGNUM': 'yes', 'ENABLE_MANILA': 'yes', 'ENABLE_MASAKARI': 'no', 'ENABLE_MISTRAL': 'no', 'ENABLE_NEUTRON_VPNAAS': 'no', 'ENABLE_OCTAVIA': 'yes', 'ENABLE_TACKER': 'no', 'ENABLE_TROVE': 'no', 'ENABLE_WATCHER': 'no', 'ENABLE_ZUN': 'no', 'FORCE_GENERATE': 'no'}, 'volumes': ['/etc/kolla/horizon/:/var/lib/kolla/config_files/:ro', '', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:80'], 'timeout': '30'}, 'haproxy': {'horizon': {'enabled': True, 'mode': 'http', 'external': False, 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_redirect': {'enabled': True, 'mode': 'redirect', 'external': False, 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'horizon_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_external_redirect': {'enabled': True, 'mode': 'redirect', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'acme_client': {'enabled': True, 'with_frontend': False, 'custom_member_list': []}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'horizon', 'value': {'container_name': 'horizon', 'group': 'horizon', 'enabled': True, 'image': 'registry.osism.tech/kolla/horizon:2024.2', 'environment': {'ENABLE_BLAZAR': 'no', 'ENABLE_CLOUDKITTY': 'no', 'ENABLE_DESIGNATE': 'yes', 'ENABLE_FWAAS': 'no', 'ENABLE_HEAT': 'no', 'ENABLE_IRONIC': 'no', 'ENABLE_MAGNUM': 'yes', 'ENABLE_MANILA': 'yes', 'ENABLE_MASAKARI': 'no', 'ENABLE_MISTRAL': 'no', 'ENABLE_NEUTRON_VPNAAS': 'no', 'ENABLE_OCTAVIA': 'yes', 'ENABLE_TACKER': 'no', 'ENABLE_TROVE': 'no', 'ENABLE_WATCHER': 'no', 'ENABLE_ZUN': 'no', 'FORCE_GENERATE': 'no'}, 'volumes': ['/etc/kolla/horizon/:/var/lib/kolla/config_files/:ro', '', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:80'], 'timeout': '30'}, 'haproxy': {'horizon': {'enabled': True, 'mode': 'http', 'external': False, 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_redirect': {'enabled': True, 'mode': 'redirect', 'external': False, 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'horizon_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_external_redirect': {'enabled': True, 'mode': 'redirect', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'acme_client': {'enabled': True, 'with_frontend': False, 'custom_member_list': []}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'horizon', 'value': {'container_name': 'horizon', 'group': 'horizon', 'enabled': True, 'image': 'registry.osism.tech/kolla/horizon:2024.2', 'environment': {'ENABLE_BLAZAR': 'no', 'ENABLE_CLOUDKITTY': 'no', 'ENABLE_DESIGNATE': 'yes', 'ENABLE_FWAAS': 'no', 'ENABLE_HEAT': 'no', 'ENABLE_IRONIC': 'no', 'ENABLE_MAGNUM': 'yes', 'ENABLE_MANILA': 'yes', 'ENABLE_MASAKARI': 'no', 'ENABLE_MISTRAL': 'no', 'ENABLE_NEUTRON_VPNAAS': 'no', 'ENABLE_OCTAVIA': 'yes', 'ENABLE_TACKER': 'no', 'ENABLE_TROVE': 'no', 'ENABLE_WATCHER': 'no', 'ENABLE_ZUN': 'no', 'FORCE_GENERATE': 'no'}, 'volumes': ['/etc/kolla/horizon/:/var/lib/kolla/config_files/:ro', '', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:80'], 'timeout': '30'}, 'haproxy': {'horizon': {'enabled': True, 'mode': 'http', 'external': False, 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_redirect': {'enabled': True, 'mode': 'redirect', 'external': False, 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'horizon_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_external_redirect': {'enabled': True, 'mode': 'redirect', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'acme_client': {'enabled': True, 'with_frontend': False, 'custom_member_list': []}}}})\u001b[0m\n\nTASK [haproxy-config : Add configuration for horizon when using single external frontend] ***\nTuesday 24 February 2026  00:52:01 +0000 (0:00:04.063)       0:02:26.820 ****** \n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'horizon', 'value': {'container_name': 'horizon', 'group': 'horizon', 'enabled': True, 'image': 'registry.osism.tech/kolla/horizon:2024.2', 'environment': {'ENABLE_BLAZAR': 'no', 'ENABLE_CLOUDKITTY': 'no', 'ENABLE_DESIGNATE': 'yes', 'ENABLE_FWAAS': 'no', 'ENABLE_HEAT': 'no', 'ENABLE_IRONIC': 'no', 'ENABLE_MAGNUM': 'yes', 'ENABLE_MANILA': 'yes', 'ENABLE_MASAKARI': 'no', 'ENABLE_MISTRAL': 'no', 'ENABLE_NEUTRON_VPNAAS': 'no', 'ENABLE_OCTAVIA': 'yes', 'ENABLE_TACKER': 'no', 'ENABLE_TROVE': 'no', 'ENABLE_WATCHER': 'no', 'ENABLE_ZUN': 'no', 'FORCE_GENERATE': 'no'}, 'volumes': ['/etc/kolla/horizon/:/var/lib/kolla/config_files/:ro', '', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:80'], 'timeout': '30'}, 'haproxy': {'horizon': {'enabled': True, 'mode': 'http', 'external': False, 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_redirect': {'enabled': True, 'mode': 'redirect', 'external': False, 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'horizon_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_external_redirect': {'enabled': True, 'mode': 'redirect', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'acme_client': {'enabled': True, 'with_frontend': False, 'custom_member_list': []}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'horizon', 'value': {'container_name': 'horizon', 'group': 'horizon', 'enabled': True, 'image': 'registry.osism.tech/kolla/horizon:2024.2', 'environment': {'ENABLE_BLAZAR': 'no', 'ENABLE_CLOUDKITTY': 'no', 'ENABLE_DESIGNATE': 'yes', 'ENABLE_FWAAS': 'no', 'ENABLE_HEAT': 'no', 'ENABLE_IRONIC': 'no', 'ENABLE_MAGNUM': 'yes', 'ENABLE_MANILA': 'yes', 'ENABLE_MASAKARI': 'no', 'ENABLE_MISTRAL': 'no', 'ENABLE_NEUTRON_VPNAAS': 'no', 'ENABLE_OCTAVIA': 'yes', 'ENABLE_TACKER': 'no', 'ENABLE_TROVE': 'no', 'ENABLE_WATCHER': 'no', 'ENABLE_ZUN': 'no', 'FORCE_GENERATE': 'no'}, 'volumes': ['/etc/kolla/horizon/:/var/lib/kolla/config_files/:ro', '', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:80'], 'timeout': '30'}, 'haproxy': {'horizon': {'enabled': True, 'mode': 'http', 'external': False, 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_redirect': {'enabled': True, 'mode': 'redirect', 'external': False, 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'horizon_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_external_redirect': {'enabled': True, 'mode': 'redirect', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'acme_client': {'enabled': True, 'with_frontend': False, 'custom_member_list': []}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'horizon', 'value': {'container_name': 'horizon', 'group': 'horizon', 'enabled': True, 'image': 'registry.osism.tech/kolla/horizon:2024.2', 'environment': {'ENABLE_BLAZAR': 'no', 'ENABLE_CLOUDKITTY': 'no', 'ENABLE_DESIGNATE': 'yes', 'ENABLE_FWAAS': 'no', 'ENABLE_HEAT': 'no', 'ENABLE_IRONIC': 'no', 'ENABLE_MAGNUM': 'yes', 'ENABLE_MANILA': 'yes', 'ENABLE_MASAKARI': 'no', 'ENABLE_MISTRAL': 'no', 'ENABLE_NEUTRON_VPNAAS': 'no', 'ENABLE_OCTAVIA': 'yes', 'ENABLE_TACKER': 'no', 'ENABLE_TROVE': 'no', 'ENABLE_WATCHER': 'no', 'ENABLE_ZUN': 'no', 'FORCE_GENERATE': 'no'}, 'volumes': ['/etc/kolla/horizon/:/var/lib/kolla/config_files/:ro', '', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:80'], 'timeout': '30'}, 'haproxy': {'horizon': {'enabled': True, 'mode': 'http', 'external': False, 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_redirect': {'enabled': True, 'mode': 'redirect', 'external': False, 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'horizon_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_external_redirect': {'enabled': True, 'mode': 'redirect', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'acme_client': {'enabled': True, 'with_frontend': False, 'custom_member_list': []}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [haproxy-config : Configuring firewall for horizon] ***********************\nTuesday 24 February 2026  00:52:03 +0000 (0:00:01.431)       0:02:28.251 ****** \n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'horizon', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'horizon_redirect', 'value': {'enabled': True, 'mode': 'redirect', 'external': False, 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'horizon_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'horizon_external_redirect', 'value': {'enabled': True, 'mode': 'redirect', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'acme_client', 'value': {'enabled': True, 'with_frontend': False, 'custom_member_list': []}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'horizon', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'horizon_redirect', 'value': {'enabled': True, 'mode': 'redirect', 'external': False, 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'horizon_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'horizon_external_redirect', 'value': {'enabled': True, 'mode': 'redirect', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'acme_client', 'value': {'enabled': True, 'with_frontend': False, 'custom_member_list': []}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'horizon', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'horizon_redirect', 'value': {'enabled': True, 'mode': 'redirect', 'external': False, 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'horizon_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'horizon_external_redirect', 'value': {'enabled': True, 'mode': 'redirect', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'acme_client', 'value': {'enabled': True, 'with_frontend': False, 'custom_member_list': []}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [proxysql-config : Copying over horizon ProxySQL users config] ************\nTuesday 24 February 2026  00:52:05 +0000 (0:00:01.807)       0:02:30.059 ****** \n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\nTASK [proxysql-config : Copying over horizon ProxySQL rules config] ************\nTuesday 24 February 2026  00:52:06 +0000 (0:00:01.524)       0:02:31.583 ****** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [include_role : influxdb] *************************************************\nTuesday 24 February 2026  00:52:09 +0000 (0:00:02.471)       0:02:34.054 ****** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [include_role : ironic] ***************************************************\nTuesday 24 February 2026  00:52:09 +0000 (0:00:00.343)       0:02:34.398 ****** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [include_role : keystone] *************************************************\nTuesday 24 February 2026  00:52:09 +0000 (0:00:00.478)       0:02:34.876 ****** \n\u001b[0;36mincluded: keystone for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [haproxy-config : Copying over keystone haproxy config] *******************\nTuesday 24 February 2026  00:52:10 +0000 (0:00:00.792)       0:02:35.669 ****** \n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone:2024.2', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-ssh:2024.2', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-fernet:2024.2', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone:2024.2', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-ssh:2024.2', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-fernet:2024.2', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone:2024.2', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-ssh:2024.2', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-fernet:2024.2', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}}) \u001b[0m\n\nTASK [haproxy-config : Add configuration for keystone when using single external frontend] ***\nTuesday 24 February 2026  00:52:15 +0000 (0:00:04.897)       0:02:40.566 ****** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone:2024.2', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-ssh:2024.2', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-fernet:2024.2', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone:2024.2', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-ssh:2024.2', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-fernet:2024.2', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone:2024.2', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-ssh:2024.2', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-fernet:2024.2', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [haproxy-config : Configuring firewall for keystone] **********************\nTuesday 24 February 2026  00:52:16 +0000 (0:00:01.083)       0:02:41.650 ****** \n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'keystone_internal', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'keystone_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'keystone_internal', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'keystone_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'keystone_internal', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'keystone_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [proxysql-config : Copying over keystone ProxySQL users config] ***********\nTuesday 24 February 2026  00:52:17 +0000 (0:00:01.044)       0:02:42.695 ****** \n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\nTASK [proxysql-config : Copying over keystone ProxySQL rules config] ***********\nTuesday 24 February 2026  00:52:19 +0000 (0:00:01.621)       0:02:44.316 ****** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [include_role : letsencrypt] **********************************************\nTuesday 24 February 2026  00:52:21 +0000 (0:00:01.915)       0:02:46.232 ****** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [include_role : magnum] ***************************************************\nTuesday 24 February 2026  00:52:21 +0000 (0:00:00.510)       0:02:46.743 ****** \n\u001b[0;36mincluded: magnum for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [haproxy-config : Copying over magnum haproxy config] *********************\nTuesday 24 February 2026  00:52:22 +0000 (0:00:00.914)       0:02:47.658 ****** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-api:2024.2', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-conductor:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.10,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-api:2024.2', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-conductor:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.11,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-api:2024.2', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-conductor:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.12,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}}) \u001b[0m\n\nTASK [haproxy-config : Add configuration for magnum when using single external frontend] ***\nTuesday 24 February 2026  00:52:26 +0000 (0:00:03.738)       0:02:51.396 ****** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-api:2024.2', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-conductor:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.10,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-api:2024.2', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-conductor:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.11,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-api:2024.2', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-conductor:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.12,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [haproxy-config : Configuring firewall for magnum] ************************\nTuesday 24 February 2026  00:52:27 +0000 (0:00:00.740)       0:02:52.137 ****** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'magnum_api', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'magnum_api_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'magnum_api', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'magnum_api_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'magnum_api', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'magnum_api_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [proxysql-config : Copying over magnum ProxySQL users config] *************\nTuesday 24 February 2026  00:52:28 +0000 (0:00:00.769)       0:02:52.907 ****** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [proxysql-config : Copying over magnum ProxySQL rules config] *************\nTuesday 24 February 2026  00:52:29 +0000 (0:00:01.333)       0:02:54.241 ****** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [include_role : manila] ***************************************************\nTuesday 24 February 2026  00:52:31 +0000 (0:00:02.157)       0:02:56.398 ****** \n\u001b[0;36mincluded: manila for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [haproxy-config : Copying over manila haproxy config] *********************\nTuesday 24 February 2026  00:52:32 +0000 (0:00:01.396)       0:02:57.794 ****** \n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'manila-api', 'value': {'container_name': 'manila_api', 'group': 'manila-api', 'image': 'registry.osism.tech/kolla/manila-api:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/manila-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8786'], 'timeout': '30'}, 'haproxy': {'manila_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8786', 'listen_port': '8786'}, 'manila_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8786', 'listen_port': '8786'}}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'manila-scheduler', 'value': {'container_name': 'manila_scheduler', 'group': 'manila-scheduler', 'image': 'registry.osism.tech/kolla/manila-scheduler:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/manila-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'manila-share', 'value': {'container_name': 'manila_share', 'group': 'manila-share', 'image': 'registry.osism.tech/kolla/manila-share:2024.2', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/manila-share/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run:/run:shared', 'kolla_logs:/var/log/kolla/', '', '/lib/modules:/lib/modules:ro', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-share 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'manila-data', 'value': {'container_name': 'manila_data', 'group': 'manila-data', 'image': 'registry.osism.tech/kolla/manila-data:2024.2', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/manila-data/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run:/run:shared', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-data 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'manila-api', 'value': {'container_name': 'manila_api', 'group': 'manila-api', 'image': 'registry.osism.tech/kolla/manila-api:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/manila-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8786'], 'timeout': '30'}, 'haproxy': {'manila_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8786', 'listen_port': '8786'}, 'manila_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8786', 'listen_port': '8786'}}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'manila-scheduler', 'value': {'container_name': 'manila_scheduler', 'group': 'manila-scheduler', 'image': 'registry.osism.tech/kolla/manila-scheduler:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/manila-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'manila-share', 'value': {'container_name': 'manila_share', 'group': 'manila-share', 'image': 'registry.osism.tech/kolla/manila-share:2024.2', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/manila-share/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run:/run:shared', 'kolla_logs:/var/log/kolla/', '', '/lib/modules:/lib/modules:ro', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-share 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'manila-api', 'value': {'container_name': 'manila_api', 'group': 'manila-api', 'image': 'registry.osism.tech/kolla/manila-api:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/manila-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8786'], 'timeout': '30'}, 'haproxy': {'manila_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8786', 'listen_port': '8786'}, 'manila_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8786', 'listen_port': '8786'}}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'manila-data', 'value': {'container_name': 'manila_data', 'group': 'manila-data', 'image': 'registry.osism.tech/kolla/manila-data:2024.2', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/manila-data/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run:/run:shared', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-data 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'manila-scheduler', 'value': {'container_name': 'manila_scheduler', 'group': 'manila-scheduler', 'image': 'registry.osism.tech/kolla/manila-scheduler:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/manila-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'manila-share', 'value': {'container_name': 'manila_share', 'group': 'manila-share', 'image': 'registry.osism.tech/kolla/manila-share:2024.2', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/manila-share/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run:/run:shared', 'kolla_logs:/var/log/kolla/', '', '/lib/modules:/lib/modules:ro', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-share 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'manila-data', 'value': {'container_name': 'manila_data', 'group': 'manila-data', 'image': 'registry.osism.tech/kolla/manila-data:2024.2', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/manila-data/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run:/run:shared', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-data 5672'], 'timeout': '30'}}}) \u001b[0m\n\nTASK [haproxy-config : Add configuration for manila when using single external frontend] ***\nTuesday 24 February 2026  00:52:37 +0000 (0:00:04.339)       0:03:02.134 ****** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'manila-api', 'value': {'container_name': 'manila_api', 'group': 'manila-api', 'image': 'registry.osism.tech/kolla/manila-api:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/manila-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8786'], 'timeout': '30'}, 'haproxy': {'manila_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8786', 'listen_port': '8786'}, 'manila_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8786', 'listen_port': '8786'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'manila-scheduler', 'value': {'container_name': 'manila_scheduler', 'group': 'manila-scheduler', 'image': 'registry.osism.tech/kolla/manila-scheduler:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/manila-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'manila-share', 'value': {'container_name': 'manila_share', 'group': 'manila-share', 'image': 'registry.osism.tech/kolla/manila-share:2024.2', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/manila-share/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run:/run:shared', 'kolla_logs:/var/log/kolla/', '', '/lib/modules:/lib/modules:ro', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-share 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'manila-data', 'value': {'container_name': 'manila_data', 'group': 'manila-data', 'image': 'registry.osism.tech/kolla/manila-data:2024.2', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/manila-data/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run:/run:shared', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-data 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'manila-api', 'value': {'container_name': 'manila_api', 'group': 'manila-api', 'image': 'registry.osism.tech/kolla/manila-api:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/manila-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8786'], 'timeout': '30'}, 'haproxy': {'manila_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8786', 'listen_port': '8786'}, 'manila_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8786', 'listen_port': '8786'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'manila-scheduler', 'value': {'container_name': 'manila_scheduler', 'group': 'manila-scheduler', 'image': 'registry.osism.tech/kolla/manila-scheduler:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/manila-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'manila-share', 'value': {'container_name': 'manila_share', 'group': 'manila-share', 'image': 'registry.osism.tech/kolla/manila-share:2024.2', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/manila-share/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run:/run:shared', 'kolla_logs:/var/log/kolla/', '', '/lib/modules:/lib/modules:ro', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-share 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'manila-data', 'value': {'container_name': 'manila_data', 'group': 'manila-data', 'image': 'registry.osism.tech/kolla/manila-data:2024.2', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/manila-data/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run:/run:shared', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-data 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'manila-api', 'value': {'container_name': 'manila_api', 'group': 'manila-api', 'image': 'registry.osism.tech/kolla/manila-api:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/manila-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8786'], 'timeout': '30'}, 'haproxy': {'manila_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8786', 'listen_port': '8786'}, 'manila_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8786', 'listen_port': '8786'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'manila-scheduler', 'value': {'container_name': 'manila_scheduler', 'group': 'manila-scheduler', 'image': 'registry.osism.tech/kolla/manila-scheduler:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/manila-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'manila-share', 'value': {'container_name': 'manila_share', 'group': 'manila-share', 'image': 'registry.osism.tech/kolla/manila-share:2024.2', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/manila-share/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run:/run:shared', 'kolla_logs:/var/log/kolla/', '', '/lib/modules:/lib/modules:ro', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-share 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'manila-data', 'value': {'container_name': 'manila_data', 'group': 'manila-data', 'image': 'registry.osism.tech/kolla/manila-data:2024.2', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/manila-data/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run:/run:shared', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-data 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [haproxy-config : Configuring firewall for manila] ************************\nTuesday 24 February 2026  00:52:37 +0000 (0:00:00.727)       0:03:02.861 ****** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'manila_api', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8786', 'listen_port': '8786'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'manila_api_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8786', 'listen_port': '8786'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'manila_api', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8786', 'listen_port': '8786'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'manila_api_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8786', 'listen_port': '8786'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'manila_api', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8786', 'listen_port': '8786'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'manila_api_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8786', 'listen_port': '8786'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [proxysql-config : Copying over manila ProxySQL users config] *************\nTuesday 24 February 2026  00:52:39 +0000 (0:00:01.397)       0:03:04.259 ****** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [proxysql-config : Copying over manila ProxySQL rules config] *************\nTuesday 24 February 2026  00:52:40 +0000 (0:00:01.385)       0:03:05.645 ****** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [include_role : mariadb] **************************************************\nTuesday 24 February 2026  00:52:42 +0000 (0:00:01.897)       0:03:07.542 ****** \n\u001b[0;36mincluded: mariadb for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [mariadb : Ensure mysql monitor user exist] *******************************\nTuesday 24 February 2026  00:52:43 +0000 (0:00:01.170)       0:03:08.713 ****** \n\u001b[0;32mok: [testbed-node-0] => (item=testbed-node-0)\u001b[0m\n\nTASK [haproxy-config : Copying over mariadb haproxy config] ********************\nTuesday 24 February 2026  00:52:46 +0000 (0:00:03.129)       0:03:11.843 ****** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/mariadb-server:2024.2', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.10', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'mariadb-clustercheck', 'value': {'container_name': 'mariadb_clustercheck', 'group': 'mariadb_shard_0', 'enabled': False, 'image': 'registry.osism.tech/kolla/mariadb-clustercheck:2024.2', 'volumes': ['/etc/kolla/mariadb-clustercheck/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.10', 'AVAILABLE_WHEN_DONOR': '1'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/mariadb-server:2024.2', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.11', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/mariadb-server:2024.2', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.12', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'mariadb-clustercheck', 'value': {'container_name': 'mariadb_clustercheck', 'group': 'mariadb_shard_0', 'enabled': False, 'image': 'registry.osism.tech/kolla/mariadb-clustercheck:2024.2', 'volumes': ['/etc/kolla/mariadb-clustercheck/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.11', 'AVAILABLE_WHEN_DONOR': '1'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'mariadb-clustercheck', 'value': {'container_name': 'mariadb_clustercheck', 'group': 'mariadb_shard_0', 'enabled': False, 'image': 'registry.osism.tech/kolla/mariadb-clustercheck:2024.2', 'volumes': ['/etc/kolla/mariadb-clustercheck/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.12', 'AVAILABLE_WHEN_DONOR': '1'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [haproxy-config : Add configuration for mariadb when using single external frontend] ***\nTuesday 24 February 2026  00:52:49 +0000 (0:00:02.346)       0:03:14.190 ****** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/mariadb-server:2024.2', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.10', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'mariadb-clustercheck', 'value': {'container_name': 'mariadb_clustercheck', 'group': 'mariadb_shard_0', 'enabled': False, 'image': 'registry.osism.tech/kolla/mariadb-clustercheck:2024.2', 'volumes': ['/etc/kolla/mariadb-clustercheck/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.10', 'AVAILABLE_WHEN_DONOR': '1'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/mariadb-server:2024.2', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.11', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'mariadb-clustercheck', 'value': {'container_name': 'mariadb_clustercheck', 'group': 'mariadb_shard_0', 'enabled': False, 'image': 'registry.osism.tech/kolla/mariadb-clustercheck:2024.2', 'volumes': ['/etc/kolla/mariadb-clustercheck/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.11', 'AVAILABLE_WHEN_DONOR': '1'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/mariadb-server:2024.2', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.12', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'mariadb-clustercheck', 'value': {'container_name': 'mariadb_clustercheck', 'group': 'mariadb_shard_0', 'enabled': False, 'image': 'registry.osism.tech/kolla/mariadb-clustercheck:2024.2', 'volumes': ['/etc/kolla/mariadb-clustercheck/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.12', 'AVAILABLE_WHEN_DONOR': '1'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [haproxy-config : Configuring firewall for mariadb] ***********************\nTuesday 24 February 2026  00:52:51 +0000 (0:00:02.385)       0:03:16.575 ****** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'mariadb', 'value': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'mariadb_external_lb', 'value': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'mariadb', 'value': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'mariadb_external_lb', 'value': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'mariadb', 'value': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'mariadb_external_lb', 'value': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [proxysql-config : Copying over mariadb ProxySQL users config] ************\nTuesday 24 February 2026  00:52:54 +0000 (0:00:02.897)       0:03:19.473 ****** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [proxysql-config : Copying over mariadb ProxySQL rules config] ************\nTuesday 24 February 2026  00:52:56 +0000 (0:00:01.952)       0:03:21.426 ****** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [include_role : masakari] *************************************************\nTuesday 24 February 2026  00:52:58 +0000 (0:00:01.536)       0:03:22.962 ****** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [include_role : memcached] ************************************************\nTuesday 24 February 2026  00:52:58 +0000 (0:00:00.350)       0:03:23.313 ****** \n\u001b[0;36mincluded: memcached for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [haproxy-config : Copying over memcached haproxy config] ******************\nTuesday 24 February 2026  00:52:59 +0000 (0:00:01.433)       0:03:24.747 ****** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'memcached', 'value': {'container_name': 'memcached', 'image': 'registry.osism.tech/kolla/memcached:2024.2', 'enabled': True, 'group': 'memcached', 'volumes': ['/etc/kolla/memcached/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen memcached 11211'], 'timeout': '30'}, 'haproxy': {'memcached': {'enabled': False, 'mode': 'tcp', 'port': '11211', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'active_passive': True}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'memcached', 'value': {'container_name': 'memcached', 'image': 'registry.osism.tech/kolla/memcached:2024.2', 'enabled': True, 'group': 'memcached', 'volumes': ['/etc/kolla/memcached/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen memcached 11211'], 'timeout': '30'}, 'haproxy': {'memcached': {'enabled': False, 'mode': 'tcp', 'port': '11211', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'active_passive': True}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'memcached', 'value': {'container_name': 'memcached', 'image': 'registry.osism.tech/kolla/memcached:2024.2', 'enabled': True, 'group': 'memcached', 'volumes': ['/etc/kolla/memcached/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen memcached 11211'], 'timeout': '30'}, 'haproxy': {'memcached': {'enabled': False, 'mode': 'tcp', 'port': '11211', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'active_passive': True}}}})\u001b[0m\n\nTASK [haproxy-config : Add configuration for memcached when using single external frontend] ***\nTuesday 24 February 2026  00:53:01 +0000 (0:00:01.566)       0:03:26.313 ****** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'memcached', 'value': {'container_name': 'memcached', 'image': 'registry.osism.tech/kolla/memcached:2024.2', 'enabled': True, 'group': 'memcached', 'volumes': ['/etc/kolla/memcached/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen memcached 11211'], 'timeout': '30'}, 'haproxy': {'memcached': {'enabled': False, 'mode': 'tcp', 'port': '11211', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'active_passive': True}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'memcached', 'value': {'container_name': 'memcached', 'image': 'registry.osism.tech/kolla/memcached:2024.2', 'enabled': True, 'group': 'memcached', 'volumes': ['/etc/kolla/memcached/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen memcached 11211'], 'timeout': '30'}, 'haproxy': {'memcached': {'enabled': False, 'mode': 'tcp', 'port': '11211', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'active_passive': True}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'memcached', 'value': {'container_name': 'memcached', 'image': 'registry.osism.tech/kolla/memcached:2024.2', 'enabled': True, 'group': 'memcached', 'volumes': ['/etc/kolla/memcached/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen memcached 11211'], 'timeout': '30'}, 'haproxy': {'memcached': {'enabled': False, 'mode': 'tcp', 'port': '11211', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'active_passive': True}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [haproxy-config : Configuring firewall for memcached] *********************\nTuesday 24 February 2026  00:53:01 +0000 (0:00:00.398)       0:03:26.711 ****** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'memcached', 'value': {'enabled': False, 'mode': 'tcp', 'port': '11211', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'active_passive': True}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'memcached', 'value': {'enabled': False, 'mode': 'tcp', 'port': '11211', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'active_passive': True}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'memcached', 'value': {'enabled': False, 'mode': 'tcp', 'port': '11211', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'active_passive': True}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [proxysql-config : Copying over memcached ProxySQL users config] **********\nTuesday 24 February 2026  00:53:02 +0000 (0:00:00.792)       0:03:27.504 ****** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [proxysql-config : Copying over memcached ProxySQL rules config] **********\nTuesday 24 February 2026  00:53:03 +0000 (0:00:00.440)       0:03:27.944 ****** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [include_role : mistral] **************************************************\nTuesday 24 February 2026  00:53:04 +0000 (0:00:01.288)       0:03:29.232 ****** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [include_role : neutron] **************************************************\nTuesday 24 February 2026  00:53:04 +0000 (0:00:00.329)       0:03:29.562 ****** \n\u001b[0;36mincluded: neutron for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [haproxy-config : Copying over neutron haproxy config] ********************\nTuesday 24 February 2026  00:53:06 +0000 (0:00:01.453)       0:03:31.016 ****** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2024.2', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-openvswitch-agent', 'value': {'container_name': 'neutron_openvswitch_agent', 'image': 'registry.osism.tech/kolla/neutron-openvswitch-agent:2024.2', 'enabled': False, 'privileged': True, 'environment': {'KOLLA_LEGACY_IPTABLES': 'false'}, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-openvswitch-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-openvswitch-agent 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-linuxbridge-agent', 'value': {'container_name': 'neutron_linuxbridge_agent', 'image': 'registry.osism.tech/kolla/neutron-linuxbridge-agent:2024.2', 'privileged': True, 'enabled': False, 'environment': {'KOLLA_LEGACY_IPTABLES': 'false'}, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-linuxbridge-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-linuxbridge-agent 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-dhcp-agent', 'value': {'container_name': 'neutron_dhcp_agent', 'image': 'registry.osism.tech/kolla/neutron-dhcp-agent:2024.2', 'privileged': True, 'enabled': False, 'group': 'neutron-dhcp-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-dhcp-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-dhcp-agent 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-l3-agent', 'value': {'container_name': 'neutron_l3_agent', 'image': 'registry.osism.tech/kolla/neutron-l3-agent:2024.2', 'privileged': True, 'enabled': False, 'environment': {'KOLLA_LEGACY_IPTABLES': 'false'}, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-l3-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', \"healthcheck_port 'neutron-l3-agent ' 5672\"], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-sriov-agent', 'value': {'container_name': 'neutron_sriov_agent', 'image': 'registry.osism.tech/kolla/neutron-sriov-agent:2024.2', 'privileged': True, 'enabled': False, 'host_in_groups': False, 'volumes': ['/etc/kolla/neutron-sriov-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-sriov-nic-agent 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-mlnx-agent', 'value': {'container_name': 'neutron_mlnx_agent', 'image': 'registry.osism.tech/kolla/neutron-mlnx-agent:2024.2', 'enabled': False, 'host_in_groups': False, 'volumes': ['/etc/kolla/neutron-mlnx-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}}}) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2024.2', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-eswitchd', 'value': {'container_name': 'neutron_eswitchd', 'image': 'registry.osism.tech/kolla/neutron-eswitchd:2024.2', 'privileged': True, 'enabled': False, 'host_in_groups': False, 'volumes': ['/etc/kolla/neutron-eswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/run/libvirt:/run/libvirt:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-metadata-agent', 'value': {'container_name': 'neutron_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'privileged': True, 'enabled': False, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-metadata-agent 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-openvswitch-agent', 'value': {'container_name': 'neutron_openvswitch_agent', 'image': 'registry.osism.tech/kolla/neutron-openvswitch-agent:2024.2', 'enabled': False, 'privileged': True, 'environment': {'KOLLA_LEGACY_IPTABLES': 'false'}, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-openvswitch-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-openvswitch-agent 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'privileged': True, 'enabled': True, 'host_in_groups': False, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-linuxbridge-agent', 'value': {'container_name': 'neutron_linuxbridge_agent', 'image': 'registry.osism.tech/kolla/neutron-linuxbridge-agent:2024.2', 'privileged': True, 'enabled': False, 'environment': {'KOLLA_LEGACY_IPTABLES': 'false'}, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-linuxbridge-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-linuxbridge-agent 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-bgp-dragent', 'value': {'container_name': 'neutron_bgp_dragent', 'image': 'registry.osism.tech/kolla/neutron-bgp-dragent:2024.2', 'privileged': True, 'enabled': False, 'group': 'neutron-bgp-dragent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-bgp-dragent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-bgp-dragent 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-dhcp-agent', 'value': {'container_name': 'neutron_dhcp_agent', 'image': 'registry.osism.tech/kolla/neutron-dhcp-agent:2024.2', 'privileged': True, 'enabled': False, 'group': 'neutron-dhcp-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-dhcp-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-dhcp-agent 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-infoblox-ipam-agent', 'value': {'container_name': 'neutron_infoblox_ipam_agent', 'image': 'registry.osism.tech/kolla/neutron-infoblox-ipam-agent:2024.2', 'privileged': True, 'enabled': False, 'group': 'neutron-infoblox-ipam-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-infoblox-ipam-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-l3-agent', 'value': {'container_name': 'neutron_l3_agent', 'image': 'registry.osism.tech/kolla/neutron-l3-agent:2024.2', 'privileged': True, 'enabled': False, 'environment': {'KOLLA_LEGACY_IPTABLES': 'false'}, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-l3-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', \"healthcheck_port 'neutron-l3-agent ' 5672\"], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-metering-agent', 'value': {'container_name': 'neutron_metering_agent', 'image': 'registry.osism.tech/kolla/neutron-metering-agent:2024.2', 'privileged': True, 'enabled': False, 'group': 'neutron-metering-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-metering-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}}}) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2024.2', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'ironic-neutron-agent', 'value': {'container_name': 'ironic_neutron_agent', 'image': 'registry.osism.tech/kolla/ironic-neutron-agent:2024.2', 'privileged': False, 'enabled': False, 'group': 'ironic-neutron-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/ironic-neutron-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port ironic-neutron-agent 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-sriov-agent', 'value': {'container_name': 'neutron_sriov_agent', 'image': 'registry.osism.tech/kolla/neutron-sriov-agent:2024.2', 'privileged': True, 'enabled': False, 'host_in_groups': False, 'volumes': ['/etc/kolla/neutron-sriov-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-sriov-nic-agent 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-openvswitch-agent', 'value': {'container_name': 'neutron_openvswitch_agent', 'image': 'registry.osism.tech/kolla/neutron-openvswitch-agent:2024.2', 'enabled': False, 'privileged': True, 'environment': {'KOLLA_LEGACY_IPTABLES': 'false'}, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-openvswitch-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-openvswitch-agent 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-tls-proxy', 'value': {'container_name': 'neutron_tls_proxy', 'group': 'neutron-server', 'host_in_groups': True, 'enabled': 'no', 'image': 'registry.osism.tech/kolla/neutron-tls-proxy:2024.2', 'volumes': ['/etc/kolla/neutron-tls-proxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl -u openstack:password 192.168.16.10:9697'], 'timeout': '30'}, 'haproxy': {'neutron_tls_proxy': {'enabled': False, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696', 'tls_backend': 'yes'}, 'neutron_tls_proxy_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696', 'tls_backend': 'yes'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-linuxbridge-agent', 'value': {'container_name': 'neutron_linuxbridge_agent', 'image': 'registry.osism.tech/kolla/neutron-linuxbridge-agent:2024.2', 'privileged': True, 'enabled': False, 'environment': {'KOLLA_LEGACY_IPTABLES': 'false'}, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-linuxbridge-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-linuxbridge-agent 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-ovn-agent', 'value': {'container_name': 'neutron_ovn_agent', 'group': 'neutron-ovn-agent', 'host_in_groups': False, 'enabled': False, 'image': 'registry.osism.tech/dockerhub/kolla/neutron-ovn-agent:2024.2', 'volumes': ['/etc/kolla/neutron-ovn-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-agent 6640'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-dhcp-agent', 'value': {'container_name': 'neutron_dhcp_agent', 'image': 'registry.osism.tech/kolla/neutron-dhcp-agent:2024.2', 'privileged': True, 'enabled': False, 'group': 'neutron-dhcp-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-dhcp-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-dhcp-agent 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-l3-agent', 'value': {'container_name': 'neutron_l3_agent', 'image': 'registry.osism.tech/kolla/neutron-l3-agent:2024.2', 'privileged': True, 'enabled': False, 'environment': {'KOLLA_LEGACY_IPTABLES': 'false'}, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-l3-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', \"healthcheck_port 'neutron-l3-agent ' 5672\"], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-sriov-agent', 'value': {'container_name': 'neutron_sriov_agent', 'image': 'registry.osism.tech/kolla/neutron-sriov-agent:2024.2', 'privileged': True, 'enabled': False, 'host_in_groups': False, 'volumes': ['/etc/kolla/neutron-sriov-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-sriov-nic-agent 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-mlnx-agent', 'value': {'container_name': 'neutron_mlnx_agent', 'image': 'registry.osism.tech/kolla/neutron-mlnx-agent:2024.2', 'enabled': False, 'host_in_groups': False, 'volumes': ['/etc/kolla/neutron-mlnx-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-eswitchd', 'value': {'container_name': 'neutron_eswitchd', 'image': 'registry.osism.tech/kolla/neutron-eswitchd:2024.2', 'privileged': True, 'enabled': False, 'host_in_groups': False, 'volumes': ['/etc/kolla/neutron-eswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/run/libvirt:/run/libvirt:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-mlnx-agent', 'value': {'container_name': 'neutron_mlnx_agent', 'image': 'registry.osism.tech/kolla/neutron-mlnx-agent:2024.2', 'enabled': False, 'host_in_groups': False, 'volumes': ['/etc/kolla/neutron-mlnx-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-metadata-agent', 'value': {'container_name': 'neutron_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'privileged': True, 'enabled': False, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-metadata-agent 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-eswitchd', 'value': {'container_name': 'neutron_eswitchd', 'image': 'registry.osism.tech/kolla/neutron-eswitchd:2024.2', 'privileged': True, 'enabled': False, 'host_in_groups': False, 'volumes': ['/etc/kolla/neutron-eswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/run/libvirt:/run/libvirt:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'privileged': True, 'enabled': True, 'host_in_groups': False, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-bgp-dragent', 'value': {'container_name': 'neutron_bgp_dragent', 'image': 'registry.osism.tech/kolla/neutron-bgp-dragent:2024.2', 'privileged': True, 'enabled': False, 'group': 'neutron-bgp-dragent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-bgp-dragent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-bgp-dragent 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-metadata-agent', 'value': {'container_name': 'neutron_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'privileged': True, 'enabled': False, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-metadata-agent 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-infoblox-ipam-agent', 'value': {'container_name': 'neutron_infoblox_ipam_agent', 'image': 'registry.osism.tech/kolla/neutron-infoblox-ipam-agent:2024.2', 'privileged': True, 'enabled': False, 'group': 'neutron-infoblox-ipam-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-infoblox-ipam-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'privileged': True, 'enabled': True, 'host_in_groups': False, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-metering-agent', 'value': {'container_name': 'neutron_metering_agent', 'image': 'registry.osism.tech/kolla/neutron-metering-agent:2024.2', 'privileged': True, 'enabled': False, 'group': 'neutron-metering-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-metering-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-bgp-dragent', 'value': {'container_name': 'neutron_bgp_dragent', 'image': 'registry.osism.tech/kolla/neutron-bgp-dragent:2024.2', 'privileged': True, 'enabled': False, 'group': 'neutron-bgp-dragent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-bgp-dragent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-bgp-dragent 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'ironic-neutron-agent', 'value': {'container_name': 'ironic_neutron_agent', 'image': 'registry.osism.tech/kolla/ironic-neutron-agent:2024.2', 'privileged': False, 'enabled': False, 'group': 'ironic-neutron-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/ironic-neutron-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port ironic-neutron-agent 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-infoblox-ipam-agent', 'value': {'container_name': 'neutron_infoblox_ipam_agent', 'image': 'registry.osism.tech/kolla/neutron-infoblox-ipam-agent:2024.2', 'privileged': True, 'enabled': False, 'group': 'neutron-infoblox-ipam-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-infoblox-ipam-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-tls-proxy', 'value': {'container_name': 'neutron_tls_proxy', 'group': 'neutron-server', 'host_in_groups': True, 'enabled': 'no', 'image': 'registry.osism.tech/kolla/neutron-tls-proxy:2024.2', 'volumes': ['/etc/kolla/neutron-tls-proxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl -u openstack:password 192.168.16.12:9697'], 'timeout': '30'}, 'haproxy': {'neutron_tls_proxy': {'enabled': False, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696', 'tls_backend': 'yes'}, 'neutron_tls_proxy_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696', 'tls_backend': 'yes'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-metering-agent', 'value': {'container_name': 'neutron_metering_agent', 'image': 'registry.osism.tech/kolla/neutron-metering-agent:2024.2', 'privileged': True, 'enabled': False, 'group': 'neutron-metering-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-metering-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-ovn-agent', 'value': {'container_name': 'neutron_ovn_agent', 'group': 'neutron-ovn-agent', 'host_in_groups': False, 'enabled': False, 'image': 'registry.osism.tech/dockerhub/kolla/neutron-ovn-agent:2024.2', 'volumes': ['/etc/kolla/neutron-ovn-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-agent 6640'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'ironic-neutron-agent', 'value': {'container_name': 'ironic_neutron_agent', 'image': 'registry.osism.tech/kolla/ironic-neutron-agent:2024.2', 'privileged': False, 'enabled': False, 'group': 'ironic-neutron-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/ironic-neutron-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port ironic-neutron-agent 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-tls-proxy', 'value': {'container_name': 'neutron_tls_proxy', 'group': 'neutron-server', 'host_in_groups': True, 'enabled': 'no', 'image': 'registry.osism.tech/kolla/neutron-tls-proxy:2024.2', 'volumes': ['/etc/kolla/neutron-tls-proxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl -u openstack:password 192.168.16.11:9697'], 'timeout': '30'}, 'haproxy': {'neutron_tls_proxy': {'enabled': False, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696', 'tls_backend': 'yes'}, 'neutron_tls_proxy_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696', 'tls_backend': 'yes'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-ovn-agent', 'value': {'container_name': 'neutron_ovn_agent', 'group': 'neutron-ovn-agent', 'host_in_groups': False, 'enabled': False, 'image': 'registry.osism.tech/dockerhub/kolla/neutron-ovn-agent:2024.2', 'volumes': ['/etc/kolla/neutron-ovn-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-agent 6640'], 'timeout': '30'}}}) \u001b[0m\n\nTASK [haproxy-config : Add configuration for neutron when using single external frontend] ***\nTuesday 24 February 2026  00:53:10 +0000 (0:00:04.136)       0:03:35.152 ****** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2024.2', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-openvswitch-agent', 'value': {'container_name': 'neutron_openvswitch_agent', 'image': 'registry.osism.tech/kolla/neutron-openvswitch-agent:2024.2', 'enabled': False, 'privileged': True, 'environment': {'KOLLA_LEGACY_IPTABLES': 'false'}, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-openvswitch-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-openvswitch-agent 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-linuxbridge-agent', 'value': {'container_name': 'neutron_linuxbridge_agent', 'image': 'registry.osism.tech/kolla/neutron-linuxbridge-agent:2024.2', 'privileged': True, 'enabled': False, 'environment': {'KOLLA_LEGACY_IPTABLES': 'false'}, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-linuxbridge-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-linuxbridge-agent 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-dhcp-agent', 'value': {'container_name': 'neutron_dhcp_agent', 'image': 'registry.osism.tech/kolla/neutron-dhcp-agent:2024.2', 'privileged': True, 'enabled': False, 'group': 'neutron-dhcp-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-dhcp-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-dhcp-agent 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-l3-agent', 'value': {'container_name': 'neutron_l3_agent', 'image': 'registry.osism.tech/kolla/neutron-l3-agent:2024.2', 'privileged': True, 'enabled': False, 'environment': {'KOLLA_LEGACY_IPTABLES': 'false'}, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-l3-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', \"healthcheck_port 'neutron-l3-agent ' 5672\"], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-sriov-agent', 'value': {'container_name': 'neutron_sriov_agent', 'image': 'registry.osism.tech/kolla/neutron-sriov-agent:2024.2', 'privileged': True, 'enabled': False, 'host_in_groups': False, 'volumes': ['/etc/kolla/neutron-sriov-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-sriov-nic-agent 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-mlnx-agent', 'value': {'container_name': 'neutron_mlnx_agent', 'image': 'registry.osism.tech/kolla/neutron-mlnx-agent:2024.2', 'enabled': False, 'host_in_groups': False, 'volumes': ['/etc/kolla/neutron-mlnx-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2024.2', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-eswitchd', 'value': {'container_name': 'neutron_eswitchd', 'image': 'registry.osism.tech/kolla/neutron-eswitchd:2024.2', 'privileged': True, 'enabled': False, 'host_in_groups': False, 'volumes': ['/etc/kolla/neutron-eswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/run/libvirt:/run/libvirt:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-openvswitch-agent', 'value': {'container_name': 'neutron_openvswitch_agent', 'image': 'registry.osism.tech/kolla/neutron-openvswitch-agent:2024.2', 'enabled': False, 'privileged': True, 'environment': {'KOLLA_LEGACY_IPTABLES': 'false'}, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-openvswitch-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-openvswitch-agent 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-metadata-agent', 'value': {'container_name': 'neutron_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'privileged': True, 'enabled': False, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-metadata-agent 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-linuxbridge-agent', 'value': {'container_name': 'neutron_linuxbridge_agent', 'image': 'registry.osism.tech/kolla/neutron-linuxbridge-agent:2024.2', 'privileged': True, 'enabled': False, 'environment': {'KOLLA_LEGACY_IPTABLES': 'false'}, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-linuxbridge-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-linuxbridge-agent 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'privileged': True, 'enabled': True, 'host_in_groups': False, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-dhcp-agent', 'value': {'container_name': 'neutron_dhcp_agent', 'image': 'registry.osism.tech/kolla/neutron-dhcp-agent:2024.2', 'privileged': True, 'enabled': False, 'group': 'neutron-dhcp-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-dhcp-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-dhcp-agent 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-bgp-dragent', 'value': {'container_name': 'neutron_bgp_dragent', 'image': 'registry.osism.tech/kolla/neutron-bgp-dragent:2024.2', 'privileged': True, 'enabled': False, 'group': 'neutron-bgp-dragent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-bgp-dragent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-bgp-dragent 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-infoblox-ipam-agent', 'value': {'container_name': 'neutron_infoblox_ipam_agent', 'image': 'registry.osism.tech/kolla/neutron-infoblox-ipam-agent:2024.2', 'privileged': True, 'enabled': False, 'group': 'neutron-infoblox-ipam-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-infoblox-ipam-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-l3-agent', 'value': {'container_name': 'neutron_l3_agent', 'image': 'registry.osism.tech/kolla/neutron-l3-agent:2024.2', 'privileged': True, 'enabled': False, 'environment': {'KOLLA_LEGACY_IPTABLES': 'false'}, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-l3-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', \"healthcheck_port 'neutron-l3-agent ' 5672\"], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2024.2', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-metering-agent', 'value': {'container_name': 'neutron_metering_agent', 'image': 'registry.osism.tech/kolla/neutron-metering-agent:2024.2', 'privileged': True, 'enabled': False, 'group': 'neutron-metering-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-metering-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-sriov-agent', 'value': {'container_name': 'neutron_sriov_agent', 'image': 'registry.osism.tech/kolla/neutron-sriov-agent:2024.2', 'privileged': True, 'enabled': False, 'host_in_groups': False, 'volumes': ['/etc/kolla/neutron-sriov-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-sriov-nic-agent 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-openvswitch-agent', 'value': {'container_name': 'neutron_openvswitch_agent', 'image': 'registry.osism.tech/kolla/neutron-openvswitch-agent:2024.2', 'enabled': False, 'privileged': True, 'environment': {'KOLLA_LEGACY_IPTABLES': 'false'}, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-openvswitch-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-openvswitch-agent 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'ironic-neutron-agent', 'value': {'container_name': 'ironic_neutron_agent', 'image': 'registry.osism.tech/kolla/ironic-neutron-agent:2024.2', 'privileged': False, 'enabled': False, 'group': 'ironic-neutron-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/ironic-neutron-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port ironic-neutron-agent 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-linuxbridge-agent', 'value': {'container_name': 'neutron_linuxbridge_agent', 'image': 'registry.osism.tech/kolla/neutron-linuxbridge-agent:2024.2', 'privileged': True, 'enabled': False, 'environment': {'KOLLA_LEGACY_IPTABLES': 'false'}, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-linuxbridge-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-linuxbridge-agent 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-mlnx-agent', 'value': {'container_name': 'neutron_mlnx_agent', 'image': 'registry.osism.tech/kolla/neutron-mlnx-agent:2024.2', 'enabled': False, 'host_in_groups': False, 'volumes': ['/etc/kolla/neutron-mlnx-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-tls-proxy', 'value': {'container_name': 'neutron_tls_proxy', 'group': 'neutron-server', 'host_in_groups': True, 'enabled': 'no', 'image': 'registry.osism.tech/kolla/neutron-tls-proxy:2024.2', 'volumes': ['/etc/kolla/neutron-tls-proxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl -u openstack:password 192.168.16.10:9697'], 'timeout': '30'}, 'haproxy': {'neutron_tls_proxy': {'enabled': False, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696', 'tls_backend': 'yes'}, 'neutron_tls_proxy_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696', 'tls_backend': 'yes'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-dhcp-agent', 'value': {'container_name': 'neutron_dhcp_agent', 'image': 'registry.osism.tech/kolla/neutron-dhcp-agent:2024.2', 'privileged': True, 'enabled': False, 'group': 'neutron-dhcp-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-dhcp-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-dhcp-agent 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-eswitchd', 'value': {'container_name': 'neutron_eswitchd', 'image': 'registry.osism.tech/kolla/neutron-eswitchd:2024.2', 'privileged': True, 'enabled': False, 'host_in_groups': False, 'volumes': ['/etc/kolla/neutron-eswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/run/libvirt:/run/libvirt:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-ovn-agent', 'value': {'container_name': 'neutron_ovn_agent', 'group': 'neutron-ovn-agent', 'host_in_groups': False, 'enabled': False, 'image': 'registry.osism.tech/dockerhub/kolla/neutron-ovn-agent:2024.2', 'volumes': ['/etc/kolla/neutron-ovn-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-agent 6640'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-l3-agent', 'value': {'container_name': 'neutron_l3_agent', 'image': 'registry.osism.tech/kolla/neutron-l3-agent:2024.2', 'privileged': True, 'enabled': False, 'environment': {'KOLLA_LEGACY_IPTABLES': 'false'}, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-l3-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', \"healthcheck_port 'neutron-l3-agent ' 5672\"], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-metadata-agent', 'value': {'container_name': 'neutron_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'privileged': True, 'enabled': False, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-metadata-agent 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-sriov-agent', 'value': {'container_name': 'neutron_sriov_agent', 'image': 'registry.osism.tech/kolla/neutron-sriov-agent:2024.2', 'privileged': True, 'enabled': False, 'host_in_groups': False, 'volumes': ['/etc/kolla/neutron-sriov-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-sriov-nic-agent 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'privileged': True, 'enabled': True, 'host_in_groups': False, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-mlnx-agent', 'value': {'container_name': 'neutron_mlnx_agent', 'image': 'registry.osism.tech/kolla/neutron-mlnx-agent:2024.2', 'enabled': False, 'host_in_groups': False, 'volumes': ['/etc/kolla/neutron-mlnx-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-bgp-dragent', 'value': {'container_name': 'neutron_bgp_dragent', 'image': 'registry.osism.tech/kolla/neutron-bgp-dragent:2024.2', 'privileged': True, 'enabled': False, 'group': 'neutron-bgp-dragent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-bgp-dragent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-bgp-dragent 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-eswitchd', 'value': {'container_name': 'neutron_eswitchd', 'image': 'registry.osism.tech/kolla/neutron-eswitchd:2024.2', 'privileged': True, 'enabled': False, 'host_in_groups': False, 'volumes': ['/etc/kolla/neutron-eswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/run/libvirt:/run/libvirt:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-infoblox-ipam-agent', 'value': {'container_name': 'neutron_infoblox_ipam_agent', 'image': 'registry.osism.tech/kolla/neutron-infoblox-ipam-agent:2024.2', 'privileged': True, 'enabled': False, 'group': 'neutron-infoblox-ipam-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-infoblox-ipam-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-metadata-agent', 'value': {'container_name': 'neutron_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'privileged': True, 'enabled': False, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-metadata-agent 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-metering-agent', 'value': {'container_name': 'neutron_metering_agent', 'image': 'registry.osism.tech/kolla/neutron-metering-agent:2024.2', 'privileged': True, 'enabled': False, 'group': 'neutron-metering-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-metering-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'ironic-neutron-agent', 'value': {'container_name': 'ironic_neutron_agent', 'image': 'registry.osism.tech/kolla/ironic-neutron-agent:2024.2', 'privileged': False, 'enabled': False, 'group': 'ironic-neutron-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/ironic-neutron-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port ironic-neutron-agent 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'privileged': True, 'enabled': True, 'host_in_groups': False, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-tls-proxy', 'value': {'container_name': 'neutron_tls_proxy', 'group': 'neutron-server', 'host_in_groups': True, 'enabled': 'no', 'image': 'registry.osism.tech/kolla/neutron-tls-proxy:2024.2', 'volumes': ['/etc/kolla/neutron-tls-proxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl -u openstack:password 192.168.16.11:9697'], 'timeout': '30'}, 'haproxy': {'neutron_tls_proxy': {'enabled': False, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696', 'tls_backend': 'yes'}, 'neutron_tls_proxy_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696', 'tls_backend': 'yes'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-bgp-dragent', 'value': {'container_name': 'neutron_bgp_dragent', 'image': 'registry.osism.tech/kolla/neutron-bgp-dragent:2024.2', 'privileged': True, 'enabled': False, 'group': 'neutron-bgp-dragent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-bgp-dragent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-bgp-dragent 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-ovn-agent', 'value': {'container_name': 'neutron_ovn_agent', 'group': 'neutron-ovn-agent', 'host_in_groups': False, 'enabled': False, 'image': 'registry.osism.tech/dockerhub/kolla/neutron-ovn-agent:2024.2', 'volumes': ['/etc/kolla/neutron-ovn-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-agent 6640'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-infoblox-ipam-agent', 'value': {'container_name': 'neutron_infoblox_ipam_agent', 'image': 'registry.osism.tech/kolla/neutron-infoblox-ipam-agent:2024.2', 'privileged': True, 'enabled': False, 'group': 'neutron-infoblox-ipam-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-infoblox-ipam-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-metering-agent', 'value': {'container_name': 'neutron_metering_agent', 'image': 'registry.osism.tech/kolla/neutron-metering-agent:2024.2', 'privileged': True, 'enabled': False, 'group': 'neutron-metering-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-metering-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'ironic-neutron-agent', 'value': {'container_name': 'ironic_neutron_agent', 'image': 'registry.osism.tech/kolla/ironic-neutron-agent:2024.2', 'privileged': False, 'enabled': False, 'group': 'ironic-neutron-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/ironic-neutron-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port ironic-neutron-agent 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-tls-proxy', 'value': {'container_name': 'neutron_tls_proxy', 'group': 'neutron-server', 'host_in_groups': True, 'enabled': 'no', 'image': 'registry.osism.tech/kolla/neutron-tls-proxy:2024.2', 'volumes': ['/etc/kolla/neutron-tls-proxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl -u openstack:password 192.168.16.12:9697'], 'timeout': '30'}, 'haproxy': {'neutron_tls_proxy': {'enabled': False, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696', 'tls_backend': 'yes'}, 'neutron_tls_proxy_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696', 'tls_backend': 'yes'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-ovn-agent', 'value': {'container_name': 'neutron_ovn_agent', 'group': 'neutron-ovn-agent', 'host_in_groups': False, 'enabled': False, 'image': 'registry.osism.tech/dockerhub/kolla/neutron-ovn-agent:2024.2', 'volumes': ['/etc/kolla/neutron-ovn-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-agent 6640'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [haproxy-config : Configuring firewall for neutron] ***********************\nTuesday 24 February 2026  00:53:11 +0000 (0:00:01.325)       0:03:36.478 ****** \n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron_server', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron_server_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron_server', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron_server_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron_server', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron_server_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [proxysql-config : Copying over neutron ProxySQL users config] ************\nTuesday 24 February 2026  00:53:13 +0000 (0:00:01.750)       0:03:38.228 ****** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [proxysql-config : Copying over neutron ProxySQL rules config] ************\nTuesday 24 February 2026  00:53:14 +0000 (0:00:01.397)       0:03:39.626 ****** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [include_role : placement] ************************************************\nTuesday 24 February 2026  00:53:16 +0000 (0:00:02.086)       0:03:41.712 ****** \n\u001b[0;36mincluded: placement for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [haproxy-config : Copying over placement haproxy config] ******************\nTuesday 24 February 2026  00:53:18 +0000 (0:00:01.206)       0:03:42.919 ****** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/placement-api:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8780'], 'timeout': '30'}, 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/placement-api:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8780'], 'timeout': '30'}, 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/placement-api:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8780'], 'timeout': '30'}, 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}}})\u001b[0m\n\nTASK [haproxy-config : Add configuration for placement when using single external frontend] ***\nTuesday 24 February 2026  00:53:21 +0000 (0:00:03.743)       0:03:46.663 ****** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/placement-api:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8780'], 'timeout': '30'}, 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/placement-api:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8780'], 'timeout': '30'}, 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/placement-api:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8780'], 'timeout': '30'}, 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [haproxy-config : Configuring firewall for placement] *********************\nTuesday 24 February 2026  00:53:22 +0000 (0:00:00.551)       0:03:47.215 ****** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'placement_api', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'placement_api_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'placement_api', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'placement_api_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'placement_api', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'placement_api_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [proxysql-config : Copying over placement ProxySQL users config] **********\nTuesday 24 February 2026  00:53:22 +0000 (0:00:00.682)       0:03:47.897 ****** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [proxysql-config : Copying over placement ProxySQL rules config] **********\nTuesday 24 February 2026  00:53:24 +0000 (0:00:01.743)       0:03:49.640 ****** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [include_role : nova] *****************************************************\nTuesday 24 February 2026  00:53:26 +0000 (0:00:01.873)       0:03:51.513 ****** \n\u001b[0;36mincluded: nova for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [haproxy-config : Copying over nova haproxy config] ***********************\nTuesday 24 February 2026  00:53:28 +0000 (0:00:01.578)       0:03:53.092 ****** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/nova-api:2024.2', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8774 '], 'timeout': '30'}, 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/nova-scheduler:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova-super-conductor', 'value': {'container_name': 'nova_super_conductor', 'group': 'nova-super-conductor', 'enabled': 'no', 'image': 'registry.osism.tech/kolla/nova-super-conductor:2024.2', 'volumes': ['/etc/kolla/nova-super-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/nova-api:2024.2', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8774 '], 'timeout': '30'}, 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/nova-scheduler:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova-super-conductor', 'value': {'container_name': 'nova_super_conductor', 'group': 'nova-super-conductor', 'enabled': 'no', 'image': 'registry.osism.tech/kolla/nova-super-conductor:2024.2', 'volumes': ['/etc/kolla/nova-super-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/nova-api:2024.2', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8774 '], 'timeout': '30'}, 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/nova-scheduler:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova-super-conductor', 'value': {'container_name': 'nova_super_conductor', 'group': 'nova-super-conductor', 'enabled': 'no', 'image': 'registry.osism.tech/kolla/nova-super-conductor:2024.2', 'volumes': ['/etc/kolla/nova-super-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}}) \u001b[0m\n\nTASK [haproxy-config : Add configuration for nova when using single external frontend] ***\nTuesday 24 February 2026  00:53:32 +0000 (0:00:04.504)       0:03:57.596 ****** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/nova-api:2024.2', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8774 '], 'timeout': '30'}, 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/nova-scheduler:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova-super-conductor', 'value': {'container_name': 'nova_super_conductor', 'group': 'nova-super-conductor', 'enabled': 'no', 'image': 'registry.osism.tech/kolla/nova-super-conductor:2024.2', 'volumes': ['/etc/kolla/nova-super-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/nova-api:2024.2', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8774 '], 'timeout': '30'}, 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/nova-scheduler:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova-super-conductor', 'value': {'container_name': 'nova_super_conductor', 'group': 'nova-super-conductor', 'enabled': 'no', 'image': 'registry.osism.tech/kolla/nova-super-conductor:2024.2', 'volumes': ['/etc/kolla/nova-super-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/nova-api:2024.2', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8774 '], 'timeout': '30'}, 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/nova-scheduler:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova-super-conductor', 'value': {'container_name': 'nova_super_conductor', 'group': 'nova-super-conductor', 'enabled': 'no', 'image': 'registry.osism.tech/kolla/nova-super-conductor:2024.2', 'volumes': ['/etc/kolla/nova-super-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [haproxy-config : Configuring firewall for nova] **************************\nTuesday 24 February 2026  00:53:33 +0000 (0:00:01.282)       0:03:58.879 ****** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova_api', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova_api_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova_metadata', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova_metadata_external', 'value': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova_api', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova_api_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova_metadata', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova_metadata_external', 'value': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova_api', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova_api_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova_metadata', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova_metadata_external', 'value': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [proxysql-config : Copying over nova ProxySQL users config] ***************\nTuesday 24 February 2026  00:53:35 +0000 (0:00:01.082)       0:03:59.961 ****** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\nTASK [proxysql-config : Copying over nova ProxySQL rules config] ***************\nTuesday 24 February 2026  00:53:36 +0000 (0:00:01.694)       0:04:01.656 ****** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [include_role : nova-cell] ************************************************\nTuesday 24 February 2026  00:53:39 +0000 (0:00:02.421)       0:04:04.078 ****** \n\u001b[0;36mincluded: nova-cell for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [nova-cell : Configure loadbalancer for nova-novncproxy] ******************\nTuesday 24 February 2026  00:53:40 +0000 (0:00:01.717)       0:04:05.795 ****** \n\u001b[0;36mincluded: /ansible/roles/nova-cell/tasks/cell_proxy_loadbalancer.yml for testbed-node-0, testbed-node-1, testbed-node-2 => (item=nova-novncproxy)\u001b[0m\n\nTASK [haproxy-config : Copying over nova-cell:nova-novncproxy haproxy config] ***\nTuesday 24 February 2026  00:53:41 +0000 (0:00:00.899)       0:04:06.695 ****** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'nova-novncproxy', 'value': {'group': 'nova-novncproxy', 'enabled': True, 'haproxy': {'nova_novncproxy': {'enabled': True, 'mode': 'http', 'external': False, 'port': '6080', 'listen_port': '6080', 'backend_http_extra': ['timeout tunnel 1h']}, 'nova_novncproxy_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6080', 'listen_port': '6080', 'backend_http_extra': ['timeout tunnel 1h']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'nova-novncproxy', 'value': {'group': 'nova-novncproxy', 'enabled': True, 'haproxy': {'nova_novncproxy': {'enabled': True, 'mode': 'http', 'external': False, 'port': '6080', 'listen_port': '6080', 'backend_http_extra': ['timeout tunnel 1h']}, 'nova_novncproxy_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6080', 'listen_port': '6080', 'backend_http_extra': ['timeout tunnel 1h']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'nova-novncproxy', 'value': {'group': 'nova-novncproxy', 'enabled': True, 'haproxy': {'nova_novncproxy': {'enabled': True, 'mode': 'http', 'external': False, 'port': '6080', 'listen_port': '6080', 'backend_http_extra': ['timeout tunnel 1h']}, 'nova_novncproxy_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6080', 'listen_port': '6080', 'backend_http_extra': ['timeout tunnel 1h']}}}})\u001b[0m\n\nTASK [haproxy-config : Add configuration for nova-cell:nova-novncproxy when using single external frontend] ***\nTuesday 24 February 2026  00:53:46 +0000 (0:00:04.791)       0:04:11.487 ****** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova-novncproxy', 'value': {'group': 'nova-novncproxy', 'enabled': True, 'haproxy': {'nova_novncproxy': {'enabled': True, 'mode': 'http', 'external': False, 'port': '6080', 'listen_port': '6080', 'backend_http_extra': ['timeout tunnel 1h']}, 'nova_novncproxy_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6080', 'listen_port': '6080', 'backend_http_extra': ['timeout tunnel 1h']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova-novncproxy', 'value': {'group': 'nova-novncproxy', 'enabled': True, 'haproxy': {'nova_novncproxy': {'enabled': True, 'mode': 'http', 'external': False, 'port': '6080', 'listen_port': '6080', 'backend_http_extra': ['timeout tunnel 1h']}, 'nova_novncproxy_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6080', 'listen_port': '6080', 'backend_http_extra': ['timeout tunnel 1h']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova-novncproxy', 'value': {'group': 'nova-novncproxy', 'enabled': True, 'haproxy': {'nova_novncproxy': {'enabled': True, 'mode': 'http', 'external': False, 'port': '6080', 'listen_port': '6080', 'backend_http_extra': ['timeout tunnel 1h']}, 'nova_novncproxy_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6080', 'listen_port': '6080', 'backend_http_extra': ['timeout tunnel 1h']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [haproxy-config : Configuring firewall for nova-cell:nova-novncproxy] *****\nTuesday 24 February 2026  00:53:47 +0000 (0:00:01.089)       0:04:12.576 ****** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova_novncproxy', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '6080', 'listen_port': '6080', 'backend_http_extra': ['timeout tunnel 1h']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova_novncproxy_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6080', 'listen_port': '6080', 'backend_http_extra': ['timeout tunnel 1h']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova_novncproxy', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '6080', 'listen_port': '6080', 'backend_http_extra': ['timeout tunnel 1h']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova_novncproxy_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6080', 'listen_port': '6080', 'backend_http_extra': ['timeout tunnel 1h']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova_novncproxy', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '6080', 'listen_port': '6080', 'backend_http_extra': ['timeout tunnel 1h']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova_novncproxy_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6080', 'listen_port': '6080', 'backend_http_extra': ['timeout tunnel 1h']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [proxysql-config : Copying over nova-cell ProxySQL users config] **********\nTuesday 24 February 2026  00:53:49 +0000 (0:00:01.729)       0:04:14.306 ****** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [proxysql-config : Copying over nova-cell ProxySQL rules config] **********\nTuesday 24 February 2026  00:53:52 +0000 (0:00:02.960)       0:04:17.267 ****** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [nova-cell : Configure loadbalancer for nova-spicehtml5proxy] *************\nTuesday 24 February 2026  00:53:55 +0000 (0:00:03.438)       0:04:20.705 ****** \n\u001b[0;36mincluded: /ansible/roles/nova-cell/tasks/cell_proxy_loadbalancer.yml for testbed-node-0, testbed-node-1, testbed-node-2 => (item=nova-spicehtml5proxy)\u001b[0m\n\nTASK [haproxy-config : Copying over nova-cell:nova-spicehtml5proxy haproxy config] ***\nTuesday 24 February 2026  00:53:57 +0000 (0:00:01.461)       0:04:22.166 ****** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova-spicehtml5proxy', 'value': {'group': 'nova-spicehtml5proxy', 'enabled': False, 'haproxy': {'nova_spicehtml5proxy': {'enabled': False, 'mode': 'http', 'external': False, 'port': '6082', 'listen_port': '6082', 'backend_http_extra': ['timeout tunnel 1h']}, 'nova_spicehtml5proxy_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6082', 'listen_port': '6082', 'backend_http_extra': ['timeout tunnel 1h']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova-spicehtml5proxy', 'value': {'group': 'nova-spicehtml5proxy', 'enabled': False, 'haproxy': {'nova_spicehtml5proxy': {'enabled': False, 'mode': 'http', 'external': False, 'port': '6082', 'listen_port': '6082', 'backend_http_extra': ['timeout tunnel 1h']}, 'nova_spicehtml5proxy_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6082', 'listen_port': '6082', 'backend_http_extra': ['timeout tunnel 1h']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova-spicehtml5proxy', 'value': {'group': 'nova-spicehtml5proxy', 'enabled': False, 'haproxy': {'nova_spicehtml5proxy': {'enabled': False, 'mode': 'http', 'external': False, 'port': '6082', 'listen_port': '6082', 'backend_http_extra': ['timeout tunnel 1h']}, 'nova_spicehtml5proxy_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6082', 'listen_port': '6082', 'backend_http_extra': ['timeout tunnel 1h']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [haproxy-config : Add configuration for nova-cell:nova-spicehtml5proxy when using single external frontend] ***\nTuesday 24 February 2026  00:53:58 +0000 (0:00:01.400)       0:04:23.567 ****** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova-spicehtml5proxy', 'value': {'group': 'nova-spicehtml5proxy', 'enabled': False, 'haproxy': {'nova_spicehtml5proxy': {'enabled': False, 'mode': 'http', 'external': False, 'port': '6082', 'listen_port': '6082', 'backend_http_extra': ['timeout tunnel 1h']}, 'nova_spicehtml5proxy_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6082', 'listen_port': '6082', 'backend_http_extra': ['timeout tunnel 1h']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova-spicehtml5proxy', 'value': {'group': 'nova-spicehtml5proxy', 'enabled': False, 'haproxy': {'nova_spicehtml5proxy': {'enabled': False, 'mode': 'http', 'external': False, 'port': '6082', 'listen_port': '6082', 'backend_http_extra': ['timeout tunnel 1h']}, 'nova_spicehtml5proxy_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6082', 'listen_port': '6082', 'backend_http_extra': ['timeout tunnel 1h']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova-spicehtml5proxy', 'value': {'group': 'nova-spicehtml5proxy', 'enabled': False, 'haproxy': {'nova_spicehtml5proxy': {'enabled': False, 'mode': 'http', 'external': False, 'port': '6082', 'listen_port': '6082', 'backend_http_extra': ['timeout tunnel 1h']}, 'nova_spicehtml5proxy_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6082', 'listen_port': '6082', 'backend_http_extra': ['timeout tunnel 1h']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [haproxy-config : Configuring firewall for nova-cell:nova-spicehtml5proxy] ***\nTuesday 24 February 2026  00:54:00 +0000 (0:00:01.385)       0:04:24.952 ****** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [proxysql-config : Copying over nova-cell ProxySQL users config] **********\nTuesday 24 February 2026  00:54:02 +0000 (0:00:02.008)       0:04:26.961 ****** \n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [proxysql-config : Copying over nova-cell ProxySQL rules config] **********\nTuesday 24 February 2026  00:54:04 +0000 (0:00:02.605)       0:04:29.567 ****** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [nova-cell : Configure loadbalancer for nova-serialproxy] *****************\nTuesday 24 February 2026  00:54:08 +0000 (0:00:03.427)       0:04:32.995 ****** \n\u001b[0;36mincluded: /ansible/roles/nova-cell/tasks/cell_proxy_loadbalancer.yml for testbed-node-0, testbed-node-1, testbed-node-2 => (item=nova-serialproxy)\u001b[0m\n\nTASK [haproxy-config : Copying over nova-cell:nova-serialproxy haproxy config] ***\nTuesday 24 February 2026  00:54:09 +0000 (0:00:00.953)       0:04:33.948 ****** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova-serialproxy', 'value': {'group': 'nova-serialproxy', 'enabled': False, 'haproxy': {'nova_serialconsole_proxy': {'enabled': False, 'mode': 'http', 'external': False, 'port': '6083', 'listen_port': '6083', 'backend_http_extra': ['timeout tunnel 10m']}, 'nova_serialconsole_proxy_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6083', 'listen_port': '6083', 'backend_http_extra': ['timeout tunnel 10m']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova-serialproxy', 'value': {'group': 'nova-serialproxy', 'enabled': False, 'haproxy': {'nova_serialconsole_proxy': {'enabled': False, 'mode': 'http', 'external': False, 'port': '6083', 'listen_port': '6083', 'backend_http_extra': ['timeout tunnel 10m']}, 'nova_serialconsole_proxy_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6083', 'listen_port': '6083', 'backend_http_extra': ['timeout tunnel 10m']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova-serialproxy', 'value': {'group': 'nova-serialproxy', 'enabled': False, 'haproxy': {'nova_serialconsole_proxy': {'enabled': False, 'mode': 'http', 'external': False, 'port': '6083', 'listen_port': '6083', 'backend_http_extra': ['timeout tunnel 10m']}, 'nova_serialconsole_proxy_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6083', 'listen_port': '6083', 'backend_http_extra': ['timeout tunnel 10m']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [haproxy-config : Add configuration for nova-cell:nova-serialproxy when using single external frontend] ***\nTuesday 24 February 2026  00:54:10 +0000 (0:00:01.553)       0:04:35.502 ****** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova-serialproxy', 'value': {'group': 'nova-serialproxy', 'enabled': False, 'haproxy': {'nova_serialconsole_proxy': {'enabled': False, 'mode': 'http', 'external': False, 'port': '6083', 'listen_port': '6083', 'backend_http_extra': ['timeout tunnel 10m']}, 'nova_serialconsole_proxy_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6083', 'listen_port': '6083', 'backend_http_extra': ['timeout tunnel 10m']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova-serialproxy', 'value': {'group': 'nova-serialproxy', 'enabled': False, 'haproxy': {'nova_serialconsole_proxy': {'enabled': False, 'mode': 'http', 'external': False, 'port': '6083', 'listen_port': '6083', 'backend_http_extra': ['timeout tunnel 10m']}, 'nova_serialconsole_proxy_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6083', 'listen_port': '6083', 'backend_http_extra': ['timeout tunnel 10m']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova-serialproxy', 'value': {'group': 'nova-serialproxy', 'enabled': False, 'haproxy': {'nova_serialconsole_proxy': {'enabled': False, 'mode': 'http', 'external': False, 'port': '6083', 'listen_port': '6083', 'backend_http_extra': ['timeout tunnel 10m']}, 'nova_serialconsole_proxy_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6083', 'listen_port': '6083', 'backend_http_extra': ['timeout tunnel 10m']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [haproxy-config : Configuring firewall for nova-cell:nova-serialproxy] ****\nTuesday 24 February 2026  00:54:12 +0000 (0:00:01.852)       0:04:37.355 ****** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [proxysql-config : Copying over nova-cell ProxySQL users config] **********\nTuesday 24 February 2026  00:54:14 +0000 (0:00:01.635)       0:04:38.990 ****** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [proxysql-config : Copying over nova-cell ProxySQL rules config] **********\nTuesday 24 February 2026  00:54:16 +0000 (0:00:02.429)       0:04:41.420 ****** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [include_role : octavia] **************************************************\nTuesday 24 February 2026  00:54:19 +0000 (0:00:03.447)       0:04:44.867 ****** \n\u001b[0;36mincluded: octavia for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [haproxy-config : Copying over octavia haproxy config] ********************\nTuesday 24 February 2026  00:54:21 +0000 (0:00:01.715)       0:04:46.583 ****** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-api:2024.2', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-driver-agent:2024.2', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-health-manager:2024.2', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-housekeeping:2024.2', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-worker:2024.2', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-api:2024.2', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-driver-agent:2024.2', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-health-manager:2024.2', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-housekeeping:2024.2', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-worker:2024.2', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-api:2024.2', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-driver-agent:2024.2', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-health-manager:2024.2', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-housekeeping:2024.2', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-worker:2024.2', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}}) \u001b[0m\n\nTASK [haproxy-config : Add configuration for octavia when using single external frontend] ***\nTuesday 24 February 2026  00:54:25 +0000 (0:00:03.854)       0:04:50.437 ****** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-api:2024.2', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-driver-agent:2024.2', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-health-manager:2024.2', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-housekeeping:2024.2', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-api:2024.2', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-worker:2024.2', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-driver-agent:2024.2', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-health-manager:2024.2', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-housekeeping:2024.2', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-worker:2024.2', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-api:2024.2', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-driver-agent:2024.2', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-health-manager:2024.2', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-housekeeping:2024.2', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-worker:2024.2', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [haproxy-config : Configuring firewall for octavia] ***********************\nTuesday 24 February 2026  00:54:26 +0000 (0:00:00.814)       0:04:51.252 ****** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'octavia_api', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'octavia_api_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'octavia_api', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'octavia_api_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'octavia_api', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'octavia_api_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [proxysql-config : Copying over octavia ProxySQL users config] ************\nTuesday 24 February 2026  00:54:28 +0000 (0:00:01.860)       0:04:53.113 ****** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [proxysql-config : Copying over octavia ProxySQL rules config] ************\nTuesday 24 February 2026  00:54:29 +0000 (0:00:01.495)       0:04:54.609 ****** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [include_role : opensearch] ***********************************************\nTuesday 24 February 2026  00:54:32 +0000 (0:00:02.349)       0:04:56.958 ****** \n\u001b[0;36mincluded: opensearch for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [haproxy-config : Copying over opensearch haproxy config] *****************\nTuesday 24 February 2026  00:54:33 +0000 (0:00:01.715)       0:04:58.673 ****** \n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/opensearch:2024.2', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/opensearch:2024.2', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/opensearch:2024.2', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/opensearch-dashboards:2024.2', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/opensearch-dashboards:2024.2', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/opensearch-dashboards:2024.2', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}}}})\u001b[0m\n\nTASK [haproxy-config : Add configuration for opensearch when using single external frontend] ***\nTuesday 24 February 2026  00:54:39 +0000 (0:00:06.124)       0:05:04.798 ****** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/opensearch:2024.2', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/opensearch-dashboards:2024.2', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/opensearch:2024.2', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/opensearch-dashboards:2024.2', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/opensearch:2024.2', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/opensearch-dashboards:2024.2', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [haproxy-config : Configuring firewall for opensearch] ********************\nTuesday 24 February 2026  00:54:40 +0000 (0:00:00.757)       0:05:05.556 ****** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'opensearch', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'opensearch-dashboards', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'opensearch_dashboards_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'opensearch', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'opensearch-dashboards', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'opensearch_dashboards_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'opensearch', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal']}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'opensearch-dashboards', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'opensearch_dashboards_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [proxysql-config : Copying over opensearch ProxySQL users config] *********\nTuesday 24 February 2026  00:54:41 +0000 (0:00:00.904)       0:05:06.461 ****** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [proxysql-config : Copying over opensearch ProxySQL rules config] *********\nTuesday 24 February 2026  00:54:42 +0000 (0:00:00.935)       0:05:07.396 ****** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [include_role : prometheus] ***********************************************\nTuesday 24 February 2026  00:54:43 +0000 (0:00:01.494)       0:05:08.890 ****** \n\u001b[0;36mincluded: prometheus for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [haproxy-config : Copying over prometheus haproxy config] *****************\nTuesday 24 February 2026  00:54:45 +0000 (0:00:01.528)       0:05:10.419 ****** \n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'prometheus-server', 'value': {'container_name': 'prometheus_server', 'group': 'prometheus', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-v2-server:2024.2', 'volumes': ['/etc/kolla/prometheus-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'prometheus_v2:/var/lib/prometheus', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'prometheus_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9091', 'active_passive': True}, 'prometheus_server_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9091', 'listen_port': '9091', 'active_passive': True}}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2024.2', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-mysqld-exporter', 'value': {'container_name': 'prometheus_mysqld_exporter', 'group': 'prometheus-mysqld-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-mysqld-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-mysqld-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-memcached-exporter', 'value': {'container_name': 'prometheus_memcached_exporter', 'group': 'prometheus-memcached-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-memcached-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-memcached-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'prometheus-server', 'value': {'container_name': 'prometheus_server', 'group': 'prometheus', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-v2-server:2024.2', 'volumes': ['/etc/kolla/prometheus-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'prometheus_v2:/var/lib/prometheus', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'prometheus_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9091', 'active_passive': True}, 'prometheus_server_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9091', 'listen_port': '9091', 'active_passive': True}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'prometheus-server', 'value': {'container_name': 'prometheus_server', 'group': 'prometheus', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-v2-server:2024.2', 'volumes': ['/etc/kolla/prometheus-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'prometheus_v2:/var/lib/prometheus', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'prometheus_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9091', 'active_passive': True}, 'prometheus_server_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9091', 'listen_port': '9091', 'active_passive': True}}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2024.2', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2024.2', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2024.2', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-mysqld-exporter', 'value': {'container_name': 'prometheus_mysqld_exporter', 'group': 'prometheus-mysqld-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-mysqld-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-mysqld-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-memcached-exporter', 'value': {'container_name': 'prometheus_memcached_exporter', 'group': 'prometheus-memcached-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-memcached-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-memcached-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-mysqld-exporter', 'value': {'container_name': 'prometheus_mysqld_exporter', 'group': 'prometheus-mysqld-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-mysqld-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-mysqld-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2024.2', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-memcached-exporter', 'value': {'container_name': 'prometheus_memcached_exporter', 'group': 'prometheus-memcached-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-memcached-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-memcached-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2024.2', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'prometheus-alertmanager', 'value': {'container_name': 'prometheus_alertmanager', 'group': 'prometheus-alertmanager', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-alertmanager:2024.2', 'volumes': ['/etc/kolla/prometheus-alertmanager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'prometheus:/var/lib/prometheus'], 'dimensions': {}, 'haproxy': {'prometheus_alertmanager': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True}, 'prometheus_alertmanager_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9093', 'listen_port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True}}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-openstack-exporter', 'value': {'container_name': 'prometheus_openstack_exporter', 'group': 'prometheus-openstack-exporter', 'enabled': False, 'environment': {'OS_COMPUTE_API_VERSION': 'latest'}, 'image': 'registry.osism.tech/kolla/prometheus-openstack-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-openstack-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'prometheus_openstack_exporter': {'enabled': False, 'mode': 'http', 'external': False, 'port': '9198', 'backend_http_extra': ['timeout server 45s']}, 'prometheus_openstack_exporter_external': {'enabled': False, 'mode': 'http', 'external': True, 'port': '9198', 'backend_http_extra': ['timeout server 45s']}}}}) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'prometheus-alertmanager', 'value': {'container_name': 'prometheus_alertmanager', 'group': 'prometheus-alertmanager', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-alertmanager:2024.2', 'volumes': ['/etc/kolla/prometheus-alertmanager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'prometheus:/var/lib/prometheus'], 'dimensions': {}, 'haproxy': {'prometheus_alertmanager': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True}, 'prometheus_alertmanager_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9093', 'listen_port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True}}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-elasticsearch-exporter', 'value': {'container_name': 'prometheus_elasticsearch_exporter', 'group': 'prometheus-elasticsearch-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-elasticsearch-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-elasticsearch-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-blackbox-exporter', 'value': {'container_name': 'prometheus_blackbox_exporter', 'group': 'prometheus-blackbox-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-blackbox-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-blackbox-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-openstack-exporter', 'value': {'container_name': 'prometheus_openstack_exporter', 'group': 'prometheus-openstack-exporter', 'enabled': False, 'environment': {'OS_COMPUTE_API_VERSION': 'latest'}, 'image': 'registry.osism.tech/kolla/prometheus-openstack-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-openstack-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'prometheus_openstack_exporter': {'enabled': False, 'mode': 'http', 'external': False, 'port': '9198', 'backend_http_extra': ['timeout server 45s']}, 'prometheus_openstack_exporter_external': {'enabled': False, 'mode': 'http', 'external': True, 'port': '9198', 'backend_http_extra': ['timeout server 45s']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-libvirt-exporter', 'value': {'container_name': 'prometheus_libvirt_exporter', 'group': 'prometheus-libvirt-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-libvirt-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-libvirt-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run/libvirt:/run/libvirt:ro'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-elasticsearch-exporter', 'value': {'container_name': 'prometheus_elasticsearch_exporter', 'group': 'prometheus-elasticsearch-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-elasticsearch-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-elasticsearch-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-blackbox-exporter', 'value': {'container_name': 'prometheus_blackbox_exporter', 'group': 'prometheus-blackbox-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-blackbox-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-blackbox-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-libvirt-exporter', 'value': {'container_name': 'prometheus_libvirt_exporter', 'group': 'prometheus-libvirt-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-libvirt-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-libvirt-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run/libvirt:/run/libvirt:ro'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'prometheus-alertmanager', 'value': {'container_name': 'prometheus_alertmanager', 'group': 'prometheus-alertmanager', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-alertmanager:2024.2', 'volumes': ['/etc/kolla/prometheus-alertmanager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'prometheus:/var/lib/prometheus'], 'dimensions': {}, 'haproxy': {'prometheus_alertmanager': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True}, 'prometheus_alertmanager_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9093', 'listen_port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True}}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-openstack-exporter', 'value': {'container_name': 'prometheus_openstack_exporter', 'group': 'prometheus-openstack-exporter', 'enabled': False, 'environment': {'OS_COMPUTE_API_VERSION': 'latest'}, 'image': 'registry.osism.tech/kolla/prometheus-openstack-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-openstack-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'prometheus_openstack_exporter': {'enabled': False, 'mode': 'http', 'external': False, 'port': '9198', 'backend_http_extra': ['timeout server 45s']}, 'prometheus_openstack_exporter_external': {'enabled': False, 'mode': 'http', 'external': True, 'port': '9198', 'backend_http_extra': ['timeout server 45s']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-elasticsearch-exporter', 'value': {'container_name': 'prometheus_elasticsearch_exporter', 'group': 'prometheus-elasticsearch-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-elasticsearch-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-elasticsearch-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-blackbox-exporter', 'value': {'container_name': 'prometheus_blackbox_exporter', 'group': 'prometheus-blackbox-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-blackbox-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-blackbox-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-libvirt-exporter', 'value': {'container_name': 'prometheus_libvirt_exporter', 'group': 'prometheus-libvirt-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-libvirt-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-libvirt-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run/libvirt:/run/libvirt:ro'], 'dimensions': {}}}) \u001b[0m\n\nTASK [haproxy-config : Add configuration for prometheus when using single external frontend] ***\nTuesday 24 February 2026  00:54:50 +0000 (0:00:05.244)       0:05:15.664 ****** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-server', 'value': {'container_name': 'prometheus_server', 'group': 'prometheus', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-v2-server:2024.2', 'volumes': ['/etc/kolla/prometheus-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'prometheus_v2:/var/lib/prometheus', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'prometheus_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9091', 'active_passive': True}, 'prometheus_server_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9091', 'listen_port': '9091', 'active_passive': True}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2024.2', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-mysqld-exporter', 'value': {'container_name': 'prometheus_mysqld_exporter', 'group': 'prometheus-mysqld-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-mysqld-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-mysqld-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-memcached-exporter', 'value': {'container_name': 'prometheus_memcached_exporter', 'group': 'prometheus-memcached-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-memcached-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-memcached-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2024.2', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-alertmanager', 'value': {'container_name': 'prometheus_alertmanager', 'group': 'prometheus-alertmanager', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-alertmanager:2024.2', 'volumes': ['/etc/kolla/prometheus-alertmanager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'prometheus:/var/lib/prometheus'], 'dimensions': {}, 'haproxy': {'prometheus_alertmanager': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True}, 'prometheus_alertmanager_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9093', 'listen_port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-openstack-exporter', 'value': {'container_name': 'prometheus_openstack_exporter', 'group': 'prometheus-openstack-exporter', 'enabled': False, 'environment': {'OS_COMPUTE_API_VERSION': 'latest'}, 'image': 'registry.osism.tech/kolla/prometheus-openstack-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-openstack-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'prometheus_openstack_exporter': {'enabled': False, 'mode': 'http', 'external': False, 'port': '9198', 'backend_http_extra': ['timeout server 45s']}, 'prometheus_openstack_exporter_external': {'enabled': False, 'mode': 'http', 'external': True, 'port': '9198', 'backend_http_extra': ['timeout server 45s']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-elasticsearch-exporter', 'value': {'container_name': 'prometheus_elasticsearch_exporter', 'group': 'prometheus-elasticsearch-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-elasticsearch-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-elasticsearch-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-blackbox-exporter', 'value': {'container_name': 'prometheus_blackbox_exporter', 'group': 'prometheus-blackbox-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-blackbox-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-blackbox-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-libvirt-exporter', 'value': {'container_name': 'prometheus_libvirt_exporter', 'group': 'prometheus-libvirt-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-libvirt-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-libvirt-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run/libvirt:/run/libvirt:ro'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-server', 'value': {'container_name': 'prometheus_server', 'group': 'prometheus', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-v2-server:2024.2', 'volumes': ['/etc/kolla/prometheus-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'prometheus_v2:/var/lib/prometheus', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'prometheus_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9091', 'active_passive': True}, 'prometheus_server_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9091', 'listen_port': '9091', 'active_passive': True}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2024.2', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-mysqld-exporter', 'value': {'container_name': 'prometheus_mysqld_exporter', 'group': 'prometheus-mysqld-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-mysqld-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-mysqld-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-memcached-exporter', 'value': {'container_name': 'prometheus_memcached_exporter', 'group': 'prometheus-memcached-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-memcached-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-memcached-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2024.2', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-alertmanager', 'value': {'container_name': 'prometheus_alertmanager', 'group': 'prometheus-alertmanager', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-alertmanager:2024.2', 'volumes': ['/etc/kolla/prometheus-alertmanager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'prometheus:/var/lib/prometheus'], 'dimensions': {}, 'haproxy': {'prometheus_alertmanager': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True}, 'prometheus_alertmanager_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9093', 'listen_port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-openstack-exporter', 'value': {'container_name': 'prometheus_openstack_exporter', 'group': 'prometheus-openstack-exporter', 'enabled': False, 'environment': {'OS_COMPUTE_API_VERSION': 'latest'}, 'image': 'registry.osism.tech/kolla/prometheus-openstack-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-openstack-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'prometheus_openstack_exporter': {'enabled': False, 'mode': 'http', 'external': False, 'port': '9198', 'backend_http_extra': ['timeout server 45s']}, 'prometheus_openstack_exporter_external': {'enabled': False, 'mode': 'http', 'external': True, 'port': '9198', 'backend_http_extra': ['timeout server 45s']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-elasticsearch-exporter', 'value': {'container_name': 'prometheus_elasticsearch_exporter', 'group': 'prometheus-elasticsearch-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-elasticsearch-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-elasticsearch-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-blackbox-exporter', 'value': {'container_name': 'prometheus_blackbox_exporter', 'group': 'prometheus-blackbox-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-blackbox-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-blackbox-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-libvirt-exporter', 'value': {'container_name': 'prometheus_libvirt_exporter', 'group': 'prometheus-libvirt-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-libvirt-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-libvirt-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run/libvirt:/run/libvirt:ro'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-server', 'value': {'container_name': 'prometheus_server', 'group': 'prometheus', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-v2-server:2024.2', 'volumes': ['/etc/kolla/prometheus-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'prometheus_v2:/var/lib/prometheus', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'prometheus_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9091', 'active_passive': True}, 'prometheus_server_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9091', 'listen_port': '9091', 'active_passive': True}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2024.2', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-mysqld-exporter', 'value': {'container_name': 'prometheus_mysqld_exporter', 'group': 'prometheus-mysqld-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-mysqld-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-mysqld-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-memcached-exporter', 'value': {'container_name': 'prometheus_memcached_exporter', 'group': 'prometheus-memcached-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-memcached-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-memcached-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2024.2', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-alertmanager', 'value': {'container_name': 'prometheus_alertmanager', 'group': 'prometheus-alertmanager', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-alertmanager:2024.2', 'volumes': ['/etc/kolla/prometheus-alertmanager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'prometheus:/var/lib/prometheus'], 'dimensions': {}, 'haproxy': {'prometheus_alertmanager': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True}, 'prometheus_alertmanager_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9093', 'listen_port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-openstack-exporter', 'value': {'container_name': 'prometheus_openstack_exporter', 'group': 'prometheus-openstack-exporter', 'enabled': False, 'environment': {'OS_COMPUTE_API_VERSION': 'latest'}, 'image': 'registry.osism.tech/kolla/prometheus-openstack-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-openstack-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'prometheus_openstack_exporter': {'enabled': False, 'mode': 'http', 'external': False, 'port': '9198', 'backend_http_extra': ['timeout server 45s']}, 'prometheus_openstack_exporter_external': {'enabled': False, 'mode': 'http', 'external': True, 'port': '9198', 'backend_http_extra': ['timeout server 45s']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-elasticsearch-exporter', 'value': {'container_name': 'prometheus_elasticsearch_exporter', 'group': 'prometheus-elasticsearch-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-elasticsearch-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-elasticsearch-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-blackbox-exporter', 'value': {'container_name': 'prometheus_blackbox_exporter', 'group': 'prometheus-blackbox-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-blackbox-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-blackbox-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-libvirt-exporter', 'value': {'container_name': 'prometheus_libvirt_exporter', 'group': 'prometheus-libvirt-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-libvirt-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-libvirt-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run/libvirt:/run/libvirt:ro'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [haproxy-config : Configuring firewall for prometheus] ********************\nTuesday 24 February 2026  00:54:51 +0000 (0:00:01.064)       0:05:16.728 ****** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus_server', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9091', 'active_passive': True}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus_server_external', 'value': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9091', 'listen_port': '9091', 'active_passive': True}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus_alertmanager', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus_alertmanager_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9093', 'listen_port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus_server', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9091', 'active_passive': True}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus_server_external', 'value': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9091', 'listen_port': '9091', 'active_passive': True}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus_alertmanager', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus_alertmanager_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9093', 'listen_port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus_server', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9091', 'active_passive': True}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus_server_external', 'value': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9091', 'listen_port': '9091', 'active_passive': True}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus_alertmanager', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus_alertmanager_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9093', 'listen_port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [proxysql-config : Copying over prometheus ProxySQL users config] *********\nTuesday 24 February 2026  00:54:52 +0000 (0:00:01.055)       0:05:17.784 ****** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [proxysql-config : Copying over prometheus ProxySQL rules config] *********\nTuesday 24 February 2026  00:54:53 +0000 (0:00:00.448)       0:05:18.232 ****** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [include_role : rabbitmq] *************************************************\nTuesday 24 February 2026  00:54:54 +0000 (0:00:01.527)       0:05:19.760 ****** \n\u001b[0;36mincluded: rabbitmq for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [haproxy-config : Copying over rabbitmq haproxy config] *******************\nTuesday 24 February 2026  00:54:56 +0000 (0:00:01.910)       0:05:21.670 ****** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'rabbitmq', 'value': {'container_name': 'rabbitmq', 'group': None, 'enabled': True, 'image': 'registry.osism.tech/kolla/rabbitmq:2024.2', 'bootstrap_environment': {'KOLLA_BOOTSTRAP': None, 'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': None, 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': None, 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'volumes': ['/etc/kolla/rabbitmq/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'rabbitmq:/var/lib/rabbitmq/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_rabbitmq'], 'timeout': '30'}, 'haproxy': {'rabbitmq_management': {'enabled': 'yes', 'mode': 'http', 'port': '15672', 'host_group': 'rabbitmq'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'rabbitmq', 'value': {'container_name': 'rabbitmq', 'group': None, 'enabled': True, 'image': 'registry.osism.tech/kolla/rabbitmq:2024.2', 'bootstrap_environment': {'KOLLA_BOOTSTRAP': None, 'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': None, 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': None, 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'volumes': ['/etc/kolla/rabbitmq/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'rabbitmq:/var/lib/rabbitmq/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_rabbitmq'], 'timeout': '30'}, 'haproxy': {'rabbitmq_management': {'enabled': 'yes', 'mode': 'http', 'port': '15672', 'host_group': 'rabbitmq'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'rabbitmq', 'value': {'container_name': 'rabbitmq', 'group': None, 'enabled': True, 'image': 'registry.osism.tech/kolla/rabbitmq:2024.2', 'bootstrap_environment': {'KOLLA_BOOTSTRAP': None, 'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': None, 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': None, 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'volumes': ['/etc/kolla/rabbitmq/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'rabbitmq:/var/lib/rabbitmq/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_rabbitmq'], 'timeout': '30'}, 'haproxy': {'rabbitmq_management': {'enabled': 'yes', 'mode': 'http', 'port': '15672', 'host_group': 'rabbitmq'}}}})\u001b[0m\n\nTASK [haproxy-config : Add configuration for rabbitmq when using single external frontend] ***\nTuesday 24 February 2026  00:54:59 +0000 (0:00:02.644)       0:05:24.315 ****** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'rabbitmq', 'value': {'container_name': 'rabbitmq', 'group': None, 'enabled': True, 'image': 'registry.osism.tech/kolla/rabbitmq:2024.2', 'bootstrap_environment': {'KOLLA_BOOTSTRAP': None, 'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': None, 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': None, 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'volumes': ['/etc/kolla/rabbitmq/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'rabbitmq:/var/lib/rabbitmq/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_rabbitmq'], 'timeout': '30'}, 'haproxy': {'rabbitmq_management': {'enabled': 'yes', 'mode': 'http', 'port': '15672', 'host_group': 'rabbitmq'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'rabbitmq', 'value': {'container_name': 'rabbitmq', 'group': None, 'enabled': True, 'image': 'registry.osism.tech/kolla/rabbitmq:2024.2', 'bootstrap_environment': {'KOLLA_BOOTSTRAP': None, 'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': None, 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': None, 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'volumes': ['/etc/kolla/rabbitmq/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'rabbitmq:/var/lib/rabbitmq/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_rabbitmq'], 'timeout': '30'}, 'haproxy': {'rabbitmq_management': {'enabled': 'yes', 'mode': 'http', 'port': '15672', 'host_group': 'rabbitmq'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'rabbitmq', 'value': {'container_name': 'rabbitmq', 'group': None, 'enabled': True, 'image': 'registry.osism.tech/kolla/rabbitmq:2024.2', 'bootstrap_environment': {'KOLLA_BOOTSTRAP': None, 'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': None, 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': None, 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'volumes': ['/etc/kolla/rabbitmq/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'rabbitmq:/var/lib/rabbitmq/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_rabbitmq'], 'timeout': '30'}, 'haproxy': {'rabbitmq_management': {'enabled': 'yes', 'mode': 'http', 'port': '15672', 'host_group': 'rabbitmq'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [haproxy-config : Configuring firewall for rabbitmq] **********************\nTuesday 24 February 2026  00:55:00 +0000 (0:00:00.969)       0:05:25.285 ****** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'rabbitmq_management', 'value': {'enabled': 'yes', 'mode': 'http', 'port': '15672', 'host_group': 'rabbitmq'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'rabbitmq_management', 'value': {'enabled': 'yes', 'mode': 'http', 'port': '15672', 'host_group': 'rabbitmq'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'rabbitmq_management', 'value': {'enabled': 'yes', 'mode': 'http', 'port': '15672', 'host_group': 'rabbitmq'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [proxysql-config : Copying over rabbitmq ProxySQL users config] ***********\nTuesday 24 February 2026  00:55:01 +0000 (0:00:00.773)       0:05:26.058 ****** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [proxysql-config : Copying over rabbitmq ProxySQL rules config] ***********\nTuesday 24 February 2026  00:55:01 +0000 (0:00:00.547)       0:05:26.606 ****** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [include_role : skyline] **************************************************\nTuesday 24 February 2026  00:55:03 +0000 (0:00:01.474)       0:05:28.080 ****** \n\u001b[0;36mincluded: skyline for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [haproxy-config : Copying over skyline haproxy config] ********************\nTuesday 24 February 2026  00:55:05 +0000 (0:00:02.040)       0:05:30.121 ****** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'skyline-apiserver', 'value': {'container_name': 'skyline_apiserver', 'group': 'skyline-apiserver', 'enabled': True, 'image': 'registry.osism.tech/kolla/skyline-apiserver:2024.2', 'volumes': ['/etc/kolla/skyline-apiserver/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9998/docs'], 'timeout': '30'}, 'haproxy': {'skyline_apiserver': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no'}, 'skyline_apiserver_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'skyline-apiserver', 'value': {'container_name': 'skyline_apiserver', 'group': 'skyline-apiserver', 'enabled': True, 'image': 'registry.osism.tech/kolla/skyline-apiserver:2024.2', 'volumes': ['/etc/kolla/skyline-apiserver/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9998/docs'], 'timeout': '30'}, 'haproxy': {'skyline_apiserver': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no'}, 'skyline_apiserver_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'skyline-apiserver', 'value': {'container_name': 'skyline_apiserver', 'group': 'skyline-apiserver', 'enabled': True, 'image': 'registry.osism.tech/kolla/skyline-apiserver:2024.2', 'volumes': ['/etc/kolla/skyline-apiserver/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9998/docs'], 'timeout': '30'}, 'haproxy': {'skyline_apiserver': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no'}, 'skyline_apiserver_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'skyline-console', 'value': {'container_name': 'skyline_console', 'group': 'skyline-console', 'enabled': True, 'image': 'registry.osism.tech/kolla/skyline-console:2024.2', 'volumes': ['/etc/kolla/skyline-console/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9999/docs'], 'timeout': '30'}, 'haproxy': {'skyline_console': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no'}, 'skyline_console_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'skyline-console', 'value': {'container_name': 'skyline_console', 'group': 'skyline-console', 'enabled': True, 'image': 'registry.osism.tech/kolla/skyline-console:2024.2', 'volumes': ['/etc/kolla/skyline-console/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9999/docs'], 'timeout': '30'}, 'haproxy': {'skyline_console': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no'}, 'skyline_console_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'skyline-console', 'value': {'container_name': 'skyline_console', 'group': 'skyline-console', 'enabled': True, 'image': 'registry.osism.tech/kolla/skyline-console:2024.2', 'volumes': ['/etc/kolla/skyline-console/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9999/docs'], 'timeout': '30'}, 'haproxy': {'skyline_console': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no'}, 'skyline_console_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no'}}}})\u001b[0m\n\nTASK [haproxy-config : Add configuration for skyline when using single external frontend] ***\nTuesday 24 February 2026  00:55:12 +0000 (0:00:07.155)       0:05:37.277 ****** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'skyline-apiserver', 'value': {'container_name': 'skyline_apiserver', 'group': 'skyline-apiserver', 'enabled': True, 'image': 'registry.osism.tech/kolla/skyline-apiserver:2024.2', 'volumes': ['/etc/kolla/skyline-apiserver/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9998/docs'], 'timeout': '30'}, 'haproxy': {'skyline_apiserver': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no'}, 'skyline_apiserver_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'skyline-console', 'value': {'container_name': 'skyline_console', 'group': 'skyline-console', 'enabled': True, 'image': 'registry.osism.tech/kolla/skyline-console:2024.2', 'volumes': ['/etc/kolla/skyline-console/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9999/docs'], 'timeout': '30'}, 'haproxy': {'skyline_console': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no'}, 'skyline_console_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'skyline-apiserver', 'value': {'container_name': 'skyline_apiserver', 'group': 'skyline-apiserver', 'enabled': True, 'image': 'registry.osism.tech/kolla/skyline-apiserver:2024.2', 'volumes': ['/etc/kolla/skyline-apiserver/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9998/docs'], 'timeout': '30'}, 'haproxy': {'skyline_apiserver': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no'}, 'skyline_apiserver_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'skyline-console', 'value': {'container_name': 'skyline_console', 'group': 'skyline-console', 'enabled': True, 'image': 'registry.osism.tech/kolla/skyline-console:2024.2', 'volumes': ['/etc/kolla/skyline-console/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9999/docs'], 'timeout': '30'}, 'haproxy': {'skyline_console': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no'}, 'skyline_console_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'skyline-apiserver', 'value': {'container_name': 'skyline_apiserver', 'group': 'skyline-apiserver', 'enabled': True, 'image': 'registry.osism.tech/kolla/skyline-apiserver:2024.2', 'volumes': ['/etc/kolla/skyline-apiserver/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9998/docs'], 'timeout': '30'}, 'haproxy': {'skyline_apiserver': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no'}, 'skyline_apiserver_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'skyline-console', 'value': {'container_name': 'skyline_console', 'group': 'skyline-console', 'enabled': True, 'image': 'registry.osism.tech/kolla/skyline-console:2024.2', 'volumes': ['/etc/kolla/skyline-console/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9999/docs'], 'timeout': '30'}, 'haproxy': {'skyline_console': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no'}, 'skyline_console_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [haproxy-config : Configuring firewall for skyline] ***********************\nTuesday 24 February 2026  00:55:13 +0000 (0:00:00.650)       0:05:37.927 ****** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'skyline_apiserver', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'skyline_apiserver_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'skyline_console', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'skyline_console_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'skyline_apiserver', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'skyline_apiserver_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'skyline_console', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'skyline_console_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'skyline_apiserver', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'skyline_apiserver_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'skyline_console', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'skyline_console_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [proxysql-config : Copying over skyline ProxySQL users config] ************\nTuesday 24 February 2026  00:55:14 +0000 (0:00:01.723)       0:05:39.650 ****** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [proxysql-config : Copying over skyline ProxySQL rules config] ************\nTuesday 24 February 2026  00:55:16 +0000 (0:00:01.374)       0:05:41.025 ****** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [include_role : swift] ****************************************************\nTuesday 24 February 2026  00:55:18 +0000 (0:00:02.216)       0:05:43.242 ****** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [include_role : tacker] ***************************************************\nTuesday 24 February 2026  00:55:18 +0000 (0:00:00.374)       0:05:43.617 ****** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [include_role : trove] ****************************************************\nTuesday 24 February 2026  00:55:19 +0000 (0:00:00.337)       0:05:43.954 ****** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [include_role : venus] ****************************************************\nTuesday 24 February 2026  00:55:19 +0000 (0:00:00.684)       0:05:44.639 ****** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [include_role : watcher] **************************************************\nTuesday 24 February 2026  00:55:20 +0000 (0:00:00.357)       0:05:44.997 ****** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [include_role : zun] ******************************************************\nTuesday 24 February 2026  00:55:20 +0000 (0:00:00.374)       0:05:45.372 ****** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [loadbalancer : Check IP addresses on the API interface] *******\nTuesday 24 February 2026  00:55:21 +0000 (0:00:01.159)       0:05:46.532 ****** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [loadbalancer : Group HA nodes by status] **********************\nTuesday 24 February 2026  00:55:22 +0000 (0:00:00.784)       0:05:47.316 ****** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [loadbalancer : Stop backup keepalived container] **************\nTuesday 24 February 2026  00:55:22 +0000 (0:00:00.401)       0:05:47.718 ****** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [loadbalancer : Stop backup haproxy container] *****************\nTuesday 24 February 2026  00:55:23 +0000 (0:00:00.974)       0:05:48.693 ****** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [loadbalancer : Stop backup proxysql container] ****************\nTuesday 24 February 2026  00:55:25 +0000 (0:00:01.285)       0:05:49.978 ****** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [loadbalancer : Start backup haproxy container] ****************\nTuesday 24 February 2026  00:55:25 +0000 (0:00:00.876)       0:05:50.855 ****** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [loadbalancer : Wait for backup haproxy to start] **************\nTuesday 24 February 2026  00:55:35 +0000 (0:00:09.869)       0:06:00.725 ****** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [loadbalancer : Start backup proxysql container] ***************\nTuesday 24 February 2026  00:55:36 +0000 (0:00:00.861)       0:06:01.587 ****** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [loadbalancer : Wait for backup proxysql to start] *************\nTuesday 24 February 2026  00:55:52 +0000 (0:00:15.529)       0:06:17.117 ****** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [loadbalancer : Start backup keepalived container] *************\nTuesday 24 February 2026  00:55:53 +0000 (0:00:00.898)       0:06:18.015 ****** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [loadbalancer : Stop master haproxy container] *****************\nTuesday 24 February 2026  00:55:58 +0000 (0:00:05.413)       0:06:23.429 ****** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [loadbalancer : Stop master proxysql container] ****************\nTuesday 24 February 2026  00:55:58 +0000 (0:00:00.452)       0:06:23.882 ****** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [loadbalancer : Stop master keepalived container] **************\nTuesday 24 February 2026  00:55:59 +0000 (0:00:00.783)       0:06:24.665 ****** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [loadbalancer : Start master haproxy container] ****************\nTuesday 24 February 2026  00:56:00 +0000 (0:00:00.448)       0:06:25.114 ****** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [loadbalancer : Start master proxysql container] ***************\nTuesday 24 February 2026  00:56:00 +0000 (0:00:00.436)       0:06:25.550 ****** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [loadbalancer : Start master keepalived container] *************\nTuesday 24 February 2026  00:56:01 +0000 (0:00:00.385)       0:06:25.935 ****** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [loadbalancer : Wait for haproxy to listen on VIP] *************\nTuesday 24 February 2026  00:56:01 +0000 (0:00:00.391)       0:06:26.327 ****** \n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nRUNNING HANDLER [loadbalancer : Wait for proxysql to listen on VIP] ************\nTuesday 24 February 2026  00:56:06 +0000 (0:00:05.365)       0:06:31.693 ****** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=123 \u001b[0m \u001b[0;33mchanged=76  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=97  \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=122 \u001b[0m \u001b[0;33mchanged=76  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=97  \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=122 \u001b[0m \u001b[0;33mchanged=76  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=97  \u001b[0m rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nTuesday 24 February 2026  00:56:07 +0000 (0:00:00.904)       0:06:32.597 ****** \n=============================================================================== \nloadbalancer : Start backup proxysql container ------------------------- 15.53s\nloadbalancer : Start backup haproxy container --------------------------- 9.87s\nhaproxy-config : Copying over skyline haproxy config -------------------- 7.16s\nhaproxy-config : Copying over opensearch haproxy config ----------------- 6.12s\nloadbalancer : Start backup keepalived container ------------------------ 5.41s\nloadbalancer : Wait for haproxy to listen on VIP ------------------------ 5.37s\nhaproxy-config : Copying over prometheus haproxy config ----------------- 5.24s\nloadbalancer : Copying over proxysql config ----------------------------- 4.95s\nhaproxy-config : Copying over keystone haproxy config ------------------- 4.90s\nhaproxy-config : Copying over nova-cell:nova-novncproxy haproxy config --- 4.79s\nhaproxy-config : Copying over aodh haproxy config ----------------------- 4.64s\nhaproxy-config : Copying over designate haproxy config ------------------ 4.59s\nhaproxy-config : Copying over glance haproxy config --------------------- 4.54s\nhaproxy-config : Copying over nova haproxy config ----------------------- 4.50s\nhaproxy-config : Copying over manila haproxy config --------------------- 4.34s\nhaproxy-config : Copying over grafana haproxy config -------------------- 4.21s\nhaproxy-config : Copying over neutron haproxy config -------------------- 4.14s\nhaproxy-config : Copying over horizon haproxy config -------------------- 4.06s\nhaproxy-config : Copying over octavia haproxy config -------------------- 3.85s\nhaproxy-config : Copying over barbican haproxy config ------------------- 3.79s\n\u001b[32m2026-02-24 00:56:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 00:56:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ca6226d5-0f70-4445-9f9f-dca4faf1e39d is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:56:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:56:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 21cb4ba5-3e33-4fe4-8954-1e96067dfc1b is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:56:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 00:56:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ca6226d5-0f70-4445-9f9f-dca4faf1e39d is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:56:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:56:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 21cb4ba5-3e33-4fe4-8954-1e96067dfc1b is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:56:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 00:56:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ca6226d5-0f70-4445-9f9f-dca4faf1e39d is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:56:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:56:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 21cb4ba5-3e33-4fe4-8954-1e96067dfc1b is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:56:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 00:56:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ca6226d5-0f70-4445-9f9f-dca4faf1e39d is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:56:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:56:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 21cb4ba5-3e33-4fe4-8954-1e96067dfc1b is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:56:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 00:56:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ca6226d5-0f70-4445-9f9f-dca4faf1e39d is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:56:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:56:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 21cb4ba5-3e33-4fe4-8954-1e96067dfc1b is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:56:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 00:56:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ca6226d5-0f70-4445-9f9f-dca4faf1e39d is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:56:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:56:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 21cb4ba5-3e33-4fe4-8954-1e96067dfc1b is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:56:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 00:56:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ca6226d5-0f70-4445-9f9f-dca4faf1e39d is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:56:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:56:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 21cb4ba5-3e33-4fe4-8954-1e96067dfc1b is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:56:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 00:56:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ca6226d5-0f70-4445-9f9f-dca4faf1e39d is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:56:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:56:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 21cb4ba5-3e33-4fe4-8954-1e96067dfc1b is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:56:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 00:56:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ca6226d5-0f70-4445-9f9f-dca4faf1e39d is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:56:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:56:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 21cb4ba5-3e33-4fe4-8954-1e96067dfc1b is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:56:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 00:56:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ca6226d5-0f70-4445-9f9f-dca4faf1e39d is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:56:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:56:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 21cb4ba5-3e33-4fe4-8954-1e96067dfc1b is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:56:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 00:56:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ca6226d5-0f70-4445-9f9f-dca4faf1e39d is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:56:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:56:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 21cb4ba5-3e33-4fe4-8954-1e96067dfc1b is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:56:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 00:56:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ca6226d5-0f70-4445-9f9f-dca4faf1e39d is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:56:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:56:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 21cb4ba5-3e33-4fe4-8954-1e96067dfc1b is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:56:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 00:56:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ca6226d5-0f70-4445-9f9f-dca4faf1e39d is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:56:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:56:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 21cb4ba5-3e33-4fe4-8954-1e96067dfc1b is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:56:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 00:56:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ca6226d5-0f70-4445-9f9f-dca4faf1e39d is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:56:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:56:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 21cb4ba5-3e33-4fe4-8954-1e96067dfc1b is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:56:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 00:56:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ca6226d5-0f70-4445-9f9f-dca4faf1e39d is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:56:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:56:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 21cb4ba5-3e33-4fe4-8954-1e96067dfc1b is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:56:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 00:56:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ca6226d5-0f70-4445-9f9f-dca4faf1e39d is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:56:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:56:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 21cb4ba5-3e33-4fe4-8954-1e96067dfc1b is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:56:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 00:57:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ca6226d5-0f70-4445-9f9f-dca4faf1e39d is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:57:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:57:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 21cb4ba5-3e33-4fe4-8954-1e96067dfc1b is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:57:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 00:57:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ca6226d5-0f70-4445-9f9f-dca4faf1e39d is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:57:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:57:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 21cb4ba5-3e33-4fe4-8954-1e96067dfc1b is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:57:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 00:57:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ca6226d5-0f70-4445-9f9f-dca4faf1e39d is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:57:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:57:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 21cb4ba5-3e33-4fe4-8954-1e96067dfc1b is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:57:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 00:57:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ca6226d5-0f70-4445-9f9f-dca4faf1e39d is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:57:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:57:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 21cb4ba5-3e33-4fe4-8954-1e96067dfc1b is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:57:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 00:57:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ca6226d5-0f70-4445-9f9f-dca4faf1e39d is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:57:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:57:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 21cb4ba5-3e33-4fe4-8954-1e96067dfc1b is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:57:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 00:57:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ca6226d5-0f70-4445-9f9f-dca4faf1e39d is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:57:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:57:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 21cb4ba5-3e33-4fe4-8954-1e96067dfc1b is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:57:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 00:57:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ca6226d5-0f70-4445-9f9f-dca4faf1e39d is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:57:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:57:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 21cb4ba5-3e33-4fe4-8954-1e96067dfc1b is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:57:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 00:57:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ca6226d5-0f70-4445-9f9f-dca4faf1e39d is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:57:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:57:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 21cb4ba5-3e33-4fe4-8954-1e96067dfc1b is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:57:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 00:57:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ca6226d5-0f70-4445-9f9f-dca4faf1e39d is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:57:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:57:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 21cb4ba5-3e33-4fe4-8954-1e96067dfc1b is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:57:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 00:57:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ca6226d5-0f70-4445-9f9f-dca4faf1e39d is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:57:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:57:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 21cb4ba5-3e33-4fe4-8954-1e96067dfc1b is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:57:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 00:57:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ca6226d5-0f70-4445-9f9f-dca4faf1e39d is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:57:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:57:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 21cb4ba5-3e33-4fe4-8954-1e96067dfc1b is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:57:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 00:57:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ca6226d5-0f70-4445-9f9f-dca4faf1e39d is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:57:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:57:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 21cb4ba5-3e33-4fe4-8954-1e96067dfc1b is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:57:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 00:57:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ca6226d5-0f70-4445-9f9f-dca4faf1e39d is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:57:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:57:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 21cb4ba5-3e33-4fe4-8954-1e96067dfc1b is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:57:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 00:57:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ca6226d5-0f70-4445-9f9f-dca4faf1e39d is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:57:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:57:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 21cb4ba5-3e33-4fe4-8954-1e96067dfc1b is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:57:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 00:57:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ca6226d5-0f70-4445-9f9f-dca4faf1e39d is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:57:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:57:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 21cb4ba5-3e33-4fe4-8954-1e96067dfc1b is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:57:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 00:57:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ca6226d5-0f70-4445-9f9f-dca4faf1e39d is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:57:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:57:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 21cb4ba5-3e33-4fe4-8954-1e96067dfc1b is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:57:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 00:57:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ca6226d5-0f70-4445-9f9f-dca4faf1e39d is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:57:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:57:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 21cb4ba5-3e33-4fe4-8954-1e96067dfc1b is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:57:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 00:57:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ca6226d5-0f70-4445-9f9f-dca4faf1e39d is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:57:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:57:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 21cb4ba5-3e33-4fe4-8954-1e96067dfc1b is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:57:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 00:57:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ca6226d5-0f70-4445-9f9f-dca4faf1e39d is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:57:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:57:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 21cb4ba5-3e33-4fe4-8954-1e96067dfc1b is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:57:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 00:57:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ca6226d5-0f70-4445-9f9f-dca4faf1e39d is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:57:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:57:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 21cb4ba5-3e33-4fe4-8954-1e96067dfc1b is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:57:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 00:58:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ca6226d5-0f70-4445-9f9f-dca4faf1e39d is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:58:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:58:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 21cb4ba5-3e33-4fe4-8954-1e96067dfc1b is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:58:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 00:58:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ca6226d5-0f70-4445-9f9f-dca4faf1e39d is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:58:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:58:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 21cb4ba5-3e33-4fe4-8954-1e96067dfc1b is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:58:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 00:58:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ca6226d5-0f70-4445-9f9f-dca4faf1e39d is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:58:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:58:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 21cb4ba5-3e33-4fe4-8954-1e96067dfc1b is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:58:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 00:58:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ca6226d5-0f70-4445-9f9f-dca4faf1e39d is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:58:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:58:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 21cb4ba5-3e33-4fe4-8954-1e96067dfc1b is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:58:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 00:58:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ca6226d5-0f70-4445-9f9f-dca4faf1e39d is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:58:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:58:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 21cb4ba5-3e33-4fe4-8954-1e96067dfc1b is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:58:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 00:58:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ca6226d5-0f70-4445-9f9f-dca4faf1e39d is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:58:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:58:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 21cb4ba5-3e33-4fe4-8954-1e96067dfc1b is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:58:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 00:58:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ca6226d5-0f70-4445-9f9f-dca4faf1e39d is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:58:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:58:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 21cb4ba5-3e33-4fe4-8954-1e96067dfc1b is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:58:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 00:58:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ca6226d5-0f70-4445-9f9f-dca4faf1e39d is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:58:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:58:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 21cb4ba5-3e33-4fe4-8954-1e96067dfc1b is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:58:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 00:58:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ca6226d5-0f70-4445-9f9f-dca4faf1e39d is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:58:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:58:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 21cb4ba5-3e33-4fe4-8954-1e96067dfc1b is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:58:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 00:58:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ca6226d5-0f70-4445-9f9f-dca4faf1e39d is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:58:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:58:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 21cb4ba5-3e33-4fe4-8954-1e96067dfc1b is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:58:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 00:58:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ca6226d5-0f70-4445-9f9f-dca4faf1e39d is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:58:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:58:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 21cb4ba5-3e33-4fe4-8954-1e96067dfc1b is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:58:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 00:58:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ca6226d5-0f70-4445-9f9f-dca4faf1e39d is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:58:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:58:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 21cb4ba5-3e33-4fe4-8954-1e96067dfc1b is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:58:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 00:58:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ca6226d5-0f70-4445-9f9f-dca4faf1e39d is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:58:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:58:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 21cb4ba5-3e33-4fe4-8954-1e96067dfc1b is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:58:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 00:58:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ca6226d5-0f70-4445-9f9f-dca4faf1e39d is in state STARTED\u001b[0m\n\n\u001b[1;35m[WARNING]: Collection community.general does not support Ansible version\u001b[0m\n\u001b[1;35m2.16.14\u001b[0m\n\nPLAY [Prepare deployment of Ceph services] *************************************\n\nTASK [ceph-facts : Include facts.yml] ******************************************\nTuesday 24 February 2026  00:47:15 +0000 (0:00:00.702)       0:00:00.702 ****** \n\u001b[0;36mincluded: /ansible/roles/ceph-facts/tasks/facts.yml for testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [ceph-facts : Check if it is atomic host] *********************************\nTuesday 24 February 2026  00:47:16 +0000 (0:00:01.114)       0:00:01.817 ****** \n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\nTASK [ceph-facts : Set_fact is_atomic] *****************************************\nTuesday 24 February 2026  00:47:18 +0000 (0:00:01.734)       0:00:03.551 ****** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-facts : Check if podman binary is present] **************************\nTuesday 24 February 2026  00:47:19 +0000 (0:00:01.156)       0:00:04.708 ****** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-facts : Set_fact container_binary] **********************************\nTuesday 24 February 2026  00:47:20 +0000 (0:00:01.104)       0:00:05.813 ****** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-facts : Set_fact ceph_cmd] ******************************************\nTuesday 24 February 2026  00:47:21 +0000 (0:00:00.859)       0:00:06.672 ****** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-facts : Set_fact discovered_interpreter_python] *********************\nTuesday 24 February 2026  00:47:21 +0000 (0:00:00.591)       0:00:07.263 ****** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-facts : Set_fact discovered_interpreter_python if not previously set] ***\nTuesday 24 February 2026  00:47:22 +0000 (0:00:00.780)       0:00:08.044 ****** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-facts : Set_fact ceph_release ceph_stable_release] ******************\nTuesday 24 February 2026  00:47:23 +0000 (0:00:00.963)       0:00:09.008 ****** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-facts : Set_fact monitor_name ansible_facts['hostname']] ************\nTuesday 24 February 2026  00:47:25 +0000 (0:00:01.446)       0:00:10.454 ****** \n\u001b[0;32mok: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item=testbed-node-0)\u001b[0m\n\u001b[0;32mok: [testbed-node-3 -> testbed-node-1(192.168.16.11)] => (item=testbed-node-1)\u001b[0m\n\u001b[0;32mok: [testbed-node-3 -> testbed-node-2(192.168.16.12)] => (item=testbed-node-2)\u001b[0m\n\nTASK [ceph-facts : Set_fact container_exec_cmd] ********************************\nTuesday 24 February 2026  00:47:25 +0000 (0:00:00.596)       0:00:11.050 ****** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-facts : Find a running mon container] *******************************\nTuesday 24 February 2026  00:47:26 +0000 (0:00:01.213)       0:00:12.264 ****** \n\u001b[0;32mok: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item=testbed-node-0)\u001b[0m\n\u001b[0;32mok: [testbed-node-3 -> testbed-node-1(192.168.16.11)] => (item=testbed-node-1)\u001b[0m\n\u001b[0;32mok: [testbed-node-3 -> testbed-node-2(192.168.16.12)] => (item=testbed-node-2)\u001b[0m\n\nTASK [ceph-facts : Check for a ceph mon socket] ********************************\nTuesday 24 February 2026  00:47:29 +0000 (0:00:02.645)       0:00:14.913 ****** \n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-0) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-1) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-facts : Check if the ceph mon socket is in-use] *********************\nTuesday 24 February 2026  00:47:30 +0000 (0:00:00.722)       0:00:15.635 ****** \n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-0', 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-1', 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-2', 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-facts : Set_fact running_mon - non_container] ***********************\nTuesday 24 February 2026  00:47:31 +0000 (0:00:01.088)       0:00:16.725 ****** \n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': {'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-0', 'ansible_loop_var': 'item'}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': {'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-1', 'ansible_loop_var': 'item'}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': {'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-2', 'ansible_loop_var': 'item'}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-facts : Set_fact running_mon - container] ***************************\nTuesday 24 February 2026  00:47:31 +0000 (0:00:00.464)       0:00:17.190 ****** \n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'stdout': '', 'stderr': '', 'rc': 0, 'cmd': ['docker', 'ps', '-q', '--filter', 'name=ceph-mon-testbed-node-0'], 'start': '2026-02-24 00:47:27.526222', 'end': '2026-02-24 00:47:27.625319', 'delta': '0:00:00.099097', 'msg': '', 'invocation': {'module_args': {'_raw_params': 'docker ps -q --filter name=ceph-mon-testbed-node-0', '_uses_shell': False, 'expand_argument_vars': True, 'stdin_add_newline': True, 'strip_empty_ends': True, 'argv': None, 'chdir': None, 'executable': None, 'creates': None, 'removes': None, 'stdin': None}}, 'stdout_lines': [], 'stderr_lines': [], 'failed': False, 'failed_when_result': False, 'item': 'testbed-node-0', 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'stdout': '', 'stderr': '', 'rc': 0, 'cmd': ['docker', 'ps', '-q', '--filter', 'name=ceph-mon-testbed-node-1'], 'start': '2026-02-24 00:47:28.794784', 'end': '2026-02-24 00:47:28.901216', 'delta': '0:00:00.106432', 'msg': '', 'invocation': {'module_args': {'_raw_params': 'docker ps -q --filter name=ceph-mon-testbed-node-1', '_uses_shell': False, 'expand_argument_vars': True, 'stdin_add_newline': True, 'strip_empty_ends': True, 'argv': None, 'chdir': None, 'executable': None, 'creates': None, 'removes': None, 'stdin': None}}, 'stdout_lines': [], 'stderr_lines': [], 'failed': False, 'failed_when_result': False, 'item': 'testbed-node-1', 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'stdout': '', 'stderr': '', 'rc': 0, 'cmd': ['docker', 'ps', '-q', '--filter', 'name=ceph-mon-testbed-node-2'], 'start': '2026-02-24 00:47:29.340894', 'end': '2026-02-24 00:47:29.439510', 'delta': '0:00:00.098616', 'msg': '', 'invocation': {'module_args': {'_raw_params': 'docker ps -q --filter name=ceph-mon-testbed-node-2', '_uses_shell': False, 'expand_argument_vars': True, 'stdin_add_newline': True, 'strip_empty_ends': True, 'argv': None, 'chdir': None, 'executable': None, 'creates': None, 'removes': None, 'stdin': None}}, 'stdout_lines': [], 'stderr_lines': [], 'failed': False, 'failed_when_result': False, 'item': 'testbed-node-2', 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-facts : Set_fact _container_exec_cmd] *******************************\nTuesday 24 February 2026  00:47:32 +0000 (0:00:00.210)       0:00:17.400 ****** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-facts : Get current fsid if cluster is already running] *************\nTuesday 24 February 2026  00:47:33 +0000 (0:00:01.197)       0:00:18.597 ****** \n\u001b[0;32mok: [testbed-node-3 -> testbed-node-0(192.168.16.10)]\u001b[0m\n\nTASK [ceph-facts : Set_fact current_fsid rc 1] *********************************\nTuesday 24 February 2026  00:47:33 +0000 (0:00:00.659)       0:00:19.257 ****** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-facts : Get current fsid] *******************************************\nTuesday 24 February 2026  00:47:36 +0000 (0:00:02.275)       0:00:21.532 ****** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-facts : Set_fact fsid] **********************************************\nTuesday 24 February 2026  00:47:37 +0000 (0:00:01.272)       0:00:22.805 ****** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-facts : Set_fact fsid from current_fsid] ****************************\nTuesday 24 February 2026  00:47:38 +0000 (0:00:01.370)       0:00:24.176 ****** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-facts : Generate cluster fsid] **************************************\nTuesday 24 February 2026  00:47:39 +0000 (0:00:00.245)       0:00:24.421 ****** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-facts : Set_fact fsid] **********************************************\nTuesday 24 February 2026  00:47:39 +0000 (0:00:00.657)       0:00:25.078 ****** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-facts : Resolve device link(s)] *************************************\nTuesday 24 February 2026  00:47:40 +0000 (0:00:01.121)       0:00:26.200 ****** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-facts : Set_fact build devices from resolved symlinks] **************\nTuesday 24 February 2026  00:47:42 +0000 (0:00:01.206)       0:00:27.407 ****** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-facts : Resolve dedicated_device link(s)] ***************************\nTuesday 24 February 2026  00:47:42 +0000 (0:00:00.746)       0:00:28.153 ****** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-facts : Set_fact build dedicated_devices from resolved symlinks] ****\nTuesday 24 February 2026  00:47:43 +0000 (0:00:01.100)       0:00:29.254 ****** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-facts : Resolve bluestore_wal_device link(s)] ***********************\nTuesday 24 February 2026  00:47:44 +0000 (0:00:00.832)       0:00:30.086 ****** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-facts : Set_fact build bluestore_wal_devices from resolved symlinks] ***\nTuesday 24 February 2026  00:47:45 +0000 (0:00:00.770)       0:00:30.857 ****** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-facts : Collect existed devices] ************************************\nTuesday 24 February 2026  00:47:46 +0000 (0:00:00.615)       0:00:31.472 ****** \n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'dm-0', 'value': {'holders': [], 'host': '', 'links': {'ids': ['dm-name-ceph--9b92f24c--dac1--5bff--a390--34d9ff9e0f4f-osd--block--9b92f24c--dac1--5bff--a390--34d9ff9e0f4f', 'dm-uuid-LVM-h1BZIKSwq0ZvfBWtXhB9IpbOvYH7db3oH2r78k3xPmfToT2WBghjfHVanEend8Zd'], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': '', 'sectors': 41934848, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'dm-1', 'value': {'holders': [], 'host': '', 'links': {'ids': ['dm-name-ceph--7f6a8fe5--6adc--52b4--b256--edcdbfd9bf4e-osd--block--7f6a8fe5--6adc--52b4--b256--edcdbfd9bf4e', 'dm-uuid-LVM-xRh1zTNtAhB8OweVwVFtzaksaCJoA7fP27eNW4gBNJ6VNIfew4kSrXXLjM4ybDHa'], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': '', 'sectors': 41934848, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'loop0', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'loop1', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'loop2', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'loop3', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'loop4', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'loop5', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'dm-0', 'value': {'holders': [], 'host': '', 'links': {'ids': ['dm-name-ceph--294dd48f--9d6d--5079--8214--59e60af9e526-osd--block--294dd48f--9d6d--5079--8214--59e60af9e526', 'dm-uuid-LVM-wjENyDIL1ovSEUDphRiOfALWwweDrSZWE8niuN9gyF3lbbybFqRYsEaNkOnZAeUN'], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': '', 'sectors': 41934848, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'dm-1', 'value': {'holders': [], 'host': '', 'links': {'ids': ['dm-name-ceph--7fbf4d63--7c6f--5c70--887c--2b73c978775b-osd--block--7fbf4d63--7c6f--5c70--887c--2b73c978775b', 'dm-uuid-LVM-aRcKlwv7ZYsrJYaz5OO903mOAEBnwLXI61utfRQTUZQKHwm4ceJ3NApz8UrwnN8y'], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': '', 'sectors': 41934848, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'loop6', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'loop0', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'loop7', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'loop1', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'loop2', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'sda', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_f90a4e80-1f36-4d14-b7da-de7b20e9fbca', 'scsi-SQEMU_QEMU_HARDDISK_f90a4e80-1f36-4d14-b7da-de7b20e9fbca'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {'sda1': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_f90a4e80-1f36-4d14-b7da-de7b20e9fbca-part1', 'scsi-SQEMU_QEMU_HARDDISK_f90a4e80-1f36-4d14-b7da-de7b20e9fbca-part1'], 'labels': ['cloudimg-rootfs'], 'masters': [], 'uuids': ['b852d8d2-8460-44aa-8998-23e4f04d73cf']}, 'sectors': 165672927, 'sectorsize': 512, 'size': '79.00 GB', 'start': '2099200', 'uuid': 'b852d8d2-8460-44aa-8998-23e4f04d73cf'}, 'sda14': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_f90a4e80-1f36-4d14-b7da-de7b20e9fbca-part14', 'scsi-SQEMU_QEMU_HARDDISK_f90a4e80-1f36-4d14-b7da-de7b20e9fbca-part14'], 'labels': [], 'masters': [], 'uuids': []}, 'sectors': 8192, 'sectorsize': 512, 'size': '4.00 MB', 'start': '2048', 'uuid': None}, 'sda15': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_f90a4e80-1f36-4d14-b7da-de7b20e9fbca-part15', 'scsi-SQEMU_QEMU_HARDDISK_f90a4e80-1f36-4d14-b7da-de7b20e9fbca-part15'], 'labels': ['UEFI'], 'masters': [], 'uuids': ['5C78-612A']}, 'sectors': 217088, 'sectorsize': 512, 'size': '106.00 MB', 'start': '10240', 'uuid': '5C78-612A'}, 'sda16': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_f90a4e80-1f36-4d14-b7da-de7b20e9fbca-part16', 'scsi-SQEMU_QEMU_HARDDISK_f90a4e80-1f36-4d14-b7da-de7b20e9fbca-part16'], 'labels': ['BOOT'], 'masters': [], 'uuids': ['09d53dc1-1e03-4286-bbb8-2b1796cf92ec']}, 'sectors': 1869825, 'sectorsize': 512, 'size': '913.00 MB', 'start': '227328', 'uuid': '09d53dc1-1e03-4286-bbb8-2b1796cf92ec'}}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 167772160, 'sectorsize': '512', 'size': '80.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'loop3', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'loop4', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'sdb', 'value': {'holders': ['ceph--9b92f24c--dac1--5bff--a390--34d9ff9e0f4f-osd--block--9b92f24c--dac1--5bff--a390--34d9ff9e0f4f'], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['lvm-pv-uuid-mec8nJ-xUip-qiWP-sj8w-uf5B-1Lt0-v8yfvH', 'scsi-0QEMU_QEMU_HARDDISK_9c6b6763-9704-41ec-9b6d-97eb80a5145c', 'scsi-SQEMU_QEMU_HARDDISK_9c6b6763-9704-41ec-9b6d-97eb80a5145c'], 'labels': [], 'masters': ['dm-0'], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'loop5', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'dm-0', 'value': {'holders': [], 'host': '', 'links': {'ids': ['dm-name-ceph--752a9ec4--4bfc--5551--82cd--92b802bdad97-osd--block--752a9ec4--4bfc--5551--82cd--92b802bdad97', 'dm-uuid-LVM-jTt9fg6XlhiXbWhW0I9T02tzYsIeoKIoaf5BxtnvIAUk6EWNKNJMqSjVsZqTXHAa'], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': '', 'sectors': 41934848, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'dm-1', 'value': {'holders': [], 'host': '', 'links': {'ids': ['dm-name-ceph--ba85a0e3--2f57--51d5--a9f1--072f0e9a8deb-osd--block--ba85a0e3--2f57--51d5--a9f1--072f0e9a8deb', 'dm-uuid-LVM-YjmBzAWpiQhdCbdayUM0w3sqNqanvbLDieybFDy5wWcdDcB5LPRFHjaRE0WCOjBW'], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': '', 'sectors': 41934848, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'loop6', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'sdc', 'value': {'holders': ['ceph--7f6a8fe5--6adc--52b4--b256--edcdbfd9bf4e-osd--block--7f6a8fe5--6adc--52b4--b256--edcdbfd9bf4e'], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['lvm-pv-uuid-HdfFdZ-81SJ-0Umt-qAzb-Dao2-7tUe-tHuhnG', 'scsi-0QEMU_QEMU_HARDDISK_6644f191-69e9-4f3a-baff-cffe4152595b', 'scsi-SQEMU_QEMU_HARDDISK_6644f191-69e9-4f3a-baff-cffe4152595b'], 'labels': [], 'masters': ['dm-1'], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'loop0', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'loop1', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'loop2', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'loop3', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'loop4', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'loop5', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'sdd', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_4534650b-23a0-46c8-8bf1-44520caee637', 'scsi-SQEMU_QEMU_HARDDISK_4534650b-23a0-46c8-8bf1-44520caee637'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'loop6', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'loop7', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'loop7', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partit\u001b[32m2026-02-24 00:58:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state SUCCESS\u001b[0m\nions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'sda', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_7a9487cb-ec9e-4508-9c5a-f692700fc5a4', 'scsi-SQEMU_QEMU_HARDDISK_7a9487cb-ec9e-4508-9c5a-f692700fc5a4'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {'sda1': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_7a9487cb-ec9e-4508-9c5a-f692700fc5a4-part1', 'scsi-SQEMU_QEMU_HARDDISK_7a9487cb-ec9e-4508-9c5a-f692700fc5a4-part1'], 'labels': ['cloudimg-rootfs'], 'masters': [], 'uuids': ['b852d8d2-8460-44aa-8998-23e4f04d73cf']}, 'sectors': 165672927, 'sectorsize': 512, 'size': '79.00 GB', 'start': '2099200', 'uuid': 'b852d8d2-8460-44aa-8998-23e4f04d73cf'}, 'sda14': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_7a9487cb-ec9e-4508-9c5a-f692700fc5a4-part14', 'scsi-SQEMU_QEMU_HARDDISK_7a9487cb-ec9e-4508-9c5a-f692700fc5a4-part14'], 'labels': [], 'masters': [], 'uuids': []}, 'sectors': 8192, 'sectorsize': 512, 'size': '4.00 MB', 'start': '2048', 'uuid': None}, 'sda15': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_7a9487cb-ec9e-4508-9c5a-f692700fc5a4-part15', 'scsi-SQEMU_QEMU_HARDDISK_7a9487cb-ec9e-4508-9c5a-f692700fc5a4-part15'], 'labels': ['UEFI'], 'masters': [], 'uuids': ['5C78-612A']}, 'sectors': 217088, 'sectorsize': 512, 'size': '106.00 MB', 'start': '10240', 'uuid': '5C78-612A'}, 'sda16': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_7a9487cb-ec9e-4508-9c5a-f692700fc5a4-part16', 'scsi-SQEMU_QEMU_HARDDISK_7a9487cb-ec9e-4508-9c5a-f692700fc5a4-part16'], 'labels': ['BOOT'], 'masters': [], 'uuids': ['09d53dc1-1e03-4286-bbb8-2b1796cf92ec']}, 'sectors': 1869825, 'sectorsize': 512, 'size': '913.00 MB', 'start': '227328', 'uuid': '09d53dc1-1e03-4286-bbb8-2b1796cf92ec'}}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 167772160, 'sectorsize': '512', 'size': '80.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'sr0', 'value': {'holders': [], 'host': 'IDE interface: Intel Corporation 82371SB PIIX3 IDE [Natoma/Triton II]', 'links': {'ids': ['ata-QEMU_DVD-ROM_QM00001'], 'labels': ['config-2'], 'masters': [], 'uuids': ['2026-02-24-00-01-58-00']}, 'model': 'QEMU DVD-ROM', 'partitions': {}, 'removable': '1', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'mq-deadline', 'sectors': 253, 'sectorsize': '2048', 'size': '506.00 KB', 'support_discard': '0', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'sda', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_47e600ec-05ca-44a6-93d0-5ca3104e3639', 'scsi-SQEMU_QEMU_HARDDISK_47e600ec-05ca-44a6-93d0-5ca3104e3639'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {'sda1': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_47e600ec-05ca-44a6-93d0-5ca3104e3639-part1', 'scsi-SQEMU_QEMU_HARDDISK_47e600ec-05ca-44a6-93d0-5ca3104e3639-part1'], 'labels': ['cloudimg-rootfs'], 'masters': [], 'uuids': ['b852d8d2-8460-44aa-8998-23e4f04d73cf']}, 'sectors': 165672927, 'sectorsize': 512, 'size': '79.00 GB', 'start': '2099200', 'uuid': 'b852d8d2-8460-44aa-8998-23e4f04d73cf'}, 'sda14': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_47e600ec-05ca-44a6-93d0-5ca3104e3639-part14', 'scsi-SQEMU_QEMU_HARDDISK_47e600ec-05ca-44a6-93d0-5ca3104e3639-part14'], 'labels': [], 'masters': [], 'uuids': []}, 'sectors': 8192, 'sectorsize': 512, 'size': '4.00 MB', 'start': '2048', 'uuid': None}, 'sda15': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_47e600ec-05ca-44a6-93d0-5ca3104e3639-part15', 'scsi-SQEMU_QEMU_HARDDISK_47e600ec-05ca-44a6-93d0-5ca3104e3639-part15'], 'labels': ['UEFI'], 'masters': [], 'uuids': ['5C78-612A']}, 'sectors': 217088, 'sectorsize': 512, 'size': '106.00 MB', 'start': '10240', 'uuid': '5C78-612A'}, 'sda16': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_47e600ec-05ca-44a6-93d0-5ca3104e3639-part16', 'scsi-SQEMU_QEMU_HARDDISK_47e600ec-05ca-44a6-93d0-5ca3104e3639-part16'], 'labels': ['BOOT'], 'masters': [], 'uuids': ['09d53dc1-1e03-4286-bbb8-2b1796cf92ec']}, 'sectors': 1869825, 'sectorsize': 512, 'size': '913.00 MB', 'start': '227328', 'uuid': '09d53dc1-1e03-4286-bbb8-2b1796cf92ec'}}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 167772160, 'sectorsize': '512', 'size': '80.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'sdb', 'value': {'holders': ['ceph--752a9ec4--4bfc--5551--82cd--92b802bdad97-osd--block--752a9ec4--4bfc--5551--82cd--92b802bdad97'], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['lvm-pv-uuid-vfwhPS-pgoq-FNQw-W72Z-epjC-Ziiv-qWNph9', 'scsi-0QEMU_QEMU_HARDDISK_1558935b-4c66-4614-97e5-cda14abd2641', 'scsi-SQEMU_QEMU_HARDDISK_1558935b-4c66-4614-97e5-cda14abd2641'], 'labels': [], 'masters': ['dm-0'], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'sdb', 'value': {'holders': ['ceph--294dd48f--9d6d--5079--8214--59e60af9e526-osd--block--294dd48f--9d6d--5079--8214--59e60af9e526'], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['lvm-pv-uuid-1V6jIs-3p8q-l9Rl-OJ2n-8L4P-5xCz-K0YqYS', 'scsi-0QEMU_QEMU_HARDDISK_f659b61f-c39f-426f-a640-6a753951661d', 'scsi-SQEMU_QEMU_HARDDISK_f659b61f-c39f-426f-a640-6a753951661d'], 'labels': [], 'masters': ['dm-0'], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'sdc', 'value': {'holders': ['ceph--ba85a0e3--2f57--51d5--a9f1--072f0e9a8deb-osd--block--ba85a0e3--2f57--51d5--a9f1--072f0e9a8deb'], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['lvm-pv-uuid-XuBysF-MTWF-hP8t-3CGS-0PEc-QL0q-u9CYq2', 'scsi-0QEMU_QEMU_HARDDISK_6f4643f0-329b-4ae2-97c7-d348648fd970', 'scsi-SQEMU_QEMU_HARDDISK_6f4643f0-329b-4ae2-97c7-d348648fd970'], 'labels': [], 'masters': ['dm-1'], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'sdc', 'value': {'holders': ['ceph--7fbf4d63--7c6f--5c70--887c--2b73c978775b-osd--block--7fbf4d63--7c6f--5c70--887c--2b73c978775b'], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['lvm-pv-uuid-aHCB2h-b0pT-WCWd-loOK-A09H-yFYp-FH2wIv', 'scsi-0QEMU_QEMU_HARDDISK_fb58bf16-a1c6-4abb-925b-29c6a6a6f100', 'scsi-SQEMU_QEMU_HARDDISK_fb58bf16-a1c6-4abb-925b-29c6a6a6f100'], 'labels': [], 'masters': ['dm-1'], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'sdd', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_5f47789f-2a1a-442a-8a3c-644a4f92e5a8', 'scsi-SQEMU_QEMU_HARDDISK_5f47789f-2a1a-442a-8a3c-644a4f92e5a8'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'sr0', 'value': {'holders': [], 'host': 'IDE interface: Intel Corporation 82371SB PIIX3 IDE [Natoma/Triton II]', 'links': {'ids': ['ata-QEMU_DVD-ROM_QM00001'], 'labels': ['config-2'], 'masters': [], 'uuids': ['2026-02-24-00-02-03-00']}, 'model': 'QEMU DVD-ROM', 'partitions': {}, 'removable': '1', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'mq-deadline', 'sectors': 253, 'sectorsize': '2048', 'size': '506.00 KB', 'support_discard': '0', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'sdd', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_5660bb62-e2c9-45b2-ae11-8ab0f829b194', 'scsi-SQEMU_QEMU_HARDDISK_5660bb62-e2c9-45b2-ae11-8ab0f829b194'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'sr0', 'value': {'holders': [], 'host': 'IDE interface: Intel Corporation 82371SB PIIX3 IDE [Natoma/Triton II]', 'links': {'ids': ['ata-QEMU_DVD-ROM_QM00001'], 'labels': ['config-2'], 'masters': [], 'uuids': ['2026-02-24-00-01-59-00']}, 'model': 'QEMU DVD-ROM', 'partitions': {}, 'removable': '1', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'mq-deadline', 'sectors': 253, 'sectorsize': '2048', 'size': '506.00 KB', 'support_discard': '0', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'loop0', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'loop1', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'loop2', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'loop3', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'loop4', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'loop5', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'loop6', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'loop7', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'sda', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_2d259cae-7b79-479e-b978-3fffee75aa51', 'scsi-SQEMU_QEMU_HARDDISK_2d259cae-7b79-479e-b978-3fffee75aa51'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {'sda1': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_2d259cae-7b79-479e-b978-3fffee75aa51-part1', 'scsi-SQEMU_QEMU_HARDDISK_2d259cae-7b79-479e-b978-3fffee75aa51-part1'], 'labels': ['cloudimg-rootfs'], 'masters': [], 'uuids': ['b852d8d2-8460-44aa-8998-23e4f04d73cf']}, 'sectors': 165672927, 'sectorsize': 512, 'size': '79.00 GB', 'start': '2099200', 'uuid': 'b852d8d2-8460-44aa-8998-23e4f04d73cf'}, 'sda14': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_2d259cae-7b79-479e-b978-3fffee75aa51-part14', 'scsi-SQEMU_QEMU_HARDDISK_2d259cae-7b79-479e-b978-3fffee75aa51-part14'], 'labels': [], 'masters': [], 'uuids': []}, 'sectors': 8192, 'sectorsize': 512, 'size': '4.00 MB', 'start': '2048', 'uuid': None}, 'sda15': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_2d259cae-7b79-479e-b978-3fffee75aa51-part15', 'scsi-SQEMU_QEMU_HARDDISK_2d259cae-7b79-479e-b978-3fffee75aa51-part15'], 'labels': ['UEFI'], 'masters': [], 'uuids': ['5C78-612A']}, 'sectors': 217088, 'sectorsize': 512, 'size': '106.00 MB', 'start': '10240', 'uuid': '5C78-612A'}, 'sda16': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_2d259cae-7b79-479e-b978-3fffee75aa51-part16', 'scsi-SQEMU_QEMU_HARDDISK_2d259cae-7b79-479e-b978-3fffee75aa51-part16'], 'labels': ['BOOT'], 'masters': [], 'uuids': ['09d53dc1-1e03-4286-bbb8-2b1796cf92ec']}, 'sectors': 1869825, 'sectorsize': 512, 'size': '913.00 MB', 'start': '227328', 'uuid': '09d53dc1-1e03-4286-bbb8-2b1796cf92ec'}}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 167772160, 'sectorsize': '512', 'size': '80.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'sr0', 'value': {'holders': [], 'host': 'IDE interface: Intel Corporation 82371SB PIIX3 IDE [Natoma/Triton II]', 'links': {'ids': ['ata-QEMU_DVD-ROM_QM00001'], 'labels': ['config-2'], 'masters': [], 'uuids': ['2026-02-24-00-01-56-00']}, 'model': 'QEMU DVD-ROM', 'partitions': {}, 'removable': '1', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'mq-deadline', 'sectors': 253, 'sectorsize': '2048', 'size': '506.00 KB', 'support_discard': '0', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'loop0', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'loop1', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'loop2', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'loop3', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'loop4', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'loop5', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'loop6', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'loop7', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'sda', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_6fef90f9-4ce5-456f-a170-152290f19313', 'scsi-SQEMU_QEMU_HARDDISK_6fef90f9-4ce5-456f-a170-152290f19313'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {'sda1': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_6fef90f9-4ce5-456f-a170-152290f19313-part1', 'scsi-SQEMU_QEMU_HARDDISK_6fef90f9-4ce5-456f-a170-152290f19313-part1'], 'labels': ['cloudimg-rootfs'], 'masters': [], 'uuids': ['b852d8d2-8460-44aa-8998-23e4f04d73cf']}, 'sectors': 165672927, 'sectorsize': 512, 'size': '79.00 GB', 'start': '2099200', 'uuid': 'b852d8d2-8460-44aa-8998-23e4f04d73cf'}, 'sda14': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_6fef90f9-4ce5-456f-a170-152290f19313-part14', 'scsi-SQEMU_QEMU_HARDDISK_6fef90f9-4ce5-456f-a170-152290f19313-part14'], 'labels': [], 'masters': [], 'uuids': []}, 'sectors': 8192, 'sectorsize': 512, 'size': '4.00 MB', 'start': '2048', 'uuid': None}, 'sda15': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_6fef90f9-4ce5-456f-a170-152290f19313-part15', 'scsi-SQEMU_QEMU_HARDDISK_6fef90f9-4ce5-456f-a170-152290f19313-part15'], 'labels': ['UEFI'], 'masters': [], 'uuids': ['5C78-612A']}, 'sectors': 217088, 'sectorsize': 512, 'size': '106.00 MB', 'start': '10240', 'uuid': '5C78-612A'}, 'sda16': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_6fef90f9-4ce5-456f-a170-152290f19313-part16', 'scsi-SQEMU_QEMU_HARDDISK_6fef90f9-4ce5-456f-a170-152290f19313-part16'], 'labels': ['BOOT'], 'masters': [], 'uuids': ['09d53dc1-1e03-4286-bbb8-2b1796cf92ec']}, 'sectors': 1869825, 'sectorsize': 512, 'size': '913.00 MB', 'start': '227328', 'uuid': '09d53dc1-1e03-4286-bbb8-2b1796cf92ec'}}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 167772160, 'sectorsize': '512', 'size': '80.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'sr0', 'value': {'holders': [], 'host': 'IDE interface: Intel Corporation 82371SB PIIX3 IDE [Natoma/Triton II]', 'links': {'ids': ['ata-QEMU_DVD-ROM_QM00001'], 'labels': ['config-2'], 'masters': [], 'uuids': ['2026-02-24-00-02-01-00']}, 'model': 'QEMU DVD-ROM', 'partitions': {}, 'removable': '1', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'mq-deadline', 'sectors': 253, 'sectorsize': '2048', 'size': '506.00 KB', 'support_discard': '0', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'loop0', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'loop1', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'loop2', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'loop3', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'loop4', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'loop5', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'loop6', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'loop7', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'sda', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_35b8d51e-716e-455d-9a95-995756ae5900', 'scsi-SQEMU_QEMU_HARDDISK_35b8d51e-716e-455d-9a95-995756ae5900'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {'sda1': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_35b8d51e-716e-455d-9a95-995756ae5900-part1', 'scsi-SQEMU_QEMU_HARDDISK_35b8d51e-716e-455d-9a95-995756ae5900-part1'], 'labels': ['cloudimg-rootfs'], 'masters': [], 'uuids': ['b852d8d2-8460-44aa-8998-23e4f04d73cf']}, 'sectors': 165672927, 'sectorsize': 512, 'size': '79.00 GB', 'start': '2099200', 'uuid': 'b852d8d2-8460-44aa-8998-23e4f04d73cf'}, 'sda14': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_35b8d51e-716e-455d-9a95-995756ae5900-part14', 'scsi-SQEMU_QEMU_HARDDISK_35b8d51e-716e-455d-9a95-995756ae5900-part14'], 'labels': [], 'masters': [], 'uuids': []}, 'sectors': 8192, 'sectorsize': 512, 'size': '4.00 MB', 'start': '2048', 'uuid': None}, 'sda15': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_35b8d51e-716e-455d-9a95-995756ae5900-part15', 'scsi-SQEMU_QEMU_HARDDISK_35b8d51e-716e-455d-9a95-995756ae5900-part15'], 'labels': ['UEFI'], 'masters': [], 'uuids': ['5C78-612A']}, 'sectors': 217088, 'sectorsize': 512, 'size': '106.00 MB', 'start': '10240', 'uuid': '5C78-612A'}, 'sda16': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_35b8d51e-716e-455d-9a95-995756ae5900-part16', 'scsi-SQEMU_QEMU_HARDDISK_35b8d51e-716e-455d-9a95-995756ae5900-part16'], 'labels': ['BOOT'], 'masters': [], 'uuids': ['09d53dc1-1e03-4286-bbb8-2b1796cf92ec']}, 'sectors': 1869825, 'sectorsize': 512, 'size': '913.00 MB', 'start': '227328', 'uuid': '09d53dc1-1e03-4286-bbb8-2b1796cf92ec'}}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 167772160, 'sectorsize': '512', 'size': '80.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'sr0', 'value': {'holders': [], 'host': 'IDE interface: Intel Corporation 82371SB PIIX3 IDE [Natoma/Triton II]', 'links': {'ids': ['ata-QEMU_DVD-ROM_QM00001'], 'labels': ['config-2'], 'masters': [], 'uuids': ['2026-02-24-00-02-05-00']}, 'model': 'QEMU DVD-ROM', 'partitions': {}, 'removable': '1', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'mq-deadline', 'sectors': 253, 'sectorsize': '2048', 'size': '506.00 KB', 'support_discard': '0', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-facts : Set_fact devices generate device list when osd_auto_discovery] ***\nTuesday 24 February 2026  00:47:48 +0000 (0:00:01.901)       0:00:33.374 ****** \n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'dm-0', 'value': {'holders': [], 'host': '', 'links': {'ids': ['dm-name-ceph--9b92f24c--dac1--5bff--a390--34d9ff9e0f4f-osd--block--9b92f24c--dac1--5bff--a390--34d9ff9e0f4f', 'dm-uuid-LVM-h1BZIKSwq0ZvfBWtXhB9IpbOvYH7db3oH2r78k3xPmfToT2WBghjfHVanEend8Zd'], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': '', 'sectors': 41934848, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'dm-1', 'value': {'holders': [], 'host': '', 'links': {'ids': ['dm-name-ceph--7f6a8fe5--6adc--52b4--b256--edcdbfd9bf4e-osd--block--7f6a8fe5--6adc--52b4--b256--edcdbfd9bf4e', 'dm-uuid-LVM-xRh1zTNtAhB8OweVwVFtzaksaCJoA7fP27eNW4gBNJ6VNIfew4kSrXXLjM4ybDHa'], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': '', 'sectors': 41934848, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop0', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop1', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop2', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'dm-0', 'value': {'holders': [], 'host': '', 'links': {'ids': ['dm-name-ceph--294dd48f--9d6d--5079--8214--59e60af9e526-osd--block--294dd48f--9d6d--5079--8214--59e60af9e526', 'dm-uuid-LVM-wjENyDIL1ovSEUDphRiOfALWwweDrSZWE8niuN9gyF3lbbybFqRYsEaNkOnZAeUN'], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': '', 'sectors': 41934848, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop3', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop4', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'dm-1', 'value': {'holders': [], 'host': '', 'links': {'ids': ['dm-name-ceph--7fbf4d63--7c6f--5c70--887c--2b73c978775b-osd--block--7fbf4d63--7c6f--5c70--887c--2b73c978775b', 'dm-uuid-LVM-aRcKlwv7ZYsrJYaz5OO903mOAEBnwLXI61utfRQTUZQKHwm4ceJ3NApz8UrwnN8y'], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': '', 'sectors': 41934848, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop5', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop6', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop0', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop0', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop1', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'dm-0', 'value': {'holders': [], 'host': '', 'links': {'ids': ['dm-name-ceph--752a9ec4--4bfc--5551--82cd--92b802bdad97-osd--block--752a9ec4--4bfc--5551--82cd--92b802bdad97', 'dm-uuid-LVM-jTt9fg6XlhiXbWhW0I9T02tzYsIeoKIoaf5BxtnvIAUk6EWNKNJMqSjVsZqTXHAa'], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': '', 'sectors': 41934848, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop7', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop2', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop1', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop3', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop4', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop5', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'dm-1', 'value': {'holders': [], 'host': '', 'links': {'ids': ['dm-name-ceph--ba85a0e3--2f57--51d5--a9f1--072f0e9a8deb-osd--block--ba85a0e3--2f57--51d5--a9f1--072f0e9a8deb', 'dm-uuid-LVM-YjmBzAWpiQhdCbdayUM0w3sqNqanvbLDieybFDy5wWcdDcB5LPRFHjaRE0WCOjBW'], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': '', 'sectors': 41934848, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop6', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop2', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sda', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_f90a4e80-1f36-4d14-b7da-de7b20e9fbca', 'scsi-SQEMU_QEMU_HARDDISK_f90a4e80-1f36-4d14-b7da-de7b20e9fbca'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {'sda1': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_f90a4e80-1f36-4d14-b7da-de7b20e9fbca-part1', 'scsi-SQEMU_QEMU_HARDDISK_f90a4e80-1f36-4d14-b7da-de7b20e9fbca-part1'], 'labels': ['cloudimg-rootfs'], 'masters': [], 'uuids': ['b852d8d2-8460-44aa-8998-23e4f04d73cf']}, 'sectors': 165672927, 'sectorsize': 512, 'size': '79.00 GB', 'start': '2099200', 'uuid': 'b852d8d2-8460-44aa-8998-23e4f04d73cf'}, 'sda14': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_f90a4e80-1f36-4d14-b7da-de7b20e9fbca-part14', 'scsi-SQEMU_QEMU_HARDDISK_f90a4e80-1f36-4d14-b7da-de7b20e9fbca-part14'], 'labels': [], 'masters': [], 'uuids': []}, 'sectors': 8192, 'sectorsize': 512, 'size': '4.00 MB', 'start': '2048', 'uuid': None}, 'sda15': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_f90a4e80-1f36-4d14-b7da-de7b20e9fbca-part15', 'scsi-SQEMU_QEMU_HARDDISK_f90a4e80-1f36-4d14-b7da-de7b20e9fbca-part15'], 'labels': ['UEFI'], 'masters': [], 'uuids': ['5C78-612A']}, 'sectors': 217088, 'sectorsize': 512, 'size': '106.00 MB', 'start': '10240', 'uuid': '5C78-612A'}, 'sda16': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_f90a4e80-1f36-4d14-b7da-de7b20e9fbca-part16', 'scsi-SQEMU_QEMU_HARDDISK_f90a4e80-1f36-4d14-b7da-de7b20e9fbca-part16'], 'labels': ['BOOT'], 'masters': [], 'uuids': ['09d53dc1-1e03-4286-bbb8-2b1796cf92ec']}, 'sectors': 1869825, 'sectorsize': 512, 'size': '913.00 MB', 'start': '227328', 'uuid': '09d53dc1-1e03-4286-bbb8-2b1796cf92ec'}}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 167772160, 'sectorsize': '512', 'size': '80.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop7', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'sda', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_2d259cae-7b79-479e-b978-3fffee75aa51', 'scsi-SQEMU_QEMU_HARDDISK_2d259cae-7b79-479e-b978-3fffee75aa51'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {'sda1': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_2d259cae-7b79-479e-b978-3fffee75aa51-part1', 'scsi-SQEMU_QEMU_HARDDISK_2d259cae-7b79-479e-b978-3fffee75aa51-part1'], 'labels': ['cloudimg-rootfs'], 'masters': [], 'uuids': ['b852d8d2-8460-44aa-8998-23e4f04d73cf']}, 'sectors': 165672927, 'sectorsize': 512, 'size': '79.00 GB', 'start': '2099200', 'uuid': 'b852d8d2-8460-44aa-8998-23e4f04d73cf'}, 'sda14': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_2d259cae-7b79-479e-b978-3fffee75aa51-part14', 'scsi-SQEMU_QEMU_HARDDISK_2d259cae-7b79-479e-b978-3fffee75aa51-part14'], 'labels': [], 'masters': [], 'uuids': []}, 'sectors': 8192, 'sectorsize': 512, 'size': '4.00 MB', 'start': '2048', 'uuid': None}, 'sda15': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_2d259cae-7b79-479e-b978-3fffee75aa51-part15', 'scsi-SQEMU_QEMU_HARDDISK_2d259cae-7b79-479e-b978-3fffee75aa51-part15'], 'labels': ['UEFI'], 'masters': [], 'uuids': ['5C78-612A']}, 'sectors': 217088, 'sectorsize': 512, 'size': '106.00 MB', 'start': '10240', 'uuid': '5C78-612A'}, 'sda16': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_2d259cae-7b79-479e-b978-3fffee75aa51-part16', 'scsi-SQEMU_QEMU_HARDDISK_2d259cae-7b79-479e-b978-3fffee75aa51-part16'], 'labels': ['BOOT'], 'masters': [], 'uuids': ['09d53dc1-1e03-4286-bbb8-2b1796cf92ec']}, 'sectors': 1869825, 'sectorsize': 512, 'size': '913.00 MB', 'start': '227328', 'uuid': '09d53dc1-1e03-4286-bbb8-2b1796cf92ec'}}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 167772160, 'sectorsize': '512', 'size': '80.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop0', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'sr0', 'value': {'holders': [], 'host': 'IDE interface: Intel Corporation 82371SB PIIX3 IDE [Natoma/Triton II]', 'links': {'ids': ['ata-QEMU_DVD-ROM_QM00001'], 'labels': ['config-2'], 'masters': [], 'uuids': ['2026-02-24-00-01-56-00']}, 'model': 'QEMU DVD-ROM', 'partitions': {}, 'removable': '1', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'mq-deadline', 'sectors': 253, 'sectorsize': '2048', 'size': '506.00 KB', 'support_discard': '0', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop3', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sdb', 'value': {'holders': ['ceph--9b92f24c--dac1--5bff--a390--34d9ff9e0f4f-osd--block--9b92f24c--dac1--5bff--a390--34d9ff9e0f4f'], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['lvm-pv-uuid-mec8nJ-xUip-qiWP-sj8w-uf5B-1Lt0-v8yfvH', 'scsi-0QEMU_QEMU_HARDDISK_9c6b6763-9704-41ec-9b6d-97eb80a5145c', 'scsi-SQEMU_QEMU_HARDDISK_9c6b6763-9704-41ec-9b6d-97eb80a5145c'], 'labels': [], 'masters': ['dm-0'], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop1', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop4', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop2', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop5', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sdc', 'value': {'holders': ['ceph--7f6a8fe5--6adc--52b4--b256--edcdbfd9bf4e-osd--block--7f6a8fe5--6adc--52b4--b256--edcdbfd9bf4e'], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['lvm-pv-uuid-HdfFdZ-81SJ-0Umt-qAzb-Dao2-7tUe-tHuhnG', 'scsi-0QEMU_QEMU_HARDDISK_6644f191-69e9-4f3a-baff-cffe4152595b', 'scsi-SQEMU_QEMU_HARDDISK_6644f191-69e9-4f3a-baff-cffe4152595b'], 'labels': [], 'masters': ['dm-1'], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop0', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop6', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop3', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop7', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop1', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sda', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_47e600ec-05ca-44a6-93d0-5ca3104e3639', 'scsi-SQEMU_QEMU_HARDDISK_47e600ec-05ca-44a6-93d0-5ca3104e3639'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {'sda1': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_47e600ec-05ca-44a6-93d0-5ca3104e3639-part1', 'scsi-SQEMU_QEMU_HARDDISK_47e600ec-05ca-44a6-93d0-5ca3104e3639-part1'], 'labels': ['cloudimg-rootfs'], 'masters': [], 'uuids': ['b852d8d2-8460-44aa-8998-23e4f04d73cf']}, 'sectors': 165672927, 'sectorsize': 512, 'size': '79.00 GB', 'start': '2099200', 'uuid': 'b852d8d2-8460-44aa-8998-23e4f04d73cf'}, 'sda14': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_47e600ec-05ca-44a6-93d0-5ca3104e3639-part14', 'scsi-SQEMU_QEMU_HARDDISK_47e600ec-05ca-44a6-93d0-5ca3104e3639-part14'], 'labels': [], 'masters': [], 'uuids': []}, 'sectors': 8192, 'sectorsize': 512, 'size': '4.00 MB', 'start': '2048', 'uuid': None}, 'sda15': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_47e600ec-05ca-44a6-93d0-5ca3104e3639-part15', 'scsi-SQEMU_QEMU_HARDDISK_47e600ec-05ca-44a6-93d0-5ca3104e3639-part15'], 'labels': ['UEFI'], 'masters': [], 'uuids': ['5C78-612A']}, 'sectors': 217088, 'sectorsize': 512, 'size': '106.00 MB', 'start': '10240', 'uuid': '5C78-612A'}, 'sda16': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_47e600ec-05ca-44a6-93d0-5ca3104e3639-part16', 'scsi-SQEMU_QEMU_HARDDISK_47e600ec-05ca-44a6-93d0-5ca3104e3639-part16'], 'labels': ['BOOT'], 'masters': [], 'uuids': ['09d53dc1-1e03-4286-bbb8-2b1796cf92ec']}, 'sectors': 1869825, 'sectorsize': 512, 'size': '913.00 MB', 'start': '227328', 'uuid': '09d53dc1-1e03-4286-bbb8-2b1796cf92ec'}}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 167772160, 'sectorsize': '512', 'size': '80.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop2', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sdb', 'value': {'holders': ['ceph--294dd48f--9d6d--5079--8214--59e60af9e526-osd--block--294dd48f--9d6d--5079--8214--59e60af9e526'], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['lvm-pv-uuid-1V6jIs-3p8q-l9Rl-OJ2n-8L4P-5xCz-K0YqYS', 'scsi-0QEMU_QEMU_HARDDISK_f659b61f-c39f-426f-a640-6a753951661d', 'scsi-SQEMU_QEMU_HARDDISK_f659b61f-c39f-426f-a640-6a753951661d'], 'labels': [], 'masters': ['dm-0'], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop4', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sdc', 'value': {'holders': ['ceph--7fbf4d63--7c6f--5c70--887c--2b73c978775b-osd--block--7fbf4d63--7c6f--5c70--887c--2b73c978775b'], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['lvm-pv-uuid-aHCB2h-b0pT-WCWd-loOK-A09H-yFYp-FH2wIv', 'scsi-0QEMU_QEMU_HARDDISK_fb58bf16-a1c6-4abb-925b-29c6a6a6f100', 'scsi-SQEMU_QEMU_HARDDISK_fb58bf16-a1c6-4abb-925b-29c6a6a6f100'], 'labels': [], 'masters': ['dm-1'], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop3', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sdd', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_5660bb62-e2c9-45b2-ae11-8ab0f829b194', 'scsi-SQEMU_QEMU_HARDDISK_5660bb62-e2c9-45b2-ae11-8ab0f829b194'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop5', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sr0', 'value': {'holders': [], 'host': 'IDE interface: Intel Corporation 82371SB PIIX3 IDE [Natoma/Triton II]', 'links': {'ids': ['ata-QEMU_DVD-ROM_QM00001'], 'labels': ['config-2'], 'masters': [], 'uuids': ['2026-02-24-00-01-59-00']}, 'model': 'QEMU DVD-ROM', 'partitions': {}, 'removable': '1', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'mq-deadline', 'sectors': 253, 'sectorsize': '2048', 'size': '506.00 KB', 'support_discard': '0', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop6', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sdd', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_4534650b-23a0-46c8-8bf1-44520caee637', 'scsi-SQEMU_QEMU_HARDDISK_4534650b-23a0-46c8-8bf1-44520caee637'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop7', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sda', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_7a9487cb-ec9e-4508-9c5a-f692700fc5a4', 'scsi-SQEMU_QEMU_HARDDISK_7a9487cb-ec9e-4508-9c5a-f692700fc5a4'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {'sda1': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_7a9487cb-ec9e-4508-9c5a-f692700fc5a4-part1', 'scsi-SQEMU_QEMU_HARDDISK_7a9487cb-ec9e-4508-9c5a-f692700fc5a4-part1'], 'labels': ['cloudimg-rootfs'], 'masters': [], 'uuids': ['b852d8d2-8460-44aa-8998-23e4f04d73cf']}, 'sectors': 165672927, 'sectorsize': 512, 'size': '79.00 GB', 'start': '2099200', 'uuid': 'b852d8d2-8460-44aa-8998-23e4f04d73cf'}, 'sda14': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_7a9487cb-ec9e-4508-9c5a-f692700fc5a4-part14', 'scsi-SQEMU_QEMU_HARDDISK_7a9487cb-ec9e-4508-9c5a-f692700fc5a4-part14'], 'labels': [], 'masters': [], 'uuids': []}, 'sectors': 8192, 'sectorsize': 512, 'size': '4.00 MB', 'start': '2048', 'uuid': None}, 'sda15': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_7a9487cb-ec9e-4508-9c5a-f692700fc5a4-part15', 'scsi-SQEMU_QEMU_HARDDISK_7a9487cb-ec9e-4508-9c5a-f692700fc5a4-part15'], 'labels': ['UEFI'], 'masters': [], 'uuids': ['5C78-612A']}, 'sectors': 217088, 'sectorsize': 512, 'size': '106.00 MB', 'start': '10240', 'uuid': '5C78-612A'}, 'sda16': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_7a9487cb-ec9e-4508-9c5a-f692700fc5a4-part16', 'scsi-SQEMU_QEMU_HARDDISK_7a9487cb-ec9e-4508-9c5a-f692700fc5a4-part16'], 'labels': ['BOOT'], 'masters': [], 'uuids': ['09d53dc1-1e03-4286-bbb8-2b1796cf92ec']}, 'sectors': 1869825, 'sectorsize': 512, 'size': '913.00 MB', 'start': '227328', 'uuid': '09d53dc1-1e03-4286-bbb8-2b1796cf92ec'}}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 167772160, 'sectorsize': '512', 'size': '80.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop0', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sdb', 'value': {'holders': ['ceph--752a9ec4--4bfc--5551--82cd--92b802bdad97-osd--block--752a9ec4--4bfc--5551--82cd--92b802bdad97'], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['lvm-pv-uuid-vfwhPS-pgoq-FNQw-W72Z-epjC-Ziiv-qWNph9', 'scsi-0QEMU_QEMU_HARDDISK_1558935b-4c66-4614-97e5-cda14abd2641', 'scsi-SQEMU_QEMU_HARDDISK_1558935b-4c66-4614-97e5-cda14abd2641'], 'labels': [], 'masters': ['dm-0'], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop4', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop1', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sdc', 'value': {'holders': ['ceph--ba85a0e3--2f57--51d5--a9f1--072f0e9a8deb-osd--block--ba85a0e3--2f57--51d5--a9f1--072f0e9a8deb'], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['lvm-pv-uuid-XuBysF-MTWF-hP8t-3CGS-0PEc-QL0q-u9CYq2', 'scsi-0QEMU_QEMU_HARDDISK_6f4643f0-329b-4ae2-97c7-d348648fd970', 'scsi-SQEMU_QEMU_HARDDISK_6f4643f0-329b-4ae2-97c7-d348648fd970'], 'labels': [], 'masters': ['dm-1'], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop2', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sdd', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_5f47789f-2a1a-442a-8a3c-644a4f92e5a8', 'scsi-SQEMU_QEMU_HARDDISK_5f47789f-2a1a-442a-8a3c-644a4f92e5a8'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sr0', 'value': {'holders': [], 'host': 'IDE interface: Intel Corporation 82371SB PIIX3 IDE [Natoma/Triton II]', 'links': {'ids': ['ata-QEMU_DVD-ROM_QM00001'], 'labels': ['config-2'], 'masters': [], 'uuids': ['2026-02-24-00-02-03-00']}, 'model': 'QEMU DVD-ROM', 'partitions': {}, 'removable': '1', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'mq-deadline', 'sectors': 253, 'sectorsize': '2048', 'size': '506.00 KB', 'support_discard': '0', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop3', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop5', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop6', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sr0', 'value': {'holders': [], 'host': 'IDE interface: Intel Corporation 82371SB PIIX3 IDE [Natoma/Triton II]', 'links': {'ids': ['ata-QEMU_DVD-ROM_QM00001'], 'labels': ['config-2'], 'masters': [], 'uuids': ['2026-02-24-00-01-58-00']}, 'model': 'QEMU DVD-ROM', 'partitions': {}, 'removable': '1', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'mq-deadline', 'sectors': 253, 'sectorsize': '2048', 'size': '506.00 KB', 'support_discard': '0', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop7', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'sda', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_6fef90f9-4ce5-456f-a170-152290f19313', 'scsi-SQEMU_QEMU_HARDDISK_6fef90f9-4ce5-456f-a170-152290f19313'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {'sda1': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_6fef90f9-4ce5-456f-a170-152290f19313-part1', 'scsi-SQEMU_QEMU_HARDDISK_6fef90f9-4ce5-456f-a170-152290f19313-part1'], 'labels': ['cloudimg-rootfs'], 'masters': [], 'uuids': ['b852d8d2-8460-44aa-8998-23e4f04d73cf']}, 'sectors': 165672927, 'sectorsize': 512, 'size': '79.00 GB', 'start': '2099200', 'uuid': 'b852d8d2-8460-44aa-8998-23e4f04d73cf'}, 'sda14': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_6fef90f9-4ce5-456f-a170-152290f19313-part14', 'scsi-SQEMU_QEMU_HARDDISK_6fef90f9-4ce5-456f-a170-152290f19313-part14'], 'labels': [], 'masters': [], 'uuids': []}, 'sectors': 8192, 'sectorsize': 512, 'size': '4.00 MB', 'start': '2048', 'uuid': None}, 'sda15': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_6fef90f9-4ce5-456f-a170-152290f19313-part15', 'scsi-SQEMU_QEMU_HARDDISK_6fef90f9-4ce5-456f-a170-152290f19313-part15'], 'labels': ['UEFI'], 'masters': [], 'uuids': ['5C78-612A']}, 'sectors': 217088, 'sectorsize': 512, 'size': '106.00 MB', 'start': '10240', 'uuid': '5C78-612A'}, 'sda16': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_6fef90f9-4ce5-456f-a170-152290f19313-part16', 'scsi-SQEMU_QEMU_HARDDISK_6fef90f9-4ce5-456f-a170-152290f19313-part16'], 'labels': ['BOOT'], 'masters': [], 'uuids': ['09d53dc1-1e03-4286-bbb8-2b1796cf92ec']}, 'sectors': 1869825, 'sectorsize': 512, 'size': '913.00 MB', 'start': '227328', 'uuid': '09d53dc1-1e03-4286-bbb8-2b1796cf92ec'}}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 167772160, 'sectorsize': '512', 'size': '80.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'sr0', 'value': {'holders': [], 'host': 'IDE interface: Intel Corporation 82371SB PIIX3 IDE [Natoma/Triton II]', 'links': {'ids': ['ata-QEMU_DVD-ROM_QM00001'], 'labels': ['config-2'], 'masters': [], 'uuids': ['2026-02-24-00-02-01-00']}, 'model': 'QEMU DVD-ROM', 'partitions': {}, 'removable': '1', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'mq-deadline', 'sectors': 253, 'sectorsize': '2048', 'size': '506.00 KB', 'support_discard': '0', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop4', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop5', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop6', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop7', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'sda', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_35b8d51e-716e-455d-9a95-995756ae5900', 'scsi-SQEMU_QEMU_HARDDISK_35b8d51e-716e-455d-9a95-995756ae5900'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {'sda1': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_35b8d51e-716e-455d-9a95-995756ae5900-part1', 'scsi-SQEMU_QEMU_HARDDISK_35b8d51e-716e-455d-9a95-995756ae5900-part1'], 'labels': ['cloudimg-rootfs'], 'masters': [], 'uuids': ['b852d8d2-8460-44aa-8998-23e4f04d73cf']}, 'sectors': 165672927, 'sectorsize': 512, 'size': '79.00 GB', 'start': '2099200', 'uuid': 'b852d8d2-8460-44aa-8998-23e4f04d73cf'}, 'sda14': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_35b8d51e-716e-455d-9a95-995756ae5900-part14', 'scsi-SQEMU_QEMU_HARDDISK_35b8d51e-716e-455d-9a95-995756ae5900-part14'], 'labels': [], 'masters': [], 'uuids': []}, 'sectors': 8192, 'sectorsize': 512, 'size': '4.00 MB', 'start': '2048', 'uuid': None}, 'sda15': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_35b8d51e-716e-455d-9a95-995756ae5900-part15', 'scsi-SQEMU_QEMU_HARDDISK_35b8d51e-716e-455d-9a95-995756ae5900-part15'], 'labels': ['UEFI'], 'masters': [], 'uuids': ['5C78-612A']}, 'sectors': 217088, 'sectorsize': 512, 'size': '106.00 MB', 'start': '10240', 'uuid': '5C78-612A'}, 'sda16': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_35b8d51e-716e-455d-9a95-995756ae5900-part16', 'scsi-SQEMU_QEMU_HARDDISK_35b8d51e-716e-455d-9a95-995756ae5900-part16'], 'labels': ['BOOT'], 'masters': [], 'uuids': ['09d53dc1-1e03-4286-bbb8-2b1796cf92ec']}, 'sectors': 1869825, 'sectorsize': 512, 'size': '913.00 MB', 'start': '227328', 'uuid': '09d53dc1-1e03-4286-bbb8-2b1796cf92ec'}}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 167772160, 'sectorsize': '512', 'size': '80.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'sr0', 'value': {'holders': [], 'host': 'IDE interface: Intel Corporation 82371SB PIIX3 IDE [Natoma/Triton II]', 'links': {'ids': ['ata-QEMU_DVD-ROM_QM00001'], 'labels': ['config-2'], 'masters': [], 'uuids': ['2026-02-24-00-02-05-00']}, 'model': 'QEMU DVD-ROM', 'partitions': {}, 'removable': '1', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'mq-deadline', 'sectors': 253, 'sectorsize': '2048', 'size': '506.00 KB', 'support_discard': '0', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-facts : Check if the ceph conf exists] ******************************\nTuesday 24 February 2026  00:47:49 +0000 (0:00:01.257)       0:00:34.631 ****** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-facts : Set default osd_pool_default_crush_rule fact] ***************\nTuesday 24 February 2026  00:47:50 +0000 (0:00:01.654)       0:00:36.285 ****** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-facts : Read osd pool default crush rule] ***************************\nTuesday 24 February 2026  00:47:52 +0000 (0:00:01.251)       0:00:37.537 ****** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-facts : Set osd_pool_default_crush_rule fact] ***********************\nTuesday 24 February 2026  00:47:53 +0000 (0:00:01.080)       0:00:38.617 ****** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-facts : Read osd pool default crush rule] ***************************\nTuesday 24 February 2026  00:47:54 +0000 (0:00:00.863)       0:00:39.481 ****** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-facts : Set osd_pool_default_crush_rule fact] ***********************\nTuesday 24 February 2026  00:47:55 +0000 (0:00:01.067)       0:00:40.548 ****** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-facts : Set_fact _monitor_addresses - ipv4] *************************\nTuesday 24 February 2026  00:47:56 +0000 (0:00:01.307)       0:00:41.856 ****** \n\u001b[0;32mok: [testbed-node-4] => (item=testbed-node-0)\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item=testbed-node-0)\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item=testbed-node-1)\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item=testbed-node-0)\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item=testbed-node-2)\u001b[0m\n\u001b[0;32mok: [testbed-node-0] => (item=testbed-node-0)\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item=testbed-node-1)\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item=testbed-node-0)\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item=testbed-node-2)\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item=testbed-node-1)\u001b[0m\n\u001b[0;32mok: [testbed-node-0] => (item=testbed-node-1)\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item=testbed-node-0)\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item=testbed-node-1)\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item=testbed-node-2)\u001b[0m\n\u001b[0;32mok: [testbed-node-0] => (item=testbed-node-2)\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item=testbed-node-1)\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item=testbed-node-2)\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item=testbed-node-2)\u001b[0m\n\nTASK [ceph-facts : Set_fact _monitor_addresses - ipv6] *************************\nTuesday 24 February 2026  00:48:00 +0000 (0:00:03.976)       0:00:45.832 ****** \n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-0) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-1) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=testbed-node-0) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=testbed-node-1) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=testbed-node-0) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=testbed-node-2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=testbed-node-1) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=testbed-node-2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-0) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-1) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=testbed-node-0) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=testbed-node-1) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=testbed-node-2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=testbed-node-0) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=testbed-node-1) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=testbed-node-2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-facts : Import_tasks set_radosgw_address.yml] ***********************\nTuesday 24 February 2026  00:48:01 +0000 (0:00:01.151)       0:00:46.984 ****** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mincluded: /ansible/roles/ceph-facts/tasks/set_radosgw_address.yml for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m\n\nTASK [ceph-facts : Set current radosgw_address_block, radosgw_address, radosgw_interface  from node \"{{ ceph_dashboard_call_item }}\"] ***\nTuesday 24 February 2026  00:48:02 +0000 (0:00:01.304)       0:00:48.288 ****** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Set_fact _radosgw_address to radosgw_address_block ipv4] ****\nTuesday 24 February 2026  00:48:03 +0000 (0:00:00.621)       0:00:48.909 ****** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Set_fact _radosgw_address to radosgw_address_block ipv6] ****\nTuesday 24 February 2026  00:48:03 +0000 (0:00:00.367)       0:00:49.277 ****** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Set_fact _radosgw_address to radosgw_address] ***************\nTuesday 24 February 2026  00:48:04 +0000 (0:00:00.863)       0:00:50.141 ****** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Set_fact _interface] ****************************************\nTuesday 24 February 2026  00:48:05 +0000 (0:00:00.791)       0:00:50.932 ****** \n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-3) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-4) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-5) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-facts : Set_fact _radosgw_address to radosgw_interface - ipv4] ******\nTuesday 24 February 2026  00:48:06 +0000 (0:00:00.474)       0:00:51.406 ****** \n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-3) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-4) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-5) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-facts : Set_fact _radosgw_address to radosgw_interface - ipv6] ******\nTuesday 24 February 2026  00:48:06 +0000 (0:00:00.718)       0:00:52.124 ****** \n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-3) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-4) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-5) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-facts : Reset rgw_instances (workaround)] ***************************\nTuesday 24 February 2026  00:48:07 +0000 (0:00:00.851)       0:00:52.976 ****** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Set_fact rgw_instances] *************************************\nTuesday 24 February 2026  00:48:08 +0000 (0:00:00.656)       0:00:53.633 ****** \n\u001b[0;32mok: [testbed-node-4] => (item=0)\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item=0)\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item=0)\u001b[0m\n\nTASK [ceph-facts : Set_fact ceph_run_cmd] **************************************\nTuesday 24 February 2026  00:48:09 +0000 (0:00:01.533)       0:00:55.166 ****** \n\u001b[0;32mok: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item=testbed-node-0)\u001b[0m\n\u001b[0;32mok: [testbed-node-3 -> testbed-node-1(192.168.16.11)] => (item=testbed-node-1)\u001b[0m\n\u001b[0;32mok: [testbed-node-3 -> testbed-node-2(192.168.16.12)] => (item=testbed-node-2)\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item=testbed-node-3)\u001b[0m\n\u001b[0;32mok: [testbed-node-3 -> testbed-node-4(192.168.16.14)] => (item=testbed-node-4)\u001b[0m\n\u001b[0;32mok: [testbed-node-3 -> testbed-node-5(192.168.16.15)] => (item=testbed-node-5)\u001b[0m\n\u001b[0;32mok: [testbed-node-3 -> testbed-manager(192.168.16.5)] => (item=testbed-manager)\u001b[0m\n\nTASK [ceph-facts : Set_fact ceph_admin_command] ********************************\nTuesday 24 February 2026  00:48:11 +0000 (0:00:01.558)       0:00:56.725 ****** \n\u001b[0;32mok: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item=testbed-node-0)\u001b[0m\n\u001b[0;32mok: [testbed-node-3 -> testbed-node-1(192.168.16.11)] => (item=testbed-node-1)\u001b[0m\n\u001b[0;32mok: [testbed-node-3 -> testbed-node-2(192.168.16.12)] => (item=testbed-node-2)\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item=testbed-node-3)\u001b[0m\n\u001b[0;32mok: [testbed-node-3 -> testbed-node-4(192.168.16.14)] => (item=testbed-node-4)\u001b[0m\n\u001b[0;32mok: [testbed-node-3 -> testbed-node-5(192.168.16.15)] => (item=testbed-node-5)\u001b[0m\n\u001b[0;32mok: [testbed-node-3 -> testbed-manager(192.168.16.5)] => (item=testbed-manager)\u001b[0m\n\nTASK [ceph-handler : Include check_running_cluster.yml] ************************\nTuesday 24 February 2026  00:48:13 +0000 (0:00:02.119)       0:00:58.845 ****** \n\u001b[0;36mincluded: /ansible/roles/ceph-handler/tasks/check_running_cluster.yml for testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [ceph-handler : Include check_running_containers.yml] *********************\nTuesday 24 February 2026  00:48:14 +0000 (0:00:01.000)       0:00:59.845 ****** \n\u001b[0;36mincluded: /ansible/roles/ceph-handler/tasks/check_running_containers.yml for testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [ceph-handler : Check for a mon container] ********************************\nTuesday 24 February 2026  00:48:15 +0000 (0:00:01.165)       0:01:01.011 ****** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Check for an osd container] *******************************\nTuesday 24 February 2026  00:48:16 +0000 (0:00:01.167)       0:01:02.178 ****** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Check for a mds container] ********************************\nTuesday 24 February 2026  00:48:17 +0000 (0:00:00.863)       0:01:03.042 ****** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Check for a rgw container] ********************************\nTuesday 24 February 2026  00:48:18 +0000 (0:00:00.699)       0:01:03.742 ****** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Check for a mgr container] ********************************\nTuesday 24 February 2026  00:48:19 +0000 (0:00:00.783)       0:01:04.525 ****** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Check for a rbd mirror container] *************************\nTuesday 24 February 2026  00:48:20 +0000 (0:00:01.199)       0:01:05.725 ****** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Check for a nfs container] ********************************\nTuesday 24 February 2026  00:48:21 +0000 (0:00:00.689)       0:01:06.414 ****** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Check for a ceph-crash container] *************************\nTuesday 24 February 2026  00:48:21 +0000 (0:00:00.724)       0:01:07.138 ****** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Check for a ceph-exporter container] **********************\nTuesday 24 February 2026  00:48:22 +0000 (0:00:01.029)       0:01:08.168 ****** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Include check_socket_non_container.yml] *******************\nTuesday 24 February 2026  00:48:23 +0000 (0:00:01.139)       0:01:09.308 ****** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_mon_status] ******************************\nTuesday 24 February 2026  00:48:24 +0000 (0:00:00.519)       0:01:09.828 ****** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_osd_status] ******************************\nTuesday 24 February 2026  00:48:25 +0000 (0:00:00.774)       0:01:10.602 ****** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_mds_status] ******************************\nTuesday 24 February 2026  00:48:25 +0000 (0:00:00.542)       0:01:11.144 ****** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_rgw_status] ******************************\nTuesday 24 February 2026  00:48:26 +0000 (0:00:00.734)       0:01:11.879 ****** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_nfs_status] ******************************\nTuesday 24 February 2026  00:48:27 +0000 (0:00:00.864)       0:01:12.743 ****** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_rbd_status] ******************************\nTuesday 24 February 2026  00:48:28 +0000 (0:00:00.962)       0:01:13.706 ****** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_mgr_status] ******************************\nTuesday 24 February 2026  00:48:28 +0000 (0:00:00.629)       0:01:14.335 ****** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_crash_status] ****************************\nTuesday 24 February 2026  00:48:29 +0000 (0:00:00.830)       0:01:15.165 ****** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_exporter_status] *************************\nTuesday 24 February 2026  00:48:30 +0000 (0:00:00.755)       0:01:15.921 ****** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-container-common : Generate systemd ceph target file] ***************\nTuesday 24 February 2026  00:48:32 +0000 (0:00:01.558)       0:01:17.479 ****** \n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\nTASK [ceph-container-common : Enable ceph.target] ******************************\nTuesday 24 February 2026  00:48:34 +0000 (0:00:02.107)       0:01:19.587 ****** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\nTASK [ceph-container-common : Include prerequisites.yml] ***********************\nTuesday 24 February 2026  00:48:37 +0000 (0:00:03.160)       0:01:22.748 ****** \n\u001b[0;36mincluded: /ansible/roles/ceph-container-common/tasks/prerequisites.yml for testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [ceph-container-common : Stop lvmetad] ************************************\nTuesday 24 February 2026  00:48:38 +0000 (0:00:01.606)       0:01:24.354 ****** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-container-common : Disable and mask lvmetad service] ****************\nTuesday 24 February 2026  00:48:39 +0000 (0:00:00.682)       0:01:25.037 ****** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-container-common : Remove ceph udev rules] **************************\nTuesday 24 February 2026  00:48:40 +0000 (0:00:01.144)       0:01:26.182 ****** \n\u001b[0;32mok: [testbed-node-3] => (item=/usr/lib/udev/rules.d/95-ceph-osd.rules)\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item=/usr/lib/udev/rules.d/95-ceph-osd.rules)\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item=/usr/lib/udev/rules.d/95-ceph-osd.rules)\u001b[0m\n\u001b[0;32mok: [testbed-node-0] => (item=/usr/lib/udev/rules.d/95-ceph-osd.rules)\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item=/usr/lib/udev/rules.d/60-ceph-by-parttypeuuid.rules)\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item=/usr/lib/udev/rules.d/60-ceph-by-parttypeuuid.rules)\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item=/usr/lib/udev/rules.d/95-ceph-osd.rules)\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item=/usr/lib/udev/rules.d/60-ceph-by-parttypeuuid.rules)\u001b[0m\n\u001b[0;32mok: [testbed-node-0] => (item=/usr/lib/udev/rules.d/60-ceph-by-parttypeuuid.rules)\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item=/usr/lib/udev/rules.d/95-ceph-osd.rules)\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item=/usr/lib/udev/rules.d/60-ceph-by-parttypeuuid.rules)\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item=/usr/lib/udev/rules.d/60-ceph-by-parttypeuuid.rules)\u001b[0m\n\nTASK [ceph-container-common : Ensure tmpfiles.d is present] ********************\nTuesday 24 February 2026  00:48:42 +0000 (0:00:01.601)       0:01:27.783 ****** \n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [ceph-container-common : Restore certificates selinux context] ************\nTuesday 24 February 2026  00:48:43 +0000 (0:00:01.176)       0:01:28.960 ****** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-container-common : Install python3 on osd nodes] ********************\nTuesday 24 February 2026  00:48:44 +0000 (0:00:00.521)       0:01:29.481 ****** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-container-common : Include registry.yml] ****************************\nTuesday 24 February 2026  00:48:44 +0000 (0:00:00.825)       0:01:30.306 ****** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-container-common : Include fetch_image.yml] *************************\nTuesday 24 February 2026  00:48:45 +0000 (0:00:00.512)       0:01:30.819 ****** \n\u001b[0;36mincluded: /ansible/roles/ceph-container-common/tasks/fetch_image.yml for testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [ceph-container-common : Pulling Ceph container image] ********************\nTuesday 24 February 2026  00:48:46 +0000 (0:00:01.045)       0:01:31.865 ****** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [ceph-container-common : Pulling alertmanager/prometheus/grafana container images] ***\nTuesday 24 February 2026  00:49:26 +0000 (0:00:40.427)       0:02:12.292 ****** \n\u001b[0;36mskipping: [testbed-node-3] => (item=docker.io/prom/alertmanager:v0.16.2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=docker.io/prom/prometheus:v2.7.2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=docker.io/grafana/grafana:6.7.4) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=docker.io/prom/alertmanager:v0.16.2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=docker.io/prom/prometheus:v2.7.2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=docker.io/grafana/grafana:6.7.4) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=docker.io/prom/alertmanager:v0.16.2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=docker.io/prom/prometheus:v2.7.2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=docker.io/grafana/grafana:6.7.4) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=docker.io/prom/alertmanager:v0.16.2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=docker.io/prom/prometheus:v2.7.2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=docker.io/grafana/grafana:6.7.4) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=docker.io/prom/alertmanager:v0.16.2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=docker.io/prom/prometheus:v2.7.2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=docker.io/grafana/grafana:6.7.4) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=docker.io/prom/alertmanager:v0.16.2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=docker.io/prom/prometheus:v2.7.2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=docker.io/grafana/grafana:6.7.4) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-container-common : Pulling node-exporter container image] ***********\nTuesday 24 February 2026  00:49:27 +0000 (0:00:00.585)       0:02:12.877 ****** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-container-common : Export local ceph dev image] *********************\nTuesday 24 February 2026  00:49:28 +0000 (0:00:00.635)       0:02:13.512 ****** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-container-common : Copy ceph dev image file] ************************\nTuesday 24 February 2026  00:49:28 +0000 (0:00:00.136)       0:02:13.649 ****** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-container-common : Load ceph dev image] *****************************\nTuesday 24 February 2026  00:49:28 +0000 (0:00:00.520)       0:02:14.169 ****** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-container-common : Remove tmp ceph dev image file] ******************\nTuesday 24 February 2026  00:49:29 +0000 (0:00:00.795)       0:02:14.965 ****** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-container-common : Get ceph version] ********************************\nTuesday 24 February 2026  00:49:30 +0000 (0:00:00.597)       0:02:15.563 ****** \n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-container-common : Set_fact ceph_version ceph_version.stdout.split] ***\nTuesday 24 February 2026  00:49:32 +0000 (0:00:02.522)       0:02:18.086 ****** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-container-common : Include release.yml] *****************************\nTuesday 24 February 2026  00:49:33 +0000 (0:00:00.633)       0:02:18.719 ****** \n\u001b[0;36mincluded: /ansible/roles/ceph-container-common/tasks/release.yml for testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [ceph-container-common : Set_fact ceph_release jewel] *********************\nTuesday 24 February 2026  00:49:34 +0000 (0:00:01.159)       0:02:19.878 ****** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-container-common : Set_fact ceph_release kraken] ********************\nTuesday 24 February 2026  00:49:35 +0000 (0:00:00.656)       0:02:20.535 ****** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-container-common : Set_fact ceph_release luminous] ******************\nTuesday 24 February 2026  00:49:35 +0000 (0:00:00.538)       0:02:21.074 ****** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-container-common : Set_fact ceph_release mimic] *********************\nTuesday 24 February 2026  00:49:36 +0000 (0:00:00.681)       0:02:21.755 ****** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-container-common : Set_fact ceph_release nautilus] ******************\nTuesday 24 February 2026  00:49:37 +0000 (0:00:00.658)       0:02:22.413 ****** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-container-common : Set_fact ceph_release octopus] *******************\nTuesday 24 February 2026  00:49:37 +0000 (0:00:00.641)       0:02:23.055 ****** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-container-common : Set_fact ceph_release pacific] *******************\nTuesday 24 February 2026  00:49:38 +0000 (0:00:00.554)       0:02:23.609 ****** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-container-common : Set_fact ceph_release quincy] ********************\nTuesday 24 February 2026  00:49:38 +0000 (0:00:00.676)       0:02:24.286 ****** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [ceph-container-common : Set_fact ceph_release reef] **********************\nTuesday 24 February 2026  00:49:39 +0000 (0:00:00.883)       0:02:25.170 ****** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-config : Include create_ceph_initial_dirs.yml] **********************\nTuesday 24 February 2026  00:49:40 +0000 (0:00:01.073)       0:02:26.243 ****** \n\u001b[0;36mincluded: /ansible/roles/ceph-config/tasks/create_ceph_initial_dirs.yml for testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [ceph-config : Create ceph initial directories] ***************************\nTuesday 24 February 2026  00:49:41 +0000 (0:00:00.943)       0:02:27.187 ****** \n\u001b[0;33mchanged: [testbed-node-3] => (item=/etc/ceph)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=/etc/ceph)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=/etc/ceph)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=/etc/ceph)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=/etc/ceph)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=/etc/ceph)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item=/var/lib/ceph/)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=/var/lib/ceph/)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=/var/lib/ceph/)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=/var/lib/ceph/)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=/var/lib/ceph/)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=/var/lib/ceph/)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=/var/lib/ceph/mon)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item=/var/lib/ceph/mon)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=/var/lib/ceph/mon)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=/var/lib/ceph/mon)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=/var/lib/ceph/mon)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=/var/lib/ceph/mon)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=/var/lib/ceph/osd)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item=/var/lib/ceph/osd)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=/var/lib/ceph/osd)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=/var/lib/ceph/osd)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=/var/lib/ceph/osd)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=/var/lib/ceph/osd)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=/var/lib/ceph/mds)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item=/var/lib/ceph/mds)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=/var/lib/ceph/mds)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=/var/lib/ceph/mds)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=/var/lib/ceph/mds)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=/var/lib/ceph/mds)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=/var/lib/ceph/tmp)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item=/var/lib/ceph/tmp)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=/var/lib/ceph/tmp)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=/var/lib/ceph/tmp)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=/var/lib/ceph/tmp)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=/var/lib/ceph/tmp)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=/var/lib/ceph/crash)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item=/var/lib/ceph/crash)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=/var/lib/ceph/crash)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=/var/lib/ceph/crash)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=/var/lib/ceph/crash)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=/var/lib/ceph/radosgw)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item=/var/lib/ceph/radosgw)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=/var/lib/ceph/radosgw)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=/var/lib/ceph/radosgw)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=/var/lib/ceph/crash)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=/var/lib/ceph/bootstrap-rgw)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=/var/lib/ceph/radosgw)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=/var/lib/ceph/bootstrap-rgw)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item=/var/lib/ceph/bootstrap-rgw)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=/var/lib/ceph/bootstrap-rgw)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=/var/lib/ceph/radosgw)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=/var/lib/ceph/bootstrap-rgw)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=/var/lib/ceph/bootstrap-mgr)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item=/var/lib/ceph/bootstrap-mgr)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=/var/lib/ceph/bootstrap-mgr)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=/var/lib/ceph/bootstrap-mgr)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=/var/lib/ceph/bootstrap-rgw)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=/var/lib/ceph/bootstrap-mgr)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item=/var/lib/ceph/bootstrap-mds)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=/var/lib/ceph/bootstrap-mds)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=/var/lib/ceph/bootstrap-mds)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=/var/lib/ceph/bootstrap-mds)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=/var/lib/ceph/bootstrap-mgr)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=/var/lib/ceph/bootstrap-mds)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=/var/lib/ceph/bootstrap-osd)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=/var/lib/ceph/bootstrap-osd)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item=/var/lib/ceph/bootstrap-osd)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=/var/lib/ceph/bootstrap-osd)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=/var/lib/ceph/bootstrap-osd)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=/var/lib/ceph/bootstrap-mds)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=/var/lib/ceph/bootstrap-rbd)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=/var/lib/ceph/bootstrap-rbd)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item=/var/lib/ceph/bootstrap-rbd)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=/var/lib/ceph/bootstrap-rbd)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=/var/lib/ceph/bootstrap-osd)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=/var/lib/ceph/bootstrap-rbd)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=/var/lib/ceph/bootstrap-rbd-mirror)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=/var/lib/ceph/bootstrap-rbd-mirror)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item=/var/lib/ceph/bootstrap-rbd-mirror)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=/var/lib/ceph/bootstrap-rbd-mirror)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=/var/lib/ceph/bootstrap-rbd-mirror)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=/var/lib/ceph/bootstrap-rbd)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=/var/run/ceph)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=/var/run/ceph)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item=/var/run/ceph)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=/var/run/ceph)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=/var/run/ceph)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=/var/log/ceph)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=/var/lib/ceph/bootstrap-rbd-mirror)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=/var/log/ceph)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=/var/log/ceph)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item=/var/log/ceph)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=/var/log/ceph)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=/var/run/ceph)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=/var/log/ceph)\u001b[0m\n\nTASK [ceph-config : Include_tasks rgw_systemd_environment_file.yml] ************\nTuesday 24 February 2026  00:49:48 +0000 (0:00:06.659)       0:02:33.847 ****** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mincluded: /ansible/roles/ceph-config/tasks/rgw_systemd_environment_file.yml for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m\n\nTASK [ceph-config : Create rados gateway instance directories] *****************\nTuesday 24 February 2026  00:49:49 +0000 (0:00:00.865)       0:02:34.713 ****** \n\u001b[0;33mchanged: [testbed-node-3] => (item={'instance_name': 'rgw0', 'radosgw_address': '192.168.16.13', 'radosgw_frontend_port': 8081})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'instance_name': 'rgw0', 'radosgw_address': '192.168.16.14', 'radosgw_frontend_port': 8081})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'instance_name': 'rgw0', 'radosgw_address': '192.168.16.15', 'radosgw_frontend_port': 8081})\u001b[0m\n\nTASK [ceph-config : Generate environment file] *********************************\nTuesday 24 February 2026  00:49:50 +0000 (0:00:00.933)       0:02:35.647 ****** \n\u001b[0;33mchanged: [testbed-node-3] => (item={'instance_name': 'rgw0', 'radosgw_address': '192.168.16.13', 'radosgw_frontend_port': 8081})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'instance_name': 'rgw0', 'radosgw_address': '192.168.16.14', 'radosgw_frontend_port': 8081})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'instance_name': 'rgw0', 'radosgw_address': '192.168.16.15', 'radosgw_frontend_port': 8081})\u001b[0m\n\nTASK [ceph-config : Reset num_osds] ********************************************\nTuesday 24 February 2026  00:49:51 +0000 (0:00:01.067)       0:02:36.715 ****** \n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-config : Count number of osds for lvm scenario] *********************\nTuesday 24 February 2026  00:49:52 +0000 (0:00:00.715)       0:02:37.430 ****** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-config : Look up for ceph-volume rejected devices] ******************\nTuesday 24 February 2026  00:49:52 +0000 (0:00:00.870)       0:02:38.301 ****** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-config : Set_fact rejected_devices] *********************************\nTuesday 24 February 2026  00:49:53 +0000 (0:00:00.652)       0:02:38.953 ****** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-config : Set_fact _devices] *****************************************\nTuesday 24 February 2026  00:49:54 +0000 (0:00:00.605)       0:02:39.559 ****** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-config : Run 'ceph-volume lvm batch --report' to see how many osds are to be created] ***\nTuesday 24 February 2026  00:49:54 +0000 (0:00:00.550)       0:02:40.109 ****** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-config : Set_fact num_osds from the output of 'ceph-volume lvm batch --report' (legacy report)] ***\nTuesday 24 February 2026  00:49:55 +0000 (0:00:00.576)       0:02:40.686 ****** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-config : Set_fact num_osds from the output of 'ceph-volume lvm batch --report' (new report)] ***\nTuesday 24 February 2026  00:49:55 +0000 (0:00:00.486)       0:02:41.172 ****** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-config : Run 'ceph-volume lvm list' to see how many osds have already been created] ***\nTuesday 24 February 2026  00:49:56 +0000 (0:00:00.578)       0:02:41.751 ****** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-config : Set_fact num_osds (add existing osds)] *********************\nTuesday 24 February 2026  00:49:59 +0000 (0:00:02.995)       0:02:44.746 ****** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-config : Set_fact _osd_memory_target] *******************************\nTuesday 24 February 2026  00:50:00 +0000 (0:00:00.780)       0:02:45.527 ****** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-config : Set osd_memory_target to cluster host config] **************\nTuesday 24 February 2026  00:50:00 +0000 (0:00:00.538)       0:02:46.065 ****** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-config : Render rgw configs] ****************************************\nTuesday 24 February 2026  00:50:01 +0000 (0:00:00.763)       0:02:46.829 ****** \n\u001b[0;32mok: [testbed-node-3] => (item={'instance_name': 'rgw0', 'radosgw_address': '192.168.16.13', 'radosgw_frontend_port': 8081})\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item={'instance_name': 'rgw0', 'radosgw_address': '192.168.16.14', 'radosgw_frontend_port': 8081})\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item={'instance_name': 'rgw0', 'radosgw_address': '192.168.16.15', 'radosgw_frontend_port': 8081})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [ceph-config : Set config to cluster] *************************************\nTuesday 24 February 2026  00:50:02 +0000 (0:00:00.628)       0:02:47.457 ****** \n\u001b[0;36mskipping: [testbed-node-3] => (item=[{'key': 'client.rgw.default.testbed-node-3.rgw0', 'value': {'log_file': '/var/log/ceph/ceph-rgw-default-testbed-node-3.rgw0.log', 'rgw_frontends': 'beast endpoint=192.168.16.13:8081'}}, {'key': 'log_file', 'value': '/var/log/ceph/ceph-rgw-default-testbed-node-3.rgw0.log'}]) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=[{'key': 'client.rgw.default.testbed-node-3.rgw0', 'value': {'log_file': '/var/log/ceph/ceph-rgw-default-testbed-node-3.rgw0.log', 'rgw_frontends': 'beast endpoint=192.168.16.13:8081'}}, {'key': 'rgw_frontends', 'value': 'beast endpoint=192.168.16.13:8081'}]) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=[{'key': 'client.rgw.default.testbed-node-4.rgw0', 'value': {'log_file': '/var/log/ceph/ceph-rgw-default-testbed-node-4.rgw0.log', 'rgw_frontends': 'beast endpoint=192.168.16.14:8081'}}, {'key': 'log_file', 'value': '/var/log/ceph/ceph-rgw-default-testbed-node-4.rgw0.log'}]) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=[{'key': 'client.rgw.default.testbed-node-4.rgw0', 'value': {'log_file': '/var/log/ceph/ceph-rgw-default-testbed-node-4.rgw0.log', 'rgw_frontends': 'beast endpoint=192.168.16.14:8081'}}, {'key': 'rgw_frontends', 'value': 'beast endpoint=192.168.16.14:8081'}]) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=[{'key': 'client.rgw.default.testbed-node-5.rgw0', 'value': {'log_file': '/var/log/ceph/ceph-rgw-default-testbed-node-5.rgw0.log', 'rgw_frontends': 'beast endpoint=192.168.16.15:8081'}}, {'key': 'log_file', 'value': '/var/log/ceph/ceph-rgw-default-testbed-node-5.rgw0.log'}]) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=[{'key': 'client.rgw.default.testbed-node-5.rgw0', 'value': {'log_file': '/var/log/ceph/ceph-rgw-default-testbed-node-5.rgw0.log', 'rgw_frontends': 'beast endpoint=192.168.16.15:8081'}}, {'key': 'rgw_frontends', 'value': 'beast endpoint=192.168.16.15:8081'}]) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-config : Set rgw configs to file] ***********************************\nTuesday 24 February 2026  00:50:02 +0000 (0:00:00.687)       0:02:48.145 ****** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-config : Create ceph conf directory] ********************************\nTuesday 24 February 2026  00:50:03 +0000 (0:00:00.744)       0:02:48.890 ****** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-facts : Set current radosgw_address_block, radosgw_address, radosgw_interface  from node \"{{ ceph_dashboard_call_item }}\"] ***\nTuesday 24 February 2026  00:50:04 +0000 (0:00:00.953)       0:02:49.843 ****** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-facts : Set_fact _radosgw_address to radosgw_address_block ipv4] ****\nTuesday 24 February 2026  00:50:05 +0000 (0:00:00.716)       0:02:50.559 ****** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-facts : Set_fact _radosgw_address to radosgw_address_block ipv6] ****\nTuesday 24 February 2026  00:50:06 +0000 (0:00:00.998)       0:02:51.558 ****** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-facts : Set_fact _radosgw_address to radosgw_address] ***************\nTuesday 24 February 2026  00:50:07 +0000 (0:00:00.838)       0:02:52.396 ****** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-facts : Set_fact _interface] ****************************************\nTuesday 24 February 2026  00:50:08 +0000 (0:00:01.233)       0:02:53.630 ****** \n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-3) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-4) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-5) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-facts : Set_fact _radosgw_address to radosgw_interface - ipv4] ******\nTuesday 24 February 2026  00:50:08 +0000 (0:00:00.358)       0:02:53.988 ****** \n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-3) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-4) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-5) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-facts : Set_fact _radosgw_address to radosgw_interface - ipv6] ******\nTuesday 24 February 2026  00:50:08 +0000 (0:00:00.351)       0:02:54.339 ****** \n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-3) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-4) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-5) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-facts : Reset rgw_instances (workaround)] ***************************\nTuesday 24 February 2026  00:50:09 +0000 (0:00:00.355)       0:02:54.694 ****** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-facts : Set_fact rgw_instances] *************************************\nTuesday 24 February 2026  00:50:09 +0000 (0:00:00.439)       0:02:55.134 ****** \n\u001b[0;36mskipping: [testbed-node-0] => (item=0) \u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item=0)\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item=0)\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=0) \u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item=0)\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=0) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-config : Generate Ceph file] ****************************************\nTuesday 24 February 2026  00:50:11 +0000 (0:00:01.794)       0:02:56.929 ****** \n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Make tempdir for scripts] **********************\nTuesday 24 February 2026  00:50:14 +0000 (0:00:03.310)       0:03:00.240 ****** \n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Mons handler] **********************************\nTuesday 24 February 2026  00:50:16 +0000 (0:00:01.379)       0:03:01.619 ****** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mincluded: /ansible/roles/ceph-handler/tasks/handler_mons.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Set _mon_handler_called before restart] ********\nTuesday 24 February 2026  00:50:17 +0000 (0:00:00.940)       0:03:02.560 ****** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Copy mon restart script] ***********************\nTuesday 24 February 2026  00:50:17 +0000 (0:00:00.281)       0:03:02.841 ****** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Restart ceph mon daemon(s)] ********************\nTuesday 24 February 2026  00:50:18 +0000 (0:00:01.312)       0:03:04.154 ****** \n\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-0) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-1) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Set _mon_handler_called after restart] *********\nTuesday 24 February 2026  00:50:19 +0000 (0:00:00.950)       0:03:05.104 ****** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Osds handler] **********************************\nTuesday 24 February 2026  00:50:20 +0000 (0:00:00.312)       0:03:05.416 ****** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mincluded: /ansible/roles/ceph-handler/tasks/handler_osds.yml for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Set_fact trigger_restart] **********************\nTuesday 24 February 2026  00:50:21 +0000 (0:00:00.992)       0:03:06.409 ****** \n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-3) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-4) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-5) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Set _osd_handler_called before restart] ********\nTuesday 24 February 2026  00:50:21 +0000 (0:00:00.419)       0:03:06.828 ****** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Unset noup flag] *******************************\nTuesday 24 February 2026  00:50:21 +0000 (0:00:00.378)       0:03:07.207 ****** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Copy osd restart script] ***********************\nTuesday 24 February 2026  00:50:22 +0000 (0:00:00.218)       0:03:07.425 ****** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Get pool list] *********************************\nTuesday 24 February 2026  00:50:22 +0000 (0:00:00.322)       0:03:07.748 ****** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Get balancer module status] ********************\nTuesday 24 February 2026  00:50:22 +0000 (0:00:00.220)       0:03:07.968 ****** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Set_fact pools_pgautoscaler_mode] **************\nTuesday 24 February 2026  00:50:22 +0000 (0:00:00.267)       0:03:08.236 ****** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Disable balancer] ******************************\nTuesday 24 February 2026  00:50:23 +0000 (0:00:00.149)       0:03:08.385 ****** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Disable pg autoscale on pools] *****************\nTuesday 24 February 2026  00:50:23 +0000 (0:00:00.750)       0:03:09.135 ****** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Restart ceph osds daemon(s)] *******************\nTuesday 24 February 2026  00:50:24 +0000 (0:00:00.238)       0:03:09.374 ****** \n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-3) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-4) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-5) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Set _osd_handler_called after restart] *********\nTuesday 24 February 2026  00:50:24 +0000 (0:00:00.440)       0:03:09.814 ****** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Re-enable pg autoscale on pools] ***************\nTuesday 24 February 2026  00:50:24 +0000 (0:00:00.336)       0:03:10.151 ****** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Re-enable balancer] ****************************\nTuesday 24 February 2026  00:50:25 +0000 (0:00:00.245)       0:03:10.396 ****** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Mdss handler] **********************************\nTuesday 24 February 2026  00:50:25 +0000 (0:00:00.205)       0:03:10.602 ****** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mincluded: /ansible/roles/ceph-handler/tasks/handler_mdss.yml for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Set _mds_handler_called before restart] ********\nTuesday 24 February 2026  00:50:26 +0000 (0:00:01.091)       0:03:11.694 ****** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Copy mds restart script] ***********************\nTuesday 24 February 2026  00:50:26 +0000 (0:00:00.344)       0:03:12.038 ****** \n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Restart ceph mds daemon(s)] ********************\nTuesday 24 February 2026  00:50:27 +0000 (0:00:01.250)       0:03:13.289 ****** \n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-3) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-4) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-5) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Set _mds_handler_called after restart] *********\nTuesday 24 February 2026  00:50:28 +0000 (0:00:00.756)       0:03:14.046 ****** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Rgws handler] **********************************\nTuesday 24 February 2026  00:50:29 +0000 (0:00:00.412)       0:03:14.458 ****** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mincluded: /ansible/roles/ceph-handler/tasks/handler_rgws.yml for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Set _rgw_handler_called before restart] ********\nTuesday 24 February 2026  00:50:30 +0000 (0:00:01.223)       0:03:15.681 ****** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Copy rgw restart script] ***********************\nTuesday 24 February 2026  00:50:30 +0000 (0:00:00.475)       0:03:16.157 ****** \n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Restart ceph rgw daemon(s)] ********************\nTuesday 24 February 2026  00:50:32 +0000 (0:00:02.006)       0:03:18.163 ****** \n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-3) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-4) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-5) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Set _rgw_handler_called after restart] *********\nTuesday 24 February 2026  00:50:33 +0000 (0:00:00.527)       0:03:18.691 ****** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Rbdmirrors handler] ****************************\nTuesday 24 February 2026  00:50:33 +0000 (0:00:00.302)       0:03:18.993 ****** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Mgrs handler] **********************************\nTuesday 24 February 2026  00:50:34 +0000 (0:00:00.668)       0:03:19.662 ****** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mincluded: /ansible/roles/ceph-handler/tasks/handler_mgrs.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Set _mgr_handler_called before restart] ********\nTuesday 24 February 2026  00:50:35 +0000 (0:00:00.754)       0:03:20.417 ****** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Copy mgr restart script] ***********************\nTuesday 24 February 2026  00:50:35 +0000 (0:00:00.543)       0:03:20.961 ****** \n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Restart ceph mgr daemon(s)] ********************\nTuesday 24 February 2026  00:50:36 +0000 (0:00:01.261)       0:03:22.222 ****** \n\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-0) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-1) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Set _mgr_handler_called after restart] *********\nTuesday 24 February 2026  00:50:37 +0000 (0:00:00.600)       0:03:22.823 ****** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nPLAY [Apply role ceph-mon] *****************************************************\n\nTASK [ceph-handler : Include check_running_cluster.yml] ************************\nTuesday 24 February 2026  00:50:38 +0000 (0:00:00.672)       0:03:23.496 ****** \n\u001b[0;36mincluded: /ansible/roles/ceph-handler/tasks/check_running_cluster.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [ceph-handler : Include check_running_containers.yml] *********************\nTuesday 24 February 2026  00:50:38 +0000 (0:00:00.742)       0:03:24.239 ****** \n\u001b[0;36mincluded: /ansible/roles/ceph-handler/tasks/check_running_containers.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [ceph-handler : Check for a mon container] ********************************\nTuesday 24 February 2026  00:50:39 +0000 (0:00:00.536)       0:03:24.776 ****** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Check for an osd container] *******************************\nTuesday 24 February 2026  00:50:40 +0000 (0:00:00.951)       0:03:25.728 ****** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Check for a mds container] ********************************\nTuesday 24 February 2026  00:50:40 +0000 (0:00:00.320)       0:03:26.048 ****** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Check for a rgw container] ********************************\nTuesday 24 February 2026  00:50:40 +0000 (0:00:00.269)       0:03:26.318 ****** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Check for a mgr container] ********************************\nTuesday 24 February 2026  00:50:41 +0000 (0:00:00.298)       0:03:26.617 ****** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Check for a rbd mirror container] *************************\nTuesday 24 February 2026  00:50:42 +0000 (0:00:00.929)       0:03:27.546 ****** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Check for a nfs container] ********************************\nTuesday 24 February 2026  00:50:42 +0000 (0:00:00.272)       0:03:27.819 ****** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Check for a ceph-crash container] *************************\nTuesday 24 February 2026  00:50:42 +0000 (0:00:00.250)       0:03:28.069 ****** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Check for a ceph-exporter container] **********************\nTuesday 24 February 2026  00:50:43 +0000 (0:00:00.630)       0:03:28.700 ****** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Include check_socket_non_container.yml] *******************\nTuesday 24 February 2026  00:50:44 +0000 (0:00:00.898)       0:03:29.599 ****** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_mon_status] ******************************\nTuesday 24 February 2026  00:50:44 +0000 (0:00:00.263)       0:03:29.863 ****** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_osd_status] ******************************\nTuesday 24 February 2026  00:50:44 +0000 (0:00:00.379)       0:03:30.242 ****** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_mds_status] ******************************\nTuesday 24 February 2026  00:50:45 +0000 (0:00:00.227)       0:03:30.470 ****** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_rgw_status] ******************************\nTuesday 24 February 2026  00:50:45 +0000 (0:00:00.253)       0:03:30.723 ****** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_nfs_status] ******************************\nTuesday 24 February 2026  00:50:45 +0000 (0:00:00.431)       0:03:31.155 ****** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_rbd_status] ******************************\nTuesday 24 February 2026  00:50:46 +0000 (0:00:00.272)       0:03:31.428 ****** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_mgr_status] ******************************\nTuesday 24 February 2026  00:50:46 +0000 (0:00:00.502)       0:03:31.930 ****** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_crash_status] ****************************\nTuesday 24 February 2026  00:50:46 +0000 (0:00:00.330)       0:03:32.261 ****** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_exporter_status] *************************\nTuesday 24 February 2026  00:50:47 +0000 (0:00:00.514)       0:03:32.775 ****** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-mon : Set_fact container_exec_cmd] **********************************\nTuesday 24 February 2026  00:50:47 +0000 (0:00:00.516)       0:03:33.292 ****** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-mon : Include deploy_monitors.yml] **********************************\nTuesday 24 February 2026  00:50:48 +0000 (0:00:00.292)       0:03:33.584 ****** \n\u001b[0;36mincluded: /ansible/roles/ceph-mon/tasks/deploy_monitors.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [ceph-mon : Check if monitor initial keyring already exists] **************\nTuesday 24 February 2026  00:50:48 +0000 (0:00:00.699)       0:03:34.283 ****** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [ceph-mon : Generate monitor initial keyring] *****************************\nTuesday 24 February 2026  00:50:49 +0000 (0:00:00.112)       0:03:34.395 ****** \n\u001b[0;33mchanged: [testbed-node-0 -> localhost]\u001b[0m\n\nTASK [ceph-mon : Set_fact _initial_mon_key_success] ****************************\nTuesday 24 February 2026  00:50:49 +0000 (0:00:00.790)       0:03:35.186 ****** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-mon : Get initial keyring when it already exists] *******************\nTuesday 24 February 2026  00:50:50 +0000 (0:00:00.299)       0:03:35.486 ****** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-mon : Create monitor initial keyring] *******************************\nTuesday 24 February 2026  00:50:50 +0000 (0:00:00.276)       0:03:35.763 ****** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [ceph-mon : Copy the initial key in /etc/ceph (for containers)] ***********\nTuesday 24 February 2026  00:50:51 +0000 (0:00:01.342)       0:03:37.106 ****** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [ceph-mon : Create monitor directory] *************************************\nTuesday 24 February 2026  00:50:52 +0000 (0:00:00.750)       0:03:37.856 ****** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\nTASK [ceph-mon : Recursively fix ownership of monitor directory] ***************\nTuesday 24 February 2026  00:50:53 +0000 (0:00:00.692)       0:03:38.548 ****** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-mon : Create admin keyring] *****************************************\nTuesday 24 February 2026  00:50:54 +0000 (0:00:01.253)       0:03:39.803 ****** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [ceph-mon : Slurp admin keyring] ******************************************\nTuesday 24 February 2026  00:50:56 +0000 (0:00:01.678)       0:03:41.481 ****** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [ceph-mon : Copy admin keyring over to mons] ******************************\nTuesday 24 February 2026  00:50:56 +0000 (0:00:00.637)       0:03:42.119 ****** \n\u001b[0;33mchanged: [testbed-node-0] => (item=None)\u001b[0m\n\u001b[0;32mok: [testbed-node-1 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m\n\u001b[0;32mok: [testbed-node-2 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=None)\u001b[0m\n\u001b[0;32mok: [testbed-node-0 -> testbed-node-1(192.168.16.11)] => (item=None)\u001b[0m\n\u001b[0;32mok: [testbed-node-2 -> testbed-node-1(192.168.16.11)] => (item=None)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0 -> testbed-node-2(192.168.16.12)] => (item=None)\u001b[0m\n\u001b[0;32mok: [testbed-node-1 -> testbed-node-2(192.168.16.12)] => (item=None)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0 -> {{ item }}]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1 -> {{ item }}]\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item=None)\u001b[0m\n\u001b[0;32mok: [testbed-node-2 -> {{ item }}]\u001b[0m\n\nTASK [ceph-mon : Import admin keyring into mon keyring] ************************\nTuesday 24 February 2026  00:51:00 +0000 (0:00:03.529)       0:03:45.649 ****** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [ceph-mon : Set_fact ceph-mon container command] **************************\nTuesday 24 February 2026  00:51:01 +0000 (0:00:01.219)       0:03:46.869 ****** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-mon : Set_fact monmaptool container command] ************************\nTuesday 24 February 2026  00:51:01 +0000 (0:00:00.316)       0:03:47.185 ****** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-mon : Generate initial monmap] **************************************\nTuesday 24 February 2026  00:51:02 +0000 (0:00:00.453)       0:03:47.639 ****** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [ceph-mon : Ceph monitor mkfs with keyring] *******************************\nTuesday 24 February 2026  00:51:03 +0000 (0:00:01.484)       0:03:49.123 ****** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [ceph-mon : Ceph monitor mkfs without keyring] ****************************\nTuesday 24 February 2026  00:51:05 +0000 (0:00:01.367)       0:03:50.491 ****** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-mon : Include start_monitor.yml] ************************************\nTuesday 24 February 2026  00:51:05 +0000 (0:00:00.365)       0:03:50.857 ****** \n\u001b[0;36mincluded: /ansible/roles/ceph-mon/tasks/start_monitor.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [ceph-mon : Ensure systemd service override directory exists] *************\nTuesday 24 February 2026  00:51:06 +0000 (0:00:00.613)       0:03:51.471 ****** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-mon : Add ceph-mon systemd service overrides] ***********************\nTuesday 24 February 2026  00:51:06 +0000 (0:00:00.264)       0:03:51.735 ****** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-mon : Include_tasks systemd.yml] ************************************\nTuesday 24 February 2026  00:51:06 +0000 (0:00:00.276)       0:03:52.012 ****** \n\u001b[0;36mincluded: /ansible/roles/ceph-mon/tasks/systemd.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [ceph-mon : Generate systemd unit file for mon container] *****************\nTuesday 24 February 2026  00:51:07 +0000 (0:00:00.822)       0:03:52.834 ****** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [ceph-mon : Generate systemd ceph-mon target file] ************************\nTuesday 24 February 2026  00:51:09 +0000 (0:00:02.202)       0:03:55.036 ****** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\nTASK [ceph-mon : Enable ceph-mon.target] ***************************************\nTuesday 24 February 2026  00:51:11 +0000 (0:00:01.387)       0:03:56.424 ****** \n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [ceph-mon : Start the monitor service] ************************************\nTuesday 24 February 2026  00:51:13 +0000 (0:00:02.089)       0:03:58.514 ****** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [ceph-mon : Include_tasks ceph_keys.yml] **********************************\nTuesday 24 February 2026  00:51:15 +0000 (0:00:02.509)       0:04:01.023 ****** \n\u001b[0;36mincluded: /ansible/roles/ceph-mon/tasks/ceph_keys.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [ceph-mon : Waiting for the monitor(s) to form the quorum...] *************\nTuesday 24 February 2026  00:51:16 +0000 (0:00:00.483)       0:04:01.507 ****** \n\u001b[1;30mFAILED - RETRYING: [testbed-node-0]: Waiting for the monitor(s) to form the quorum... (10 retries left).\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [ceph-mon : Fetch ceph initial keys] **************************************\nTuesday 24 February 2026  00:51:38 +0000 (0:00:21.909)       0:04:23.416 ****** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\nTASK [ceph-mon : Include secure_cluster.yml] ***********************************\nTuesday 24 February 2026  00:51:47 +0000 (0:00:09.272)       0:04:32.688 ****** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-mon : Set cluster configs] ******************************************\nTuesday 24 February 2026  00:51:47 +0000 (0:00:00.463)       0:04:33.151 ****** \n\u001b[0;33mchanged: [testbed-node-0] => (item=[{'key': 'global', 'value': {'public_network': '192.168.16.0/20', 'cluster_network': '192.168.16.0/20', 'osd_pool_default_crush_rule': -1, 'ms_bind_ipv6': 'False', 'ms_bind_ipv4': 'True', 'osd_crush_chooseleaf_type': '__omit_place_holder__a1d5f0c2cf158f4a69bae0c92e95cd3badd9c9e3'}}, {'key': 'public_network', 'value': '192.168.16.0/20'}])\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=[{'key': 'global', 'value': {'public_network': '192.168.16.0/20', 'cluster_network': '192.168.16.0/20', 'osd_pool_default_crush_rule': -1, 'ms_bind_ipv6': 'False', 'ms_bind_ipv4': 'True', 'osd_crush_chooseleaf_type': '__omit_place_holder__a1d5f0c2cf158f4a69bae0c92e95cd3badd9c9e3'}}, {'key': 'cluster_network', 'value': '192.168.16.0/20'}])\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=[{'key': 'global', 'value': {'public_network': '192.168.16.0/20', 'cluster_network': '192.168.16.0/20', 'osd_pool_default_crush_rule': -1, 'ms_bind_ipv6': 'False', 'ms_bind_ipv4': 'True', 'osd_crush_chooseleaf_type': '__omit_place_holder__a1d5f0c2cf158f4a69bae0c92e95cd3badd9c9e3'}}, {'key': 'osd_pool_default_crush_rule', 'value': -1}])\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=[{'key': 'global', 'value': {'public_network': '192.168.16.0/20', 'cluster_network': '192.168.16.0/20', 'osd_pool_default_crush_rule': -1, 'ms_bind_ipv6': 'False', 'ms_bind_ipv4': 'True', 'osd_crush_chooseleaf_type': '__omit_place_holder__a1d5f0c2cf158f4a69bae0c92e95cd3badd9c9e3'}}, {'key': 'ms_bind_ipv6', 'value': 'False'}])\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=[{'key': 'global', 'value': {'public_network': '192.168.16.0/20', 'cluster_network': '192.168.16.0/20', 'osd_pool_default_crush_rule': -1, 'ms_bind_ipv6': 'False', 'ms_bind_ipv4': 'True', 'osd_crush_chooseleaf_type': '__omit_place_holder__a1d5f0c2cf158f4a69bae0c92e95cd3badd9c9e3'}}, {'key': 'ms_bind_ipv4', 'value': 'True'}])\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=[{'key': 'global', 'value': {'public_network': '192.168.16.0/20', 'cluster_network': '192.168.16.0/20', 'osd_pool_default_crush_rule': -1, 'ms_bind_ipv6': 'False', 'ms_bind_ipv4': 'True', 'osd_crush_chooseleaf_type': '__omit_place_holder__a1d5f0c2cf158f4a69bae0c92e95cd3badd9c9e3'}}, {'key': 'osd_crush_chooseleaf_type', 'value': '__omit_place_holder__a1d5f0c2cf158f4a69bae0c92e95cd3badd9c9e3'}]) \u001b[0m\n\nRUNNING HANDLER [ceph-handler : Make tempdir for scripts] **********************\nTuesday 24 February 2026  00:52:02 +0000 (0:00:14.800)       0:04:47.952 ****** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Mons handler] **********************************\nTuesday 24 February 2026  00:52:02 +0000 (0:00:00.395)       0:04:48.348 ****** \n\u001b[0;36mincluded: /ansible/roles/ceph-handler/tasks/handler_mons.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Set _mon_handler_called before restart] ********\nTuesday 24 February 2026  00:52:03 +0000 (0:00:00.868)       0:04:49.217 ****** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Copy mon restart script] ***********************\nTuesday 24 February 2026  00:52:04 +0000 (0:00:00.472)       0:04:49.689 ****** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Restart ceph mon daemon(s)] ********************\nTuesday 24 February 2026  00:52:04 +0000 (0:00:00.489)       0:04:50.179 ****** \n\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-0) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-1) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Set _mon_handler_called after restart] *********\nTuesday 24 February 2026  00:52:06 +0000 (0:00:01.477)       0:04:51.656 ****** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nPLAY [Apply role ceph-mgr] *****************************************************\n\nTASK [ceph-handler : Include check_running_cluster.yml] ************************\nTuesday 24 February 2026  00:52:07 +0000 (0:00:00.780)       0:04:52.436 ****** \n\u001b[0;36mincluded: /ansible/roles/ceph-handler/tasks/check_running_cluster.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [ceph-handler : Include check_running_containers.yml] *********************\nTuesday 24 February 2026  00:52:07 +0000 (0:00:00.577)       0:04:53.013 ****** \n\u001b[0;36mincluded: /ansible/roles/ceph-handler/tasks/check_running_containers.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [ceph-handler : Check for a mon container] ********************************\nTuesday 24 February 2026  00:52:08 +0000 (0:00:00.579)       0:04:53.593 ****** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Check for an osd container] *******************************\nTuesday 24 February 2026  00:52:09 +0000 (0:00:00.773)       0:04:54.366 ****** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Check for a mds container] ********************************\nTuesday 24 February 2026  00:52:09 +0000 (0:00:00.387)       0:04:54.753 ****** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Check for a rgw container] ********************************\nTuesday 24 February 2026  00:52:09 +0000 (0:00:00.376)       0:04:55.130 ****** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Check for a mgr container] ********************************\nTuesday 24 February 2026  00:52:10 +0000 (0:00:00.303)       0:04:55.433 ****** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Check for a rbd mirror container] *************************\nTuesday 24 February 2026  00:52:10 +0000 (0:00:00.843)       0:04:56.277 ****** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Check for a nfs container] ********************************\nTuesday 24 February 2026  00:52:11 +0000 (0:00:00.313)       0:04:56.590 ****** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [ceph-handler : Check for a ceph-crash container] *************************\nTuesday 24 February 2026  00:52:11 +0000 (0:00:00.679)       0:04:57.269 ****** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Check for a ceph-exporter container] **********************\nTuesday 24 February 2026  00:52:12 +0000 (0:00:01.017)       0:04:58.287 ****** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\nTASK [ceph-handler : Include check_socket_non_container.yml] *******************\nTuesday 24 February 2026  00:52:13 +0000 (0:00:00.785)       0:04:59.073 ****** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_mon_status] ******************************\nTuesday 24 February 2026  00:52:13 +0000 (0:00:00.265)       0:04:59.338 ****** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_osd_status] ******************************\nTuesday 24 February 2026  00:52:14 +0000 (0:00:00.277)       0:04:59.616 ****** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_mds_status] ******************************\nTuesday 24 February 2026  00:52:14 +0000 (0:00:00.426)       0:05:00.042 ****** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_rgw_status] ******************************\nTuesday 24 February 2026  00:52:14 +0000 (0:00:00.256)       0:05:00.299 ****** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_nfs_status] ******************************\nTuesday 24 February 2026  00:52:15 +0000 (0:00:00.267)       0:05:00.567 ****** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_rbd_status] ******************************\nTuesday 24 February 2026  00:52:15 +0000 (0:00:00.300)       0:05:00.868 ****** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_mgr_status] ******************************\nTuesday 24 February 2026  00:52:15 +0000 (0:00:00.455)       0:05:01.323 ****** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_crash_status] ****************************\nTuesday 24 February 2026  00:52:16 +0000 (0:00:00.318)       0:05:01.641 ****** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_exporter_status] *************************\nTuesday 24 February 2026  00:52:16 +0000 (0:00:00.275)       0:05:01.916 ****** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-mgr : Set_fact container_exec_cmd] **********************************\nTuesday 24 February 2026  00:52:17 +0000 (0:00:00.720)       0:05:02.637 ****** \n\u001b[0;32mok: [testbed-node-0] => (item=testbed-node-0)\u001b[0m\n\u001b[0;32mok: [testbed-node-0 -> testbed-node-1(192.168.16.11)] => (item=testbed-node-1)\u001b[0m\n\u001b[0;32mok: [testbed-node-0 -> testbed-node-2(192.168.16.12)] => (item=testbed-node-2)\u001b[0m\n\nTASK [ceph-mgr : Include common.yml] *******************************************\nTuesday 24 February 2026  00:52:17 +0000 (0:00:00.474)       0:05:03.112 ****** \n\u001b[0;36mincluded: /ansible/roles/ceph-mgr/tasks/common.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [ceph-mgr : Create mgr directory] *****************************************\nTuesday 24 February 2026  00:52:18 +0000 (0:00:00.515)       0:05:03.627 ****** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [ceph-mgr : Fetch ceph mgr keyring] ***************************************\nTuesday 24 February 2026  00:52:18 +0000 (0:00:00.605)       0:05:04.233 ****** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-mgr : Create ceph mgr keyring(s) on a mon node] *********************\nTuesday 24 February 2026  00:52:19 +0000 (0:00:00.430)       0:05:04.663 ****** \n\u001b[0;33mchanged: [testbed-node-0] => (item=None)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=None)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=None)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0 -> {{ groups[mon_group_name][0] }}]\u001b[0m\n\nTASK [ceph-mgr : Set_fact _mgr_keys] *******************************************\nTuesday 24 February 2026  00:52:29 +0000 (0:00:10.213)       0:05:14.877 ****** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-mgr : Get keys from monitors] ***************************************\nTuesday 24 February 2026  00:52:29 +0000 (0:00:00.371)       0:05:15.248 ****** \n\u001b[0;36mskipping: [testbed-node-0] => (item=None) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=None) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=None) \u001b[0m\n\u001b[0;32mok: [testbed-node-0] => (item=None)\u001b[0m\n\u001b[0;32mok: [testbed-node-1 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m\n\u001b[0;32mok: [testbed-node-2 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m\n\nTASK [ceph-mgr : Copy ceph key(s) if needed] ***********************************\nTuesday 24 February 2026  00:52:32 +0000 (0:00:02.351)       0:05:17.600 ****** \n\u001b[0;36mskipping: [testbed-node-0] => (item=None) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=None) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=None) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=None)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=None)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=None)\u001b[0m\n\nTASK [ceph-mgr : Set mgr key permissions] **************************************\nTuesday 24 February 2026  00:52:33 +0000 (0:00:01.488)       0:05:19.089 ****** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-mgr : Append dashboard modules to ceph_mgr_modules] *****************\nTuesday 24 February 2026  00:52:35 +0000 (0:00:01.523)       0:05:20.613 ****** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-mgr : Include pre_requisite.yml] ************************************\nTuesday 24 February 2026  00:52:35 +0000 (0:00:00.332)       0:05:20.945 ****** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-mgr : Include start_mgr.yml] ****************************************\nTuesday 24 February 2026  00:52:35 +0000 (0:00:00.298)       0:05:21.244 ****** \n\u001b[0;36mincluded: /ansible/roles/ceph-mgr/tasks/start_mgr.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [ceph-mgr : Ensure systemd service override directory exists] *************\nTuesday 24 February 2026  00:52:36 +0000 (0:00:00.723)       0:05:21.968 ****** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-mgr : Add ceph-mgr systemd service overrides] ***********************\nTuesday 24 February 2026  00:52:36 +0000 (0:00:00.324)       0:05:22.292 ****** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-mgr : Include_tasks systemd.yml] ************************************\nTuesday 24 February 2026  00:52:37 +0000 (0:00:00.329)       0:05:22.621 ****** \n\u001b[0;36mincluded: /ansible/roles/ceph-mgr/tasks/systemd.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [ceph-mgr : Generate systemd unit file] ***********************************\nTuesday 24 February 2026  00:52:38 +0000 (0:00:00.805)       0:05:23.427 ****** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\nTASK [ceph-mgr : Generate systemd ceph-mgr target file] ************************\nTuesday 24 February 2026  00:52:39 +0000 (0:00:01.273)       0:05:24.701 ****** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [ceph-mgr : Enable ceph-mgr.target] ***************************************\nTuesday 24 February 2026  00:52:40 +0000 (0:00:01.225)       0:05:25.927 ****** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [ceph-mgr : Systemd start mgr] ********************************************\nTuesday 24 February 2026  00:52:42 +0000 (0:00:02.019)       0:05:27.946 ****** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [ceph-mgr : Include mgr_modules.yml] **************************************\nTuesday 24 February 2026  00:52:44 +0000 (0:00:02.114)       0:05:30.060 ****** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mincluded: /ansible/roles/ceph-mgr/tasks/mgr_modules.yml for testbed-node-2\u001b[0m\n\nTASK [ceph-mgr : Wait for all mgr to be up] ************************************\nTuesday 24 February 2026  00:52:45 +0000 (0:00:00.542)       0:05:30.603 ****** \n\u001b[1;30mFAILED - RETRYING: [testbed-node-2 -> testbed-node-0]: Wait for all mgr to be up (30 retries left).\u001b[0m\n\u001b[1;30mFAILED - RETRYING: [testbed-node-2 -> testbed-node-0]: Wait for all mgr to be up (29 retries left).\u001b[0m\n\u001b[1;30mFAILED - RETRYING: [testbed-node-2 -> testbed-node-0]: Wait for all mgr to be up (28 retries left).\u001b[0m\n\u001b[1;30mFAILED - RETRYING: [testbed-node-2 -> testbed-node-0]: Wait for all mgr to be up (27 retries left).\u001b[0m\n\u001b[1;30mFAILED - RETRYING: [testbed-node-2 -> testbed-node-0]: Wait for all mgr to be up (26 retries left).\u001b[0m\n\u001b[1;30mFAILED - RETRYING: [testbed-node-2 -> testbed-node-0]: Wait for all mgr to be up (25 retries left).\u001b[0m\n\u001b[0;32mok: [testbed-node-2 -> testbed-node-0(192.168.16.10)]\u001b[0m\n\nTASK [ceph-mgr : Get enabled modules from ceph-mgr] ****************************\nTuesday 24 February 2026  00:53:21 +0000 (0:00:36.565)       0:06:07.169 ****** \n\u001b[0;32mok: [testbed-node-2 -> testbed-node-0(192.168.16.10)]\u001b[0m\n\nTASK [ceph-mgr : Set _ceph_mgr_modules fact (convert _ceph_mgr_modules.stdout to a dict)] ***\nTuesday 24 February 2026  00:53:23 +0000 (0:00:01.367)       0:06:08.536 ****** \n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-mgr : Set _disabled_ceph_mgr_modules fact] **************************\nTuesday 24 February 2026  00:53:23 +0000 (0:00:00.277)       0:06:08.813 ****** \n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-mgr : Disable ceph mgr enabled modules] *****************************\nTuesday 24 February 2026  00:53:23 +0000 (0:00:00.140)       0:06:08.954 ****** \n\u001b[0;33mchanged: [testbed-node-2 -> testbed-node-0(192.168.16.10)] => (item=iostat)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2 -> testbed-node-0(192.168.16.10)] => (item=nfs)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2 -> testbed-node-0(192.168.16.10)] => (item=restful)\u001b[0m\n\nTASK [ceph-mgr : Add modules to ceph-mgr] **************************************\nTuesday 24 February 2026  00:53:30 +0000 (0:00:06.715)       0:06:15.670 ****** \n\u001b[0;36mskipping: [testbed-node-2] => (item=balancer) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-2 -> testbed-node-0(192.168.16.10)] => (item=dashboard)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2 -> testbed-node-0(192.168.16.10)] => (item=prometheus)\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=status) \u001b[0m\n\nRUNNING HANDLER [ceph-handler : Make tempdir for scripts] **********************\nTuesday 24 February 2026  00:53:36 +0000 (0:00:05.722)       0:06:21.392 ****** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Mgrs handler] **********************************\nTuesday 24 February 2026  00:53:36 +0000 (0:00:00.724)       0:06:22.116 ****** \n\u001b[0;36mincluded: /ansible/roles/ceph-handler/tasks/handler_mgrs.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Set _mgr_handler_called before restart] ********\nTuesday 24 February 2026  00:53:37 +0000 (0:00:00.907)       0:06:23.024 ****** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Copy mgr restart script] ***********************\nTuesday 24 February 2026  00:53:38 +0000 (0:00:00.373)       0:06:23.398 ****** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Restart ceph mgr daemon(s)] ********************\nTuesday 24 February 2026  00:53:39 +0000 (0:00:01.324)       0:06:24.723 ****** \n\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-0) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-1) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Set _mgr_handler_called after restart] *********\nTuesday 24 February 2026  00:53:39 +0000 (0:00:00.618)       0:06:25.341 ****** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nPLAY [Apply role ceph-osd] *****************************************************\n\nTASK [ceph-handler : Include check_running_cluster.yml] ************************\nTuesday 24 February 2026  00:53:40 +0000 (0:00:00.952)       0:06:26.293 ****** \n\u001b[0;36mincluded: /ansible/roles/ceph-handler/tasks/check_running_cluster.yml for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m\n\nTASK [ceph-handler : Include check_running_containers.yml] *********************\nTuesday 24 February 2026  00:53:41 +0000 (0:00:00.506)       0:06:26.800 ****** \n\u001b[0;36mincluded: /ansible/roles/ceph-handler/tasks/check_running_containers.yml for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m\n\nTASK [ceph-handler : Check for a mon container] ********************************\nTuesday 24 February 2026  00:53:42 +0000 (0:00:00.764)       0:06:27.565 ****** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Check for an osd container] *******************************\nTuesday 24 February 2026  00:53:42 +0000 (0:00:00.314)       0:06:27.879 ****** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Check for a mds container] ********************************\nTuesday 24 February 2026  00:53:43 +0000 (0:00:00.810)       0:06:28.690 ****** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Check for a rgw container] ********************************\nTuesday 24 February 2026  00:53:44 +0000 (0:00:00.843)       0:06:29.533 ****** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Check for a mgr container] ********************************\nTuesday 24 February 2026  00:53:45 +0000 (0:00:01.264)       0:06:30.797 ****** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Check for a rbd mirror container] *************************\nTuesday 24 February 2026  00:53:45 +0000 (0:00:00.368)       0:06:31.166 ****** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Check for a nfs container] ********************************\nTuesday 24 February 2026  00:53:46 +0000 (0:00:00.307)       0:06:31.473 ****** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Check for a ceph-crash container] *************************\nTuesday 24 February 2026  00:53:46 +0000 (0:00:00.323)       0:06:31.797 ****** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Check for a ceph-exporter container] **********************\nTuesday 24 February 2026  00:53:47 +0000 (0:00:01.020)       0:06:32.818 ****** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Include check_socket_non_container.yml] *******************\nTuesday 24 February 2026  00:53:48 +0000 (0:00:00.694)       0:06:33.512 ****** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_mon_status] ******************************\nTuesday 24 February 2026  00:53:48 +0000 (0:00:00.339)       0:06:33.852 ****** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_osd_status] ******************************\nTuesday 24 February 2026  00:53:48 +0000 (0:00:00.338)       0:06:34.190 ****** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_mds_status] ******************************\nTuesday 24 February 2026  00:53:49 +0000 (0:00:00.629)       0:06:34.819 ****** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_rgw_status] ******************************\nTuesday 24 February 2026  00:53:49 +0000 (0:00:00.366)       0:06:35.186 ****** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_nfs_status] ******************************\nTuesday 24 February 2026  00:53:50 +0000 (0:00:00.346)       0:06:35.533 ****** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_rbd_status] ******************************\nTuesday 24 February 2026  00:53:50 +0000 (0:00:00.309)       0:06:35.842 ****** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_mgr_status] ******************************\nTuesday 24 February 2026  00:53:51 +0000 (0:00:00.702)       0:06:36.544 ****** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_crash_status] ****************************\nTuesday 24 February 2026  00:53:51 +0000 (0:00:00.384)       0:06:36.928 ****** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_exporter_status] *************************\nTuesday 24 February 2026  00:53:51 +0000 (0:00:00.370)       0:06:37.299 ****** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-osd : Set_fact add_osd] *********************************************\nTuesday 24 February 2026  00:53:52 +0000 (0:00:00.554)       0:06:37.853 ****** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-osd : Set_fact container_exec_cmd] **********************************\nTuesday 24 February 2026  00:53:53 +0000 (0:00:00.625)       0:06:38.478 ****** \n\u001b[0;32mok: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item=testbed-node-0)\u001b[0m\n\u001b[0;32mok: [testbed-node-3 -> testbed-node-1(192.168.16.11)] => (item=testbed-node-1)\u001b[0m\n\u001b[0;32mok: [testbed-node-3 -> testbed-node-2(192.168.16.12)] => (item=testbed-node-2)\u001b[0m\n\nTASK [ceph-osd : Include_tasks system_tuning.yml] ******************************\nTuesday 24 February 2026  00:53:53 +0000 (0:00:00.643)       0:06:39.122 ****** \n\u001b[0;36mincluded: /ansible/roles/ceph-osd/tasks/system_tuning.yml for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m\n\nTASK [ceph-osd : Create tmpfiles.d directory] **********************************\nTuesday 24 February 2026  00:53:54 +0000 (0:00:00.654)       0:06:39.777 ****** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-osd : Disable transparent hugepage] *********************************\nTuesday 24 February 2026  00:53:54 +0000 (0:00:00.584)       0:06:40.361 ****** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-osd : Get default vm.min_free_kbytes] *******************************\nTuesday 24 February 2026  00:53:55 +0000 (0:00:00.355)       0:06:40.716 ****** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-osd : Set_fact vm_min_free_kbytes] **********************************\nTuesday 24 February 2026  00:53:55 +0000 (0:00:00.617)       0:06:41.334 ****** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-osd : Apply operating system tuning] ********************************\nTuesday 24 February 2026  00:53:56 +0000 (0:00:00.394)       0:06:41.728 ****** \n\u001b[0;33mchanged: [testbed-node-3] => (item={'name': 'fs.aio-max-nr', 'value': '1048576', 'enable': True})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'name': 'fs.aio-max-nr', 'value': '1048576', 'enable': True})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'name': 'fs.file-max', 'value': 26234859})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'name': 'fs.file-max', 'value': 26234859})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'name': 'vm.zone_reclaim_mode', 'value': 0})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'name': 'fs.aio-max-nr', 'value': '1048576', 'enable': True})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'name': 'vm.zone_reclaim_mode', 'value': 0})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'name': 'vm.swappiness', 'value': 10})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'name': 'vm.swappiness', 'value': 10})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'name': 'vm.min_free_kbytes', 'value': '67584'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'name': 'fs.file-max', 'value': 26234859})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'name': 'vm.zone_reclaim_mode', 'value': 0})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'name': 'vm.swappiness', 'value': 10})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'name': 'vm.min_free_kbytes', 'value': '67584'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'name': 'vm.min_free_kbytes', 'value': '67584'})\u001b[0m\n\nTASK [ceph-osd : Install dependencies] *****************************************\nTuesday 24 February 2026  00:54:00 +0000 (0:00:04.284)       0:06:46.013 ****** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-osd : Include_tasks common.yml] *************************************\nTuesday 24 February 2026  00:54:00 +0000 (0:00:00.308)       0:06:46.321 ****** \n\u001b[0;36mincluded: /ansible/roles/ceph-osd/tasks/common.yml for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m\n\nTASK [ceph-osd : Create bootstrap-osd and osd directories] *********************\nTuesday 24 February 2026  00:54:01 +0000 (0:00:00.544)       0:06:46.866 ****** \n\u001b[0;32mok: [testbed-node-3] => (item=/var/lib/ceph/bootstrap-osd/)\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item=/var/lib/ceph/bootstrap-osd/)\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item=/var/lib/ceph/bootstrap-osd/)\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item=/var/lib/ceph/osd/)\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item=/var/lib/ceph/osd/)\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item=/var/lib/ceph/osd/)\u001b[0m\n\nTASK [ceph-osd : Get keys from monitors] ***************************************\nTuesday 24 February 2026  00:54:02 +0000 (0:00:01.375)       0:06:48.241 ****** \n\u001b[0;32mok: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=None) \u001b[0m\n\u001b[0;32mok: [testbed-node-3 -> {{ groups.get(mon_group_name)[0] }}]\u001b[0m\n\nTASK [ceph-osd : Copy ceph key(s) if needed] ***********************************\nTuesday 24 February 2026  00:54:05 +0000 (0:00:02.637)       0:06:50.878 ****** \n\u001b[0;33mchanged: [testbed-node-3] => (item=None)\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=None) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=None)\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=None) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=None)\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=None) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nTASK [ceph-osd : Set noup flag] ************************************************\nTuesday 24 February 2026  00:54:06 +0000 (0:00:01.342)       0:06:52.221 ****** \n\u001b[0;33mchanged: [testbed-node-3 -> testbed-node-0(192.168.16.10)]\u001b[0m\n\nTASK [ceph-osd : Include_tasks scenarios/lvm.yml] ******************************\nTuesday 24 February 2026  00:54:09 +0000 (0:00:02.618)       0:06:54.839 ****** \n\u001b[0;36mincluded: /ansible/roles/ceph-osd/tasks/scenarios/lvm.yml for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m\n\nTASK [ceph-osd : Use ceph-volume to create osds] *******************************\nTuesday 24 February 2026  00:54:10 +0000 (0:00:00.555)       0:06:55.395 ****** \n\u001b[0;33mchanged: [testbed-node-5] => (item={'data': 'osd-block-752a9ec4-4bfc-5551-82cd-92b802bdad97', 'data_vg': 'ceph-752a9ec4-4bfc-5551-82cd-92b802bdad97'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'data': 'osd-block-9b92f24c-dac1-5bff-a390-34d9ff9e0f4f', 'data_vg': 'ceph-9b92f24c-dac1-5bff-a390-34d9ff9e0f4f'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'data': 'osd-block-294dd48f-9d6d-5079-8214-59e60af9e526', 'data_vg': 'ceph-294dd48f-9d6d-5079-8214-59e60af9e526'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'data': 'osd-block-ba85a0e3-2f57-51d5-a9f1-072f0e9a8deb', 'data_vg': 'ceph-ba85a0e3-2f57-51d5-a9f1-072f0e9a8deb'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'data': 'osd-block-7f6a8fe5-6adc-52b4-b256-edcdbfd9bf4e', 'data_vg': 'ceph-7f6a8fe5-6adc-52b4-b256-edcdbfd9bf4e'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'data': 'osd-block-7fbf4d63-7c6f-5c70-887c-2b73c978775b', 'data_vg': 'ceph-7fbf4d63-7c6f-5c70-887c-2b73c978775b'})\u001b[0m\n\nTASK [ceph-osd : Include_tasks scenarios/lvm-batch.yml] ************************\nTuesday 24 February 2026  00:54:58 +0000 (0:00:48.430)       0:07:43.825 ****** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-osd : Include_tasks start_osds.yml] *********************************\nTuesday 24 February 2026  00:54:58 +0000 (0:00:00.408)       0:07:44.234 ****** \n\u001b[0;36mincluded: /ansible/roles/ceph-osd/tasks/start_osds.yml for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m\n\nTASK [ceph-osd : Get osd ids] **************************************************\nTuesday 24 February 2026  00:54:59 +0000 (0:00:00.681)       0:07:44.915 ****** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-osd : Collect osd ids] **********************************************\nTuesday 24 February 2026  00:55:00 +0000 (0:00:01.064)       0:07:45.979 ****** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\nTASK [ceph-osd : Include_tasks systemd.yml] ************************************\nTuesday 24 February 2026  00:55:03 +0000 (0:00:02.510)       0:07:48.490 ****** \n\u001b[0;36mincluded: /ansible/roles/ceph-osd/tasks/systemd.yml for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m\n\nTASK [ceph-osd : Generate systemd unit file] ***********************************\nTuesday 24 February 2026  00:55:03 +0000 (0:00:00.502)       0:07:48.992 ****** \n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nTASK [ceph-osd : Generate systemd ceph-osd target file] ************************\nTuesday 24 February 2026  00:55:05 +0000 (0:00:01.478)       0:07:50.471 ****** \n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nTASK [ceph-osd : Enable ceph-osd.target] ***************************************\nTuesday 24 February 2026  00:55:06 +0000 (0:00:01.145)       0:07:51.617 ****** \n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\nTASK [ceph-osd : Ensure systemd service override directory exists] *************\nTuesday 24 February 2026  00:55:08 +0000 (0:00:01.909)       0:07:53.526 ****** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-osd : Add ceph-osd systemd service overrides] ***********************\nTuesday 24 February 2026  00:55:08 +0000 (0:00:00.359)       0:07:53.886 ****** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-osd : Ensure /var/lib/ceph/osd/<cluster>-<item> is present] *********\nTuesday 24 February 2026  00:55:09 +0000 (0:00:00.618)       0:07:54.505 ****** \n\u001b[0;32mok: [testbed-node-3] => (item=0)\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item=5)\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item=3)\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item=4)\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item=1)\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item=2)\u001b[0m\n\nTASK [ceph-osd : Write run file in /var/lib/ceph/osd/xxxx/run] *****************\nTuesday 24 February 2026  00:55:10 +0000 (0:00:01.020)       0:07:55.525 ****** \n\u001b[0;33mchanged: [testbed-node-3] => (item=0)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=3)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=5)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item=4)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=1)\u001b[0m\n\nTASK [ceph-osd : Systemd start osd] ********************************************\nTuesday 24 February 2026  00:55:12 +0000 (0:00:02.525)       0:07:58.051 ****** \n\u001b[0;33mchanged: [testbed-node-5] => (item=3)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item=0)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=5)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item=4)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=1)\u001b[0m\n\nTASK [ceph-osd : Unset noup flag] **********************************************\nTuesday 24 February 2026  00:55:16 +0000 (0:00:03.882)       0:08:01.933 ****** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-0(192.168.16.10)]\u001b[0m\n\nTASK [ceph-osd : Wait for all osd to be up] ************************************\nTuesday 24 February 2026  00:55:19 +0000 (0:00:03.041)       0:08:04.975 ****** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[1;30mFAILED - RETRYING: [testbed-node-5 -> testbed-node-0]: Wait for all osd to be up (60 retries left).\u001b[0m\n\u001b[0;32mok: [testbed-node-5 -> testbed-node-0(192.168.16.10)]\u001b[0m\n\nTASK [ceph-osd : Include crush_rules.yml] **************************************\nTuesday 24 February 2026  00:55:32 +0000 (0:00:12.824)       0:08:17.799 ****** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Make tempdir for scripts] **********************\nTuesday 24 February 2026  00:55:33 +0000 (0:00:01.281)       0:08:19.081 ****** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Osds handler] **********************************\nTuesday 24 February 2026  00:55:34 +0000 (0:00:00.400)       0:08:19.482 ****** \n\u001b[0;36mincluded: /ansible/roles/ceph-handler/tasks/handler_osds.yml for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Set_fact trigger_restart] **********************\nTuesday 24 February 2026  00:55:34 +0000 (0:00:00.543)       0:08:20.026 ****** \n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-3) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-4) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-5) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Set _osd_handler_called before restart] ********\nTuesday 24 February 2026  00:55:35 +0000 (0:00:01.062)       0:08:21.088 ****** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Unset noup flag] *******************************\nTuesday 24 February 2026  00:55:36 +0000 (0:00:00.332)       0:08:21.421 ****** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Copy osd restart script] ***********************\nTuesday 24 February 2026  00:55:36 +0000 (0:00:00.232)       0:08:21.654 ****** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Get pool list] *********************************\nTuesday 24 February 2026  00:55:36 +0000 (0:00:00.388)       0:08:22.042 ****** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Get balancer module status] ********************\nTuesday 24 February 2026  00:55:36 +0000 (0:00:00.231)       0:08:22.273 ****** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Set_fact pools_pgautoscaler_mode] **************\nTuesday 24 February 2026  00:55:37 +0000 (0:00:00.215)       0:08:22.489 ****** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Disable balancer] ******************************\nTuesday 24 February 2026  00:55:37 +0000 (0:00:00.122)       0:08:22.611 ****** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Disable pg autoscale on pools] *****************\nTuesday 24 February 2026  00:55:37 +0000 (0:00:00.371)       0:08:22.983 ****** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Restart ceph osds daemon(s)] *******************\nTuesday 24 February 2026  00:55:38 +0000 (0:00:01.051)       0:08:24.035 ****** \n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-3) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-4) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-5) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Set _osd_handler_called after restart] *********\nTuesday 24 February 2026  00:55:39 +0000 (0:00:00.466)       0:08:24.501 ****** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Re-enable pg autoscale on pools] ***************\nTuesday 24 February 2026  00:55:39 +0000 (0:00:00.355)       0:08:24.857 ****** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Re-enable balancer] ****************************\nTuesday 24 February 2026  00:55:39 +0000 (0:00:00.261)       0:08:25.118 ****** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nPLAY [Apply role ceph-crash] ***************************************************\n\nTASK [ceph-handler : Include check_running_cluster.yml] ************************\nTuesday 24 February 2026  00:55:40 +0000 (0:00:00.733)       0:08:25.852 ****** \n\u001b[0;36mincluded: /ansible/roles/ceph-handler/tasks/check_running_cluster.yml for testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [ceph-handler : Include check_running_containers.yml] *********************\nTuesday 24 February 2026  00:55:41 +0000 (0:00:01.317)       0:08:27.169 ****** \n\u001b[0;36mincluded: /ansible/roles/ceph-handler/tasks/check_running_containers.yml for testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [ceph-handler : Check for a mon container] ********************************\nTuesday 24 February 2026  00:55:43 +0000 (0:00:01.401)       0:08:28.571 ****** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Check for an osd container] *******************************\nTuesday 24 February 2026  00:55:44 +0000 (0:00:01.389)       0:08:29.961 ****** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Check for a mds container] ********************************\nTuesday 24 February 2026  00:55:45 +0000 (0:00:00.719)       0:08:30.681 ****** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Check for a rgw container] ********************************\nTuesday 24 February 2026  00:55:46 +0000 (0:00:01.022)       0:08:31.703 ****** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Check for a mgr container] ********************************\nTuesday 24 February 2026  00:55:47 +0000 (0:00:00.710)       0:08:32.413 ****** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Check for a rbd mirror container] *************************\nTuesday 24 February 2026  00:55:48 +0000 (0:00:01.334)       0:08:33.748 ****** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Check for a nfs container] ********************************\nTuesday 24 February 2026  00:55:49 +0000 (0:00:00.717)       0:08:34.466 ****** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Check for a ceph-crash container] *************************\nTuesday 24 February 2026  00:55:50 +0000 (0:00:00.931)       0:08:35.398 ****** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Check for a ceph-exporter container] **********************\nTuesday 24 February 2026  00:55:51 +0000 (0:00:01.101)       0:08:36.499 ****** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Include check_socket_non_container.yml] *******************\nTuesday 24 February 2026  00:55:52 +0000 (0:00:01.492)       0:08:37.991 ****** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_mon_status] ******************************\nTuesday 24 February 2026  00:55:53 +0000 (0:00:00.663)       0:08:38.655 ****** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_osd_status] ******************************\nTuesday 24 February 2026  00:55:54 +0000 (0:00:01.115)       0:08:39.771 ****** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_mds_status] ******************************\nTuesday 24 February 2026  00:55:55 +0000 (0:00:00.796)       0:08:40.567 ****** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_rgw_status] ******************************\nTuesday 24 February 2026  00:55:56 +0000 (0:00:01.059)       0:08:41.627 ****** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_nfs_status] ******************************\nTuesday 24 February 2026  00:55:57 +0000 (0:00:00.835)       0:08:42.462 ****** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_rbd_status] ******************************\nTuesday 24 February 2026  00:55:58 +0000 (0:00:00.942)       0:08:43.405 ****** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_mgr_status] ******************************\nTuesday 24 February 2026  00:55:58 +0000 (0:00:00.771)       0:08:44.177 ****** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_crash_status] ****************************\nTuesday 24 February 2026  00:55:59 +0000 (0:00:01.109)       0:08:45.286 ****** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_exporter_status] *************************\nTuesday 24 February 2026  00:56:00 +0000 (0:00:00.723)       0:08:46.010 ****** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [ceph-crash : Create client.crash keyring] ********************************\nTuesday 24 February 2026  00:56:02 +0000 (0:00:01.441)       0:08:47.452 ****** \n\u001b[0;33mchanged: [testbed-node-3 -> testbed-node-0(192.168.16.10)]\u001b[0m\n\nTASK [ceph-crash : Get keys from monitors] *************************************\nTuesday 24 February 2026  00:56:06 +0000 (0:00:04.655)       0:08:52.108 ****** \n\u001b[0;32mok: [testbed-node-3 -> testbed-node-0(192.168.16.10)]\u001b[0m\n\nTASK [ceph-crash : Copy ceph key(s) if needed] *********************************\nTuesday 24 February 2026  00:56:09 +0000 (0:00:02.380)       0:08:54.489 ****** \n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [ceph-crash : Create /var/lib/ceph/crash/posted] **************************\nTuesday 24 February 2026  00:56:11 +0000 (0:00:02.195)       0:08:56.684 ****** \n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [ceph-crash : Include_tasks systemd.yml] **********************************\nTuesday 24 February 2026  00:56:12 +0000 (0:00:01.034)       0:08:57.719 ****** \n\u001b[0;36mincluded: /ansible/roles/ceph-crash/tasks/systemd.yml for testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [ceph-crash : Generate systemd unit file for ceph-crash container] ********\nTuesday 24 February 2026  00:56:13 +0000 (0:00:01.392)       0:08:59.112 ****** \n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [ceph-crash : Start the ceph-crash service] *******************************\nTuesday 24 February 2026  00:56:15 +0000 (0:00:02.048)       0:09:01.160 ****** \n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Ceph crash handler] ****************************\nTuesday 24 February 2026  00:56:19 +0000 (0:00:03.697)       0:09:04.857 ****** \n\u001b[0;36mincluded: /ansible/roles/ceph-handler/tasks/handler_crash.yml for testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Set _crash_handler_called before restart] ******\nTuesday 24 February 2026  00:56:22 +0000 (0:00:02.520)       0:09:07.378 ****** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Restart the ceph-crash service] ****************\nTuesday 24 February 2026  00:56:23 +0000 (0:00:01.480)       0:09:08.859 ****** \n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Set _crash_handler_called after restart] *******\nTuesday 24 February 2026  00:56:27 +0000 (0:00:03.599)       0:09:12.458 ****** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nPLAY [Apply role ceph-mds] *****************************************************\n\nTASK [ceph-handler : Include check_running_cluster.yml] ************************\nTuesday 24 February 2026  00:56:28 +0000 (0:00:01.310)       0:09:13.769 ****** \n\u001b[0;36mincluded: /ansible/roles/ceph-handler/tasks/check_running_cluster.yml for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m\n\nTASK [ceph-handler : Include check_running_containers.yml] *********************\nTuesday 24 February 2026  00:56:29 +0000 (0:00:00.776)       0:09:14.545 ****** \n\u001b[0;36mincluded: /ansible/roles/ceph-handler/tasks/check_running_containers.yml for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m\n\nTASK [ceph-handler : Check for a mon container] ********************************\nTuesday 24 February 2026  00:56:29 +0000 (0:00:00.814)       0:09:15.360 ****** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Check for an osd container] *******************************\nTuesday 24 February 2026  00:56:30 +0000 (0:00:00.436)       0:09:15.796 ****** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Check for a mds container] ********************************\nTuesday 24 February 2026  00:56:31 +0000 (0:00:00.758)       0:09:16.554 ****** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Check for a rgw container] ********************************\nTuesday 24 February 2026  00:56:32 +0000 (0:00:01.039)       0:09:17.594 ****** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Check for a mgr container] ********************************\nTuesday 24 February 2026  00:56:32 +0000 (0:00:00.734)       0:09:18.329 ****** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Check for a rbd mirror container] *************************\nTuesday 24 February 2026  00:56:33 +0000 (0:00:00.345)       0:09:18.675 ****** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Check for a nfs container] ********************************\nTuesday 24 February 2026  00:56:33 +0000 (0:00:00.307)       0:09:18.982 ****** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Check for a ceph-crash container] *************************\nTuesday 24 February 2026  00:56:34 +0000 (0:00:00.656)       0:09:19.638 ****** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Check for a ceph-exporter container] **********************\nTuesday 24 February 2026  00:56:35 +0000 (0:00:00.763)       0:09:20.402 ****** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Include check_socket_non_container.yml] *******************\nTuesday 24 February 2026  00:56:35 +0000 (0:00:00.936)       0:09:21.338 ****** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_mon_status] ******************************\nTuesday 24 February 2026  00:56:36 +0000 (0:00:00.347)       0:09:21.685 ****** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_osd_status] ******************************\nTuesday 24 February 2026  00:56:37 +0000 (0:00:00.850)       0:09:22.536 ****** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_mds_status] ******************************\nTuesday 24 February 2026  00:56:37 +0000 (0:00:00.412)       0:09:22.949 ****** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_rgw_status] ******************************\nTuesday 24 February 2026  00:56:38 +0000 (0:00:00.480)       0:09:23.429 ****** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_nfs_status] ******************************\nTuesday 24 February 2026  00:56:38 +0000 (0:00:00.413)       0:09:23.843 ****** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_rbd_status] ******************************\nTuesday 24 February 2026  00:56:39 +0000 (0:00:00.698)       0:09:24.541 ****** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_mgr_status] ******************************\nTuesday 24 February 2026  00:56:39 +0000 (0:00:00.336)       0:09:24.877 ****** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_crash_status] ****************************\nTuesday 24 February 2026  00:56:39 +0000 (0:00:00.370)       0:09:25.247 ****** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_exporter_status] *************************\nTuesday 24 February 2026  00:56:40 +0000 (0:00:00.482)       0:09:25.731 ****** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-mds : Include create_mds_filesystems.yml] ***************************\nTuesday 24 February 2026  00:56:41 +0000 (0:00:00.951)       0:09:26.682 ****** \n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mincluded: /ansible/roles/ceph-mds/tasks/create_mds_filesystems.yml for testbed-node-3\u001b[0m\n\nTASK [ceph-facts : Get current default crush rule details] *********************\nTuesday 24 February 2026  00:56:41 +0000 (0:00:00.412)       0:09:27.095 ****** \n\u001b[0;32mok: [testbed-node-3 -> testbed-node-0(192.168.16.10)]\u001b[0m\n\nTASK [ceph-facts : Get current default crush rule name] ************************\nTuesday 24 February 2026  00:56:44 +0000 (0:00:02.379)       0:09:29.475 ****** \n\u001b[0;36mskipping: [testbed-node-3] => (item={'rule_id': 0, 'rule_name': 'replicated_rule', 'type': 1, 'steps': [{'op': 'take', 'item': -1, 'item_name': 'default'}, {'op': 'chooseleaf_firstn', 'num': 0, 'type': 'host'}, {'op': 'emit'}]}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-mds : Create filesystem pools] **************************************\nTuesday 24 February 2026  00:56:44 +0000 (0:00:00.246)       0:09:29.721 ****** \n\u001b[0;33mchanged: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item={'application': 'cephfs', 'erasure_profile': '', 'expected_num_objects': '', 'min_size': 0, 'name': 'cephfs_data', 'pg_num': 16, 'pgp_num': 16, 'rule_name': 'replicated_rule', 'size': 3, 'type': 1})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item={'application': 'cephfs', 'erasure_profile': '', 'expected_num_objects': '', 'min_size': 0, 'name': 'cephfs_metadata', 'pg_num': 16, 'pgp_num': 16, 'rule_name': 'replicated_rule', 'size': 3, 'type': 1})\u001b[0m\n\nTASK [ceph-mds : Create ceph filesystem] ***************************************\nTuesday 24 February 2026  00:56:54 +0000 (0:00:10.517)       0:09:40.239 ****** \n\u001b[0;33mchanged: [testbed-node-3 -> testbed-node-0(192.168.16.10)]\u001b[0m\n\nTASK [ceph-mds : Include common.yml] *******************************************\nTuesday 24 February 2026  00:56:58 +0000 (0:00:03.807)       0:09:44.046 ****** \n\u001b[0;36mincluded: /ansible/roles/ceph-mds/tasks/common.yml for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m\n\nTASK [ceph-mds : Create bootstrap-mds and mds directories] *********************\nTuesday 24 February 2026  00:56:59 +0000 (0:00:00.698)       0:09:44.745 ****** \n\u001b[0;32mok: [testbed-node-3] => (item=/var/lib/ceph/bootstrap-mds/)\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item=/var/lib/ceph/bootstrap-mds/)\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item=/var/lib/ceph/bootstrap-mds/)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=/var/lib/ceph/mds/ceph-testbed-node-4)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item=/var/lib/ceph/mds/ceph-testbed-node-3)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=/var/lib/ceph/mds/ceph-testbed-node-5)\u001b[0m\n\nTASK [ceph-mds : Get keys from monitors] ***************************************\nTuesday 24 February 2026  00:57:00 +0000 (0:00:01.112)       0:09:45.858 ****** \n\u001b[0;32mok: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=None) \u001b[0m\n\u001b[0;32mok: [testbed-node-3 -> {{ groups.get(mon_group_name)[0] }}]\u001b[0m\n\nTASK [ceph-mds : Copy ceph key(s) if needed] ***********************************\nTuesday 24 February 2026  00:57:03 +0000 (0:00:02.730)       0:09:48.589 ****** \n\u001b[0;33mchanged: [testbed-node-3] => (item=None)\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=None) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=None)\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=None) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=None)\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=None) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nTASK [ceph-mds : Create mds keyring] *******************************************\nTuesday 24 February 2026  00:57:04 +0000 (0:00:01.619)       0:09:50.209 ****** \n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\nTASK [ceph-mds : Non_containerized.yml] ****************************************\nTuesday 24 February 2026  00:57:07 +0000 (0:00:02.951)       0:09:53.160 ****** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-mds : Containerized.yml] ********************************************\nTuesday 24 February 2026  00:57:08 +0000 (0:00:00.381)       0:09:53.542 ****** \n\u001b[0;36mincluded: /ansible/roles/ceph-mds/tasks/containerized.yml for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m\n\nTASK [ceph-mds : Include_tasks systemd.yml] ************************************\nTuesday 24 February 2026  00:57:09 +0000 (0:00:01.029)       0:09:54.571 ****** \n\u001b[0;36mincluded: /ansible/roles/ceph-mds/tasks/systemd.yml for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m\n\nTASK [ceph-mds : Generate systemd unit file] ***********************************\nTuesday 24 February 2026  00:57:09 +0000 (0:00:00.670)       0:09:55.242 ****** \n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nTASK [ceph-mds : Generate systemd ceph-mds target file] ************************\nTuesday 24 February 2026  00:57:11 +0000 (0:00:01.400)       0:09:56.643 ****** \n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nTASK [ceph-mds : Enable ceph-mds.target] ***************************************\nTuesday 24 February 2026  00:57:13 +0000 (0:00:01.768)       0:09:58.411 ****** \n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nTASK [ceph-mds : Systemd start mds container] **********************************\nTuesday 24 February 2026  00:57:14 +0000 (0:00:01.862)       0:10:00.274 ****** \n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\nTASK [ceph-mds : Wait for mds socket to exist] *********************************\nTuesday 24 February 2026  00:57:17 +0000 (0:00:02.157)       0:10:02.431 ****** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Make tempdir for scripts] **********************\nTuesday 24 February 2026  00:57:18 +0000 (0:00:01.577)       0:10:04.009 ****** \n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Mdss handler] **********************************\nTuesday 24 February 2026  00:57:19 +0000 (0:00:00.753)       0:10:04.763 ****** \n\u001b[0;36mincluded: /ansible/roles/ceph-handler/tasks/handler_mdss.yml for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Set _mds_handler_called before restart] ********\nTuesday 24 February 2026  00:57:20 +0000 (0:00:00.985)       0:10:05.749 ****** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Copy mds restart script] ***********************\nTuesday 24 February 2026  00:57:20 +0000 (0:00:00.391)       0:10:06.140 ****** \n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Restart ceph mds daemon(s)] ********************\nTuesday 24 February 2026  00:57:22 +0000 (0:00:01.286)       0:10:07.427 ****** \n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-3) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-4) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-5) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Set _mds_handler_called after restart] *********\nTuesday 24 February 2026  00:57:23 +0000 (0:00:01.240)       0:10:08.668 ****** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nPLAY [Apply role ceph-rgw] *****************************************************\n\nTASK [ceph-handler : Include check_running_cluster.yml] ************************\nTuesday 24 February 2026  00:57:24 +0000 (0:00:01.020)       0:10:09.688 ****** \n\u001b[0;36mincluded: /ansible/roles/ceph-handler/tasks/check_running_cluster.yml for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m\n\nTASK [ceph-handler : Include check_running_containers.yml] *********************\nTuesday 24 February 2026  00:57:24 +0000 (0:00:00.589)       0:10:10.277 ****** \n\u001b[0;36mincluded: /ansible/roles/ceph-handler/tasks/check_running_containers.yml for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m\n\nTASK [ceph-handler : Check for a mon container] ********************************\nTuesday 24 February 2026  00:57:25 +0000 (0:00:00.975)       0:10:11.253 ****** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Check for an osd container] *******************************\nTuesday 24 February 2026  00:57:26 +0000 (0:00:00.340)       0:10:11.594 ****** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Check for a mds container] ********************************\nTuesday 24 February 2026  00:57:26 +0000 (0:00:00.759)       0:10:12.354 ****** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Check for a rgw container] ********************************\nTuesday 24 February 2026  00:57:28 +0000 (0:00:01.124)       0:10:13.479 ****** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Check for a mgr container] ********************************\nTuesday 24 February 2026  00:57:28 +0000 (0:00:00.779)       0:10:14.258 ****** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Check for a rbd mirror container] *************************\nTuesday 24 February 2026  00:57:29 +0000 (0:00:00.342)       0:10:14.601 ****** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Check for a nfs container] ********************************\nTuesday 24 February 2026  00:57:29 +0000 (0:00:00.343)       0:10:14.944 ****** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Check for a ceph-crash container] *************************\nTuesday 24 February 2026  00:57:29 +0000 (0:00:00.348)       0:10:15.293 ****** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Check for a ceph-exporter container] **********************\nTuesday 24 February 2026  00:57:31 +0000 (0:00:01.224)       0:10:16.518 ****** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Include check_socket_non_container.yml] *******************\nTuesday 24 February 2026  00:57:31 +0000 (0:00:00.785)       0:10:17.303 ****** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_mon_status] ******************************\nTuesday 24 February 2026  00:57:32 +0000 (0:00:00.326)       0:10:17.629 ****** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_osd_status] ******************************\nTuesday 24 February 2026  00:57:32 +0000 (0:00:00.345)       0:10:17.975 ****** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_mds_status] ******************************\nTuesday 24 February 2026  00:57:33 +0000 (0:00:00.805)       0:10:18.780 ****** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_rgw_status] ******************************\nTuesday 24 February 2026  00:57:33 +0000 (0:00:00.361)       0:10:19.141 ****** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_nfs_status] ******************************\nTuesday 24 February 2026  00:57:34 +0000 (0:00:00.414)       0:10:19.556 ****** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_rbd_status] ******************************\nTuesday 24 February 2026  00:57:34 +0000 (0:00:00.317)       0:10:19.874 ****** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_mgr_status] ******************************\nTuesday 24 February 2026  00:57:35 +0000 (0:00:00.657)       0:10:20.531 ****** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_crash_status] ****************************\nTuesday 24 February 2026  00:57:35 +0000 (0:00:00.396)       0:10:20.928 ****** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-handler : Set_fact handler_exporter_status] *************************\nTuesday 24 February 2026  00:57:35 +0000 (0:00:00.415)       0:10:21.343 ****** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-rgw : Include common.yml] *******************************************\nTuesday 24 February 2026  00:57:36 +0000 (0:00:00.855)       0:10:22.198 ****** \n\u001b[0;36mincluded: /ansible/roles/ceph-rgw/tasks/common.yml for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m\n\nTASK [ceph-rgw : Get keys from monitors] ***************************************\nTuesday 24 February 2026  00:57:37 +0000 (0:00:00.591)       0:10:22.790 ****** \n\u001b[0;32mok: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=None) \u001b[0m\n\u001b[0;32mok: [testbed-node-3 -> {{ groups.get(mon_group_name)[0] }}]\u001b[0m\n\nTASK [ceph-rgw : Copy ceph key(s) if needed] ***********************************\nTuesday 24 February 2026  00:57:39 +0000 (0:00:02.439)       0:10:25.230 ****** \n\u001b[0;33mchanged: [testbed-node-3] => (item=None)\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=None) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=None)\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=None) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=None)\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=None) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nTASK [ceph-rgw : Copy SSL certificate & key data to certificate path] **********\nTuesday 24 February 2026  00:57:41 +0000 (0:00:01.656)       0:10:26.886 ****** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-rgw : Include_tasks pre_requisite.yml] ******************************\nTuesday 24 February 2026  00:57:41 +0000 (0:00:00.376)       0:10:27.262 ****** \n\u001b[0;36mincluded: /ansible/roles/ceph-rgw/tasks/pre_requisite.yml for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m\n\nTASK [ceph-rgw : Create rados gateway directories] *****************************\nTuesday 24 February 2026  00:57:42 +0000 (0:00:00.552)       0:10:27.814 ****** \n\u001b[0;33mchanged: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item={'instance_name': 'rgw0', 'radosgw_address': '192.168.16.13', 'radosgw_frontend_port': 8081})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4 -> testbed-node-0(192.168.16.10)] => (item={'instance_name': 'rgw0', 'radosgw_address': '192.168.16.14', 'radosgw_frontend_port': 8081})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item={'instance_name': 'rgw0', 'radosgw_address': '192.168.16.15', 'radosgw_frontend_port': 8081})\u001b[0m\n\nTASK [ceph-rgw : Create rgw keyrings] ******************************************\nTuesday 24 February 2026  00:57:44 +0000 (0:00:01.634)       0:10:29.449 ****** \n\u001b[0;33mchanged: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3 -> {{ groups[mon_group_name][0] if groups.get(mon_group_name, []) | length > 0 else 'localhost' }}]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4 -> {{ groups[mon_group_name][0] if groups.get(mon_group_name, []) | length > 0 else 'localhost' }}]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5 -> {{ groups[mon_group_name][0] if groups.get(mon_group_name, []) | length > 0 else 'localhost' }}]\u001b[0m\n\nTASK [ceph-rgw : Get keys from monitors] ***************************************\nTuesday 24 February 2026  00:57:49 +0000 (0:00:05.535)       0:10:34.985 ****** \n\u001b[0;32mok: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m\n\u001b[0;32mok: [testbed-node-3 -> {{ groups.get(mon_group_name)[0] }}]\u001b[0m\n\u001b[0;32mok: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m\n\u001b[0;32mok: [testbed-node-5 -> {{ groups.get(mon_group_name)[0] }}]\u001b[0m\n\u001b[0;32mok: [testbed-node-4 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m\n\u001b[0;32mok: [testbed-node-4 -> {{ groups.get(mon_group_name)[0] }}]\u001b[0m\n\nTASK [ceph-rgw : Copy ceph key(s) if needed] ***********************************\nTuesday 24 February 2026  00:57:52 +0000 (0:00:02.978)       0:10:37.963 ****** \n\u001b[0;33mchanged: [testbed-node-3] => (item=None)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=None)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=None)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nTASK [ceph-rgw : Rgw pool creation tasks] **************************************\nTuesday 24 February 2026  00:57:53 +0000 (0:00:01.388)       0:10:39.351 ****** \n\u001b[0;36mincluded: /ansible/roles/ceph-rgw/tasks/rgw_create_pools.yml for testbed-node-3\u001b[0m\n\nTASK [ceph-rgw : Create ec profile] ********************************************\nTuesday 24 February 2026  00:57:54 +0000 (0:00:00.251)       0:10:39.603 ****** \n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'default.rgw.buckets.data', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'default.rgw.buckets.index', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'default.rgw.control', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'default.rgw.log', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'default.rgw.meta', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-rgw : Set crush rule] ***********************************************\nTuesday 24 February 2026  00:57:55 +0000 (0:00:01.301)       0:10:40.905 ****** \n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'default.rgw.buckets.data', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'default.rgw.buckets.index', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'default.rgw.control', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'default.rgw.log', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'default.rgw.meta', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-rgw : Create rgw pools] *********************************************\nTuesday 24 February 2026  00:57:56 +0000 (0:00:00.671)       0:10:41.577 ****** \n\u001b[0;33mchanged: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item={'key': 'default.rgw.buckets.data', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item={'key': 'default.rgw.buckets.index', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item={'key': 'default.rgw.control', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item={'key': 'default.rgw.log', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item={'key': 'default.rgw.meta', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}})\u001b[0m\n\nTASK [ceph-rgw : Include_tasks openstack-keystone.yml] *************************\nTuesday 24 February 2026  00:58:27 +0000 (0:00:31.785)       0:11:13.362 ****** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-rgw : Include_tasks start_radosgw.yml] ******************************\nTuesday 24 February 2026  00:58:28 +0000 (0:00:00.358)       0:11:13.721 ****** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-rgw : Include start_docker_rgw.yml] *********************************\nTuesday 24 February 2026  00:58:28 +0000 (0:00:00.350)       0:11:14.072 ****** \n\u001b[0;36mincluded: /ansible/roles/ceph-rgw/tasks/start_docker_rgw.yml for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m\n\nTASK [ceph-rgw : Include_task systemd.yml] *************************************\nTuesday 24 February 2026  00:58:29 +0000 (0:00:00.875)       0:11:14.947 ****** \n\u001b[0;36mincluded: /ansible/roles/ceph-rgw/tasks/systemd.yml for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m\n\nTASK [ceph-rgw : Generate systemd unit file] ***********************************\nTuesday 24 February 2026  00:58:30 +0000 (0:00:00.553)       0:11:15.500 ****** \n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nTASK [ceph-rgw : Generate systemd ceph-radosgw target file] ********************\nTuesday 24 February 2026  00:58:31 +0000 (0:00:01.332)       0:11:16.833 ****** \n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nTASK [ceph-rgw : Enable ceph-radosgw.target] ***********************************\nTuesday 24 February 2026  00:58:33 +0000 (0:00:01.637)       0:11:18.470 ****** \n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\nTASK [ceph-rgw : Systemd start rgw container] **********************************\nTuesday 24 February 2026  00:58:35 +0000 (0:00:02.045)       0:11:20.515 ****** \n\u001b[0;33mchanged: [testbed-node-4] => (item={'instance_name': 'rgw0', 'radosgw_address': '192.168.16.14', 'radosgw_frontend_port': 8081})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'instance_name': 'rgw0', 'radosgw_address': '192.168.16.13', 'radosgw_frontend_port': 8081})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'instance_name': 'rgw0', 'radosgw_address': '192.168.16.15', 'radosgw_frontend_port': 8081})\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Make tempdir for scripts] **********************\nTuesday 24 February 2026  00:58:37 +0000 (0:00:02.766)       0:11:23.282 ****** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Rgws handler] **********************************\nTuesday 24 February 2026  00:58:38 +0000 (0:00:00.390)       0:11:23.672 ****** \n\u001b[0;36mincluded: /ansible/roles/ceph-handler/tasks/handler_rgws.yml for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Set _rgw_handler_called before restart] ********\nTuesday 24 February 2026  00:58:38 +0000 (0:00:00.572)       0:11:24.245 ****** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Copy rgw restart script] ***********************\nTuesday 24 February 2026  00:58:39 +0000 (0:00:00.663)       0:11:24.908 ****** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Restart ceph rgw daemon(s)] ********************\nTuesday 24 February 2026  00:58:39 +0000 (0:00:00.389)       0:11:25.298 ****** \n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-3) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-4) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-5) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nRUNNING HANDLER [ceph-handler : Set _rgw_handler_called after restart] *********\nTuesday 24 February 2026  00:58:40 +0000 (0:00:00.704)       0:11:26.003 ****** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=134 \u001b[0m \u001b[0;33mchanged=35  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=125 \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=127 \u001b[0m \u001b[0;33mchanged=32  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=120 \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=134 \u001b[0m \u001b[0;33mchanged=33  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=119 \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-3\u001b[0m             : \u001b[0;32mok=193 \u001b[0m \u001b[0;33mchanged=45  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=162 \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-4\u001b[0m             : \u001b[0;32mok=175 \u001b[0m \u001b[0;33mchanged=40  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=123 \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-5\u001b[0m             : \u001b[0;32mok=177 \u001b[0m \u001b[0;33mchanged=41  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=121 \u001b[0m rescued=0    ignored=0   \n\n\n\nTASKS RECAP ********************************************************************\nTuesday 24 February 2026  00:58:40 +0000 (0:00:00.266)       0:11:26.270 ****** \n=============================================================================== \nceph-osd : Use ceph-volume to create osds ------------------------------ 48.43s\nceph-container-common : Pulling Ceph container image ------------------- 40.43s\nceph-mgr : Wait for all mgr to be up ----------------------------------- 36.57s\nceph-rgw : Create rgw pools -------------------------------------------- 31.79s\nceph-mon : Waiting for the monitor(s) to form the quorum... ------------ 21.91s\nceph-mon : Set cluster configs ----------------------------------------- 14.80s\nceph-osd : Wait for all osd to be up ----------------------------------- 12.82s\nceph-mds : Create filesystem pools ------------------------------------- 10.52s\nceph-mgr : Create ceph mgr keyring(s) on a mon node -------------------- 10.21s\nceph-mon : Fetch ceph initial keys -------------------------------------- 9.27s\nceph-mgr : Disable ceph mgr enabled modules ----------------------------- 6.72s\nceph-config : Create ceph initial directories --------------------------- 6.66s\nceph-mgr : Add modules to ceph-mgr -------------------------------------- 5.72s\nceph-rgw : Create rgw keyrings ------------------------------------------ 5.54s\nceph-crash : Create client.crash keyring -------------------------------- 4.66s\nceph-osd : Apply operating system tuning -------------------------------- 4.28s\nceph-facts : Set_fact _monitor_addresses - ipv4 ------------------------- 3.98s\nceph-osd : Systemd start osd -------------------------------------------- 3.88s\nceph-mds : Create ceph filesystem --------------------------------------- 3.81s\nceph-crash : Start the ceph-crash service ------------------------------- 3.70s\n\u001b[32m2026-02-24 00:58:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 21cb4ba5-3e33-4fe4-8954-1e96067dfc1b is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:58:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 00:58:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ca6226d5-0f70-4445-9f9f-dca4faf1e39d is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:58:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 460230b5-3593-4a25-becb-4c4727892a2e is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:58:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 21cb4ba5-3e33-4fe4-8954-1e96067dfc1b is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:58:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 00:58:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ca6226d5-0f70-4445-9f9f-dca4faf1e39d is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:58:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 460230b5-3593-4a25-becb-4c4727892a2e is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:58:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 21cb4ba5-3e33-4fe4-8954-1e96067dfc1b is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:58:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 00:58:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ca6226d5-0f70-4445-9f9f-dca4faf1e39d is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:58:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 460230b5-3593-4a25-becb-4c4727892a2e is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:58:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 21cb4ba5-3e33-4fe4-8954-1e96067dfc1b is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:58:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 00:58:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ca6226d5-0f70-4445-9f9f-dca4faf1e39d is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:58:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 460230b5-3593-4a25-becb-4c4727892a2e is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:58:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 21cb4ba5-3e33-4fe4-8954-1e96067dfc1b is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:58:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 00:58:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ca6226d5-0f70-4445-9f9f-dca4faf1e39d is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:58:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 460230b5-3593-4a25-becb-4c4727892a2e is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:58:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 21cb4ba5-3e33-4fe4-8954-1e96067dfc1b is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:58:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 00:59:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ca6226d5-0f70-4445-9f9f-dca4faf1e39d is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:59:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 460230b5-3593-4a25-becb-4c4727892a2e is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:59:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 21cb4ba5-3e33-4fe4-8954-1e96067dfc1b is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:59:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 00:59:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ca6226d5-0f70-4445-9f9f-dca4faf1e39d is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:59:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 460230b5-3593-4a25-becb-4c4727892a2e is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:59:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 21cb4ba5-3e33-4fe4-8954-1e96067dfc1b is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:59:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 00:59:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ca6226d5-0f70-4445-9f9f-dca4faf1e39d is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:59:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 460230b5-3593-4a25-becb-4c4727892a2e is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:59:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 21cb4ba5-3e33-4fe4-8954-1e96067dfc1b is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:59:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 00:59:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ca6226d5-0f70-4445-9f9f-dca4faf1e39d is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:59:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 460230b5-3593-4a25-becb-4c4727892a2e is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:59:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 21cb4ba5-3e33-4fe4-8954-1e96067dfc1b is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:59:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 00:59:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ca6226d5-0f70-4445-9f9f-dca4faf1e39d is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:59:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 460230b5-3593-4a25-becb-4c4727892a2e is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:59:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 21cb4ba5-3e33-4fe4-8954-1e96067dfc1b is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:59:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 00:59:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ca6226d5-0f70-4445-9f9f-dca4faf1e39d is in state SUCCESS\u001b[0m\n\n\nPLAY [Group hosts based on configuration] **************************************\n\nTASK [Group hosts based on Kolla action] ***************************************\nTuesday 24 February 2026  00:56:14 +0000 (0:00:00.367)       0:00:00.367 ****** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [Group hosts based on enabled services] ***********************************\nTuesday 24 February 2026  00:56:14 +0000 (0:00:00.377)       0:00:00.744 ****** \n\u001b[0;32mok: [testbed-node-0] => (item=enable_opensearch_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item=enable_opensearch_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item=enable_opensearch_True)\u001b[0m\n\nPLAY [Apply role opensearch] ***************************************************\n\nTASK [opensearch : include_tasks] **********************************************\nTuesday 24 February 2026  00:56:15 +0000 (0:00:00.495)       0:00:01.240 ****** \n\u001b[0;36mincluded: /ansible/roles/opensearch/tasks/deploy.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [opensearch : Setting sysctl values] **************************************\nTuesday 24 February 2026  00:56:15 +0000 (0:00:00.627)       0:00:01.868 ****** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'name': 'vm.max_map_count', 'value': 262144})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'name': 'vm.max_map_count', 'value': 262144})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'name': 'vm.max_map_count', 'value': 262144})\u001b[0m\n\nTASK [opensearch : Ensuring config directories exist] **************************\nTuesday 24 February 2026  00:56:16 +0000 (0:00:00.877)       0:00:02.746 ****** \n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/opensearch:2024.2', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/opensearch:2024.2', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/opensearch:2024.2', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/opensearch-dashboards:2024.2', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/opensearch-dashboards:2024.2', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/opensearch-dashboards:2024.2', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}}}})\u001b[0m\n\nTASK [opensearch : include_tasks] **********************************************\nTuesday 24 February 2026  00:56:18 +0000 (0:00:02.098)       0:00:04.844 ****** \n\u001b[0;36mincluded: /ansible/roles/opensearch/tasks/copy-certs.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [service-cert-copy : opensearch | Copying over extra CA certificates] *****\nTuesday 24 February 2026  00:56:19 +0000 (0:00:00.746)       0:00:05.591 ****** \n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/opensearch:2024.2', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/opensearch:2024.2', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/opensearch:2024.2', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/opensearch-dashboards:2024.2', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/opensearch-dashboards:2024.2', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/opensearch-dashboards:2024.2', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}}}})\u001b[0m\n\nTASK [service-cert-copy : opensearch | Copying over backend internal TLS certificate] ***\nTuesday 24 February 2026  00:56:23 +0000 (0:00:03.963)       0:00:09.554 ****** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/opensearch:2024.2', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/opensearch-dashboards:2024.2', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/opensearch:2024.2', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/opensearch-dashboards:2024.2', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/opensearch:2024.2', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/opensearch-dashboards:2024.2', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [service-cert-copy : opensearch | Copying over backend internal TLS key] ***\nTuesday 24 February 2026  00:56:25 +0000 (0:00:01.992)       0:00:11.547 ****** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/opensearch:2024.2', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/opensearch-dashboards:2024.2', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/opensearch:2024.2', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/opensearch-dashboards:2024.2', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/opensearch:2024.2', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/opensearch-dashboards:2024.2', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [opensearch : Copying over config.json files for services] ****************\nTuesday 24 February 2026  00:56:26 +0000 (0:00:01.278)       0:00:12.826 ****** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/opensearch:2024.2', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/opensearch:2024.2', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/opensearch:2024.2', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/opensearch-dashboards:2024.2', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/opensearch-dashboards:2024.2', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/opensearch-dashboards:2024.2', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}}}})\u001b[0m\n\nTASK [opensearch : Copying over opensearch service config file] ****************\nTuesday 24 February 2026  00:56:29 +0000 (0:00:02.769)       0:00:15.595 ****** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [opensearch : Copying over opensearch-dashboards config file] *************\nTuesday 24 February 2026  00:56:32 +0000 (0:00:03.450)       0:00:19.046 ****** \n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [opensearch : Check opensearch containers] ********************************\nTuesday 24 February 2026  00:56:35 +0000 (0:00:02.551)       0:00:21.597 ****** \n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/opensearch:2024.2', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/opensearch:2024.2', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/opensearch:2024.2', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/opensearch-dashboards:2024.2', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/opensearch-dashboards:2024.2', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/opensearch-dashboards:2024.2', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}}}})\u001b[0m\n\nTASK [opensearch : include_tasks] **********************************************\nTuesday 24 February 2026  00:56:38 +0000 (0:00:02.653)       0:00:24.251 ****** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [opensearch : Flush handlers] *********************************************\nTuesday 24 February 2026  00:56:38 +0000 (0:00:00.367)       0:00:24.619 ****** \n\nTASK [opensearch : Flush handlers] *********************************************\nTuesday 24 February 2026  00:56:38 +0000 (0:00:00.095)       0:00:24.715 ****** \n\nTASK [opensearch : Flush handlers] *********************************************\nTuesday 24 February 2026  00:56:38 +0000 (0:00:00.076)       0:00:24.791 ****** \n\nRUNNING HANDLER [opensearch : Disable shard allocation] ************************\nTuesday 24 February 2026  00:56:38 +0000 (0:00:00.076)       0:00:24.867 ****** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nRUNNING HANDLER [opensearch : Perform a flush] *********************************\nTuesday 24 February 2026  00:56:39 +0000 (0:00:00.857)       0:00:25.725 ****** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nRUNNING HANDLER [opensearch : Restart opensearch container] ********************\nTuesday 24 February 2026  00:56:39 +0000 (0:00:00.269)       0:00:25.994 ****** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\nRUNNING HANDLER [opensearch : Restart opensearch-dashboards container] *********\nTuesday 24 February 2026  00:57:43 +0000 (0:01:04.121)       0:01:30.115 ****** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\nTASK [opensearch : include_tasks] **********************************************\nTuesday 24 February 2026  00:58:58 +0000 (0:01:14.250)       0:02:44.366 ****** \n\u001b[0;36mincluded: /ansible/roles/opensearch/tasks/post-config.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [opensearch : Wait for OpenSearch to become ready] ************************\nTuesday 24 February 2026  00:58:59 +0000 (0:00:00.848)       0:02:45.214 ****** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [opensearch : Wait for OpenSearch cluster to become healthy] **************\nTuesday 24 February 2026  00:59:02 +0000 (0:00:03.085)       0:02:48.300 ****** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [opensearch : Check if a log retention policy exists] *********************\nTuesday 24 February 2026  00:59:05 +0000 (0:00:02.915)       0:02:51.215 ****** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [opensearch : Create new log retention policy] ****************************\nTuesday 24 February 2026  00:59:07 +0000 (0:00:02.839)       0:02:54.055 ****** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [opensearch : Apply retention policy to existing indices] *****************\nTuesday 24 February 2026  00:59:11 +0000 (0:00:03.362)       0:02:57.418 ****** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=19  \u001b[0m \u001b[0;33mchanged=11  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=5   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=14  \u001b[0m \u001b[0;33mchanged=9   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=3   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=14  \u001b[0m \u001b[0;33mchanged=9   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=3   \u001b[0m rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nTuesday 24 February 2026  00:59:14 +0000 (0:00:03.258)       0:03:00.676 ****** \n=============================================================================== \nopensearch : Restart opensearch-dashboards container ------------------- 74.25s\nopensearch : Restart opensearch container ------------------------------ 64.12s\nservice-cert-copy : opensearch | Copying over extra CA certificates ----- 3.96s\nopensearch : Copying over opensearch service config file ---------------- 3.45s\nopensearch : Create new log retention policy ---------------------------- 3.36s\nopensearch : Apply retention policy to existing indices ----------------- 3.26s\nopensearch : Wait for OpenSearch to become ready ------------------------ 3.09s\nopensearch : Wait for OpenSearch cluster to become healthy -------------- 2.92s\nopensearch : Check if a log retention policy exists --------------------- 2.84s\nopensearch : Copying over config.json files for services ---------------- 2.77s\nopensearch : Check opensearch containers -------------------------------- 2.65s\nopensearch : Copying over opensearch-dashboards config file ------------- 2.55s\nopensearch : Ensuring config directories exist -------------------------- 2.10s\nservice-cert-copy : opensearch | Copying over backend internal TLS certificate --- 1.99s\nservice-cert-copy : opensearch | Copying over backend internal TLS key --- 1.28s\nopensearch : Setting sysctl values -------------------------------------- 0.88s\nopensearch : Disable shard allocation ----------------------------------- 0.86s\nopensearch : include_tasks ---------------------------------------------- 0.85s\nopensearch : include_tasks ---------------------------------------------- 0.75s\nopensearch : include_tasks ---------------------------------------------- 0.63s\n\u001b[32m2026-02-24 00:59:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 460230b5-3593-4a25-becb-4c4727892a2e is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:59:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 21cb4ba5-3e33-4fe4-8954-1e96067dfc1b is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:59:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 00:59:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 460230b5-3593-4a25-becb-4c4727892a2e is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:59:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 21cb4ba5-3e33-4fe4-8954-1e96067dfc1b is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:59:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 00:59:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 460230b5-3593-4a25-becb-4c4727892a2e is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:59:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 21cb4ba5-3e33-4fe4-8954-1e96067dfc1b is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:59:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 00:59:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 460230b5-3593-4a25-becb-4c4727892a2e is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:59:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 21cb4ba5-3e33-4fe4-8954-1e96067dfc1b is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:59:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 00:59:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 460230b5-3593-4a25-becb-4c4727892a2e is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:59:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 21cb4ba5-3e33-4fe4-8954-1e96067dfc1b is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:59:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 00:59:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 460230b5-3593-4a25-becb-4c4727892a2e is in state STARTED\u001b[0m\n\n\nPLAY [Set kolla_action_mariadb] ************************************************\n\nTASK [Inform the user about the following task] ********************************\nTuesday 24 February 2026  00:56:13 +0000 (0:00:00.107)       0:00:00.107 ****** \n\u001b[0;32mok: [localhost] => {\u001b[0m\n\u001b[0;32m    \"msg\": \"The task 'Check MariaDB service' fails if the MariaDB service has not yet been deployed. This is fine.\"\u001b[0m\n\u001b[0;32m}\u001b[0m\n\nTASK [Check MariaDB service] ***************************************************\nTuesday 24 February 2026  00:56:13 +0000 (0:00:00.068)       0:00:00.175 ****** \n\u001b[0;31mfatal: [localhost]: FAILED! => {\"changed\": false, \"elapsed\": 2, \"msg\": \"Timeout when waiting for search string MariaDB in 192.168.16.9:3306\"}\u001b[0m\n\u001b[0;36m...ignoring\u001b[0m\n\nTASK [Set kolla_action_mariadb = upgrade if MariaDB is already running] ********\nTuesday 24 February 2026  00:56:16 +0000 (0:00:02.965)       0:00:03.141 ****** \n\u001b[0;36mskipping: [localhost]\u001b[0m\n\nTASK [Set kolla_action_mariadb = kolla_action_ng] ******************************\nTuesday 24 February 2026  00:56:16 +0000 (0:00:00.062)       0:00:03.203 ****** \n\u001b[0;32mok: [localhost]\u001b[0m\n\nPLAY [Group hosts based on configuration] **************************************\n\nTASK [Group hosts based on Kolla action] ***************************************\nTuesday 24 February 2026  00:56:17 +0000 (0:00:00.166)       0:00:03.370 ****** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [Group hosts based on enabled services] ***********************************\nTuesday 24 February 2026  00:56:17 +0000 (0:00:00.350)       0:00:03.721 ****** \n\u001b[0;32mok: [testbed-node-0] => (item=enable_mariadb_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item=enable_mariadb_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item=enable_mariadb_True)\u001b[0m\n\nPLAY [Apply role mariadb] ******************************************************\n\nTASK [mariadb : Group MariaDB hosts based on shards] ***************************\nTuesday 24 February 2026  00:56:18 +0000 (0:00:00.639)       0:00:04.360 ****** \n\u001b[0;32mok: [testbed-node-0] => (item=testbed-node-0)\u001b[0m\n\u001b[0;32mok: [testbed-node-0] => (item=testbed-node-1)\u001b[0m\n\u001b[0;32mok: [testbed-node-0] => (item=testbed-node-2)\u001b[0m\n\nTASK [mariadb : include_tasks] *************************************************\nTuesday 24 February 2026  00:56:18 +0000 (0:00:00.480)       0:00:04.841 ****** \n\u001b[0;36mincluded: /ansible/roles/mariadb/tasks/deploy.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [mariadb : Ensuring config directories exist] *****************************\nTuesday 24 February 2026  00:56:19 +0000 (0:00:00.972)       0:00:05.813 ****** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/mariadb-server:2024.2', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.10', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/mariadb-server:2024.2', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.12', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/mariadb-server:2024.2', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.11', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}})\u001b[0m\n\nTASK [mariadb : Ensuring database backup config directory exists] **************\nTuesday 24 February 2026  00:56:24 +0000 (0:00:05.059)       0:00:10.873 ****** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [mariadb : Copying over my.cnf for mariabackup] ***************************\nTuesday 24 February 2026  00:56:25 +0000 (0:00:01.083)       0:00:11.957 ****** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [mariadb : Copying over config.json files for services] *******************\nTuesday 24 February 2026  00:56:27 +0000 (0:00:02.056)       0:00:14.013 ****** \n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/mariadb-server:2024.2', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.11', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/mariadb-server:2024.2', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.10', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/mariadb-server:2024.2', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.12', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}})\u001b[0m\n\nTASK [mariadb : Copying over config.json files for mariabackup] ****************\nTuesday 24 February 2026  00:56:32 +0000 (0:00:04.539)       0:00:18.553 ****** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [mariadb : Copying over galera.cnf] ***************************************\nTuesday 24 February 2026  00:56:33 +0000 (0:00:01.322)       0:00:19.875 ****** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [mariadb : include_tasks] *************************************************\nTuesday 24 February 2026  00:56:39 +0000 (0:00:05.462)       0:00:25.338 ****** \n\u001b[0;36mincluded: /ansible/roles/mariadb/tasks/copy-certs.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [service-cert-copy : mariadb | Copying over extra CA certificates] ********\nTuesday 24 February 2026  00:56:39 +0000 (0:00:00.627)       0:00:25.965 ****** \n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/mariadb-server:2024.2', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.11', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/mariadb-server:2024.2', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.12', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/mariadb-server:2024.2', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.10', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [service-cert-copy : mariadb | Copying over backend internal TLS certificate] ***\nTuesday 24 February 2026  00:56:44 +0000 (0:00:04.403)       0:00:30.368 ****** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/mariadb-server:2024.2', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.10', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/mariadb-server:2024.2', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.12', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/mariadb-server:2024.2', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.11', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [service-cert-copy : mariadb | Copying over backend internal TLS key] *****\nTuesday 24 February 2026  00:56:47 +0000 (0:00:03.636)       0:00:34.005 ****** \n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/mariadb-server:2024.2', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.11', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/mariadb-server:2024.2', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.10', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/mariadb-server:2024.2', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.12', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [mariadb : Check mariadb containers] **************************************\nTuesday 24 February 2026  00:56:51 +0000 (0:00:03.614)       0:00:37.620 ****** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/mariadb-server:2024.2', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.10', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/mariadb-server:2024.2', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.11', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/mariadb-server:2024.2', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.12', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}})\u001b[0m\n\nTASK [mariadb : Create MariaDB volume] *****************************************\nTuesday 24 February 2026  00:56:55 +0000 (0:00:04.441)       0:00:42.062 ****** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [mariadb : Divide hosts by their MariaDB volume availability] *************\nTuesday 24 February 2026  00:56:56 +0000 (0:00:00.816)       0:00:42.878 ****** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [mariadb : Establish whether the cluster has already existed] *************\nTuesday 24 February 2026  00:56:57 +0000 (0:00:00.608)       0:00:43.487 ****** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [mariadb : Check MariaDB service port liveness] ***************************\nTuesday 24 February 2026  00:56:57 +0000 (0:00:00.350)       0:00:43.837 ****** \n\u001b[0;31mfatal: [testbed-node-1]: FAILED! => {\"changed\": false, \"elapsed\": 10, \"msg\": \"Timeout when waiting for search string MariaDB in 192.168.16.11:3306\"}\u001b[0m\n\u001b[0;36m...ignoring\u001b[0m\n\u001b[0;31mfatal: [testbed-node-0]: FAILED! => {\"changed\": false, \"elapsed\": 10, \"msg\": \"Timeout when waiting for search string MariaDB in 192.168.16.10:3306\"}\u001b[0m\n\u001b[0;36m...ignoring\u001b[0m\n\u001b[0;31mfatal: [testbed-node-2]: FAILED! => {\"changed\": false, \"elapsed\": 10, \"msg\": \"Timeout when waiting for search string MariaDB in 192.168.16.12:3306\"}\u001b[0m\n\u001b[0;36m...ignoring\u001b[0m\n\nTASK [mariadb : Divide hosts by their MariaDB service port liveness] ***********\nTuesday 24 February 2026  00:57:08 +0000 (0:00:10.988)       0:00:54.825 ****** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [mariadb : Fail on existing but stopped cluster] **************************\nTuesday 24 February 2026  00:57:09 +0000 (0:00:00.470)       0:00:55.296 ****** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [mariadb : Check MariaDB service WSREP sync status] ***********************\nTuesday 24 February 2026  00:57:09 +0000 (0:00:00.760)       0:00:56.056 ****** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [mariadb : Extract MariaDB service WSREP sync status] *********************\nTuesday 24 February 2026  00:57:10 +0000 (0:00:00.641)       0:00:56.698 ****** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [mariadb : Divide hosts by their MariaDB service WSREP sync status] *******\nTuesday 24 February 2026  00:57:10 +0000 (0:00:00.511)       0:00:57.210 ****** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [mariadb : Fail when MariaDB services are not synced across the whole cluster] ***\nTuesday 24 February 2026  00:57:11 +0000 (0:00:00.475)       0:00:57.686 ****** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [mariadb : include_tasks] *************************************************\nTuesday 24 February 2026  00:57:12 +0000 (0:00:00.881)       0:00:58.567 ****** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mincluded: /ansible/roles/mariadb/tasks/bootstrap_cluster.yml for testbed-node-0\u001b[0m\n\nTASK [mariadb : Running MariaDB bootstrap container] ***************************\nTuesday 24 February 2026  00:57:12 +0000 (0:00:00.451)       0:00:59.018 ****** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [mariadb : Store bootstrap host name into facts] **************************\nTuesday 24 February 2026  00:57:23 +0000 (0:00:10.911)       0:01:09.930 ****** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [mariadb : include_tasks] *************************************************\nTuesday 24 February 2026  00:57:23 +0000 (0:00:00.147)       0:01:10.077 ****** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [mariadb : Starting first MariaDB container] *******************\nTuesday 24 February 2026  00:57:24 +0000 (0:00:01.184)       0:01:11.262 ****** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nRUNNING HANDLER [mariadb : Wait for first MariaDB service port liveness] *******\nTuesday 24 February 2026  00:57:33 +0000 (0:00:08.946)       0:01:20.208 ****** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nRUNNING HANDLER [mariadb : Wait for first MariaDB service to sync WSREP] *******\nTuesday 24 February 2026  00:57:35 +0000 (0:00:01.641)       0:01:21.850 ****** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nRUNNING HANDLER [mariadb : Ensure MariaDB is running normally on bootstrap host] ***\nTuesday 24 February 2026  00:57:38 +0000 (0:00:02.889)       0:01:24.739 ****** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nRUNNING HANDLER [mariadb : Restart MariaDB on existing cluster members] ********\nTuesday 24 February 2026  00:57:38 +0000 (0:00:00.124)       0:01:24.863 ****** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [mariadb : Start MariaDB on new nodes] *************************\nTuesday 24 February 2026  00:57:38 +0000 (0:00:00.342)       0:01:25.206 ****** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[1;35m[WARNING]: Could not match supplied host pattern, ignoring: mariadb_restart\u001b[0m\n\nPLAY [Restart mariadb services] ************************************************\n\u001b[0;36mskipping: no hosts matched\u001b[0m\n\nPLAY [Start mariadb services] **************************************************\n\nTASK [mariadb : Restart MariaDB container] *************************************\nTuesday 24 February 2026  00:57:39 +0000 (0:00:00.603)       0:01:25.810 ****** \n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\nTASK [mariadb : Wait for MariaDB service port liveness] ************************\nTuesday 24 February 2026  00:57:58 +0000 (0:00:18.602)       0:01:44.413 ****** \n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\nTASK [mariadb : Wait for MariaDB service to sync WSREP] ************************\nTuesday 24 February 2026  00:58:13 +0000 (0:00:15.600)       0:02:00.013 ****** \n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\nPLAY [Start mariadb services] **************************************************\n\nTASK [mariadb : Restart MariaDB container] *************************************\nTuesday 24 February 2026  00:58:16 +0000 (0:00:02.675)       0:02:02.689 ****** \n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [mariadb : Wait for MariaDB service port liveness] ************************\nTuesday 24 February 2026  00:58:35 +0000 (0:00:19.127)       0:02:21.817 ****** \n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [mariadb : Wait for MariaDB service to sync WSREP] ************************\nTuesday 24 February 2026  00:58:51 +0000 (0:00:15.622)       0:02:37.439 ****** \n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nPLAY [Restart bootstrap mariadb service] ***************************************\n\nTASK [mariadb : Restart MariaDB container] *************************************\nTuesday 24 February 2026  00:58:53 +0000 (0:00:02.735)       0:02:40.175 ****** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [mariadb : Wait for MariaDB service port liveness] ************************\nTuesday 24 February 2026  00:59:06 +0000 (0:00:12.798)       0:02:52.973 ****** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [mariadb : Wait for MariaDB service to sync WSREP] ************************\nTuesday 24 February 2026  00:59:11 +0000 (0:00:04.639)       0:02:57.613 ****** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nPLAY [Apply mariadb post-configuration] ****************************************\n\nTASK [Include mariadb post-deploy.yml] *****************************************\nTuesday 24 February 2026  00:59:14 +0000 (0:00:02.977)       0:03:00.590 ****** \n\u001b[0;36mincluded: mariadb for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [mariadb : Creating shard root mysql user] ********************************\nTuesday 24 February 2026  00:59:14 +0000 (0:00:00.492)       0:03:01.083 ****** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [mariadb : Creating mysql monitor user] ***********************************\nTuesday 24 February 2026  00:59:17 +0000 (0:00:02.868)       0:03:03.951 ****** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [mariadb : Creating database backup user and setting permissions] *********\nTuesday 24 February 2026  00:59:20 +0000 (0:00:03.083)       0:03:07.034 ****** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [mariadb : Granting permissions on Mariabackup database to backup user] ***\nTuesday 24 February 2026  00:59:23 +0000 (0:00:02.893)       0:03:09.928 ****** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [mariadb : Wait for MariaDB service to be ready through VIP] **************\nTuesday 24 February 2026  00:59:26 +0000 (0:00:02.522)       0:03:12.450 ****** \n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [Include mariadb post-upgrade.yml] ****************************************\nTuesday 24 February 2026  00:59:29 +0000 (0:00:03.582)       0:03:16.032 ****** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;32mlocalhost\u001b[0m                  : \u001b[0;32mok=3   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=1   \u001b[0m rescued=0    \u001b[1;35mignored=1   \u001b[0m\n\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=34  \u001b[0m \u001b[0;33mchanged=16  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=11  \u001b[0m rescued=0    \u001b[1;35mignored=1   \u001b[0m\n\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=20  \u001b[0m \u001b[0;33mchanged=7   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=18  \u001b[0m rescued=0    \u001b[1;35mignored=1   \u001b[0m\n\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=20  \u001b[0m \u001b[0;33mchanged=7   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=18  \u001b[0m rescued=0    \u001b[1;35mignored=1   \u001b[0m\n\n\nTASKS RECAP ********************************************************************\nTuesday 24 February 2026  00:59:30 +0000 (0:00:00.274)       0:03:16.306 ****** \n=============================================================================== \nmariadb : Restart MariaDB container ------------------------------------ 37.73s\nmariadb : Wait for MariaDB service port liveness ----------------------- 31.22s\nmariadb : Restart MariaDB container ------------------------------------ 12.80s\nmariadb : Check MariaDB service port liveness -------------------------- 10.99s\nmariadb : Running MariaDB bootstrap container -------------------------- 10.91s\nmariadb : Starting first MariaDB container ------------------------------ 8.95s\nmariadb : Copying over galera.cnf --------------------------------------- 5.46s\nmariadb : Wait for MariaDB service to sync WSREP ------------------------ 5.41s\nmariadb : Ensuring config directories exist ----------------------------- 5.06s\nmariadb : Wait for MariaDB service port liveness ------------------------ 4.64s\nmariadb : Copying over config.json files for services ------------------- 4.54s\nmariadb : Check mariadb containers -------------------------------------- 4.44s\nservice-cert-copy : mariadb | Copying over extra CA certificates -------- 4.40s\nservice-cert-copy : mariadb | Copying over backend internal TLS certificate --- 3.64s\nservice-cert-copy : mariadb | Copying over backend internal TLS key ----- 3.61s\nmariadb : Wait for MariaDB service to be ready through VIP -------------- 3.58s\nmariadb : Creating mysql monitor user ----------------------------------- 3.08s\nmariadb : Wait for MariaDB service to sync WSREP ------------------------ 2.98s\nCheck MariaDB service --------------------------------------------------- 2.97s\nmariadb : Creating database backup user and setting permissions --------- 2.89s\n\u001b[32m2026-02-24 00:59:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 21cb4ba5-3e33-4fe4-8954-1e96067dfc1b is in state SUCCESS\u001b[0m\n\u001b[32m2026-02-24 00:59:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 00:59:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c587e61e-d18f-4e39-9404-027cd3697185 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:59:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9cb83792-8f4e-4574-a78c-9b0e654d6fd6 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:59:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 460230b5-3593-4a25-becb-4c4727892a2e is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:59:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 00:59:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c587e61e-d18f-4e39-9404-027cd3697185 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:59:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9cb83792-8f4e-4574-a78c-9b0e654d6fd6 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:59:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 460230b5-3593-4a25-becb-4c4727892a2e is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:59:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 00:59:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c587e61e-d18f-4e39-9404-027cd3697185 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:59:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9cb83792-8f4e-4574-a78c-9b0e654d6fd6 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:59:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 460230b5-3593-4a25-becb-4c4727892a2e is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:59:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 00:59:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c587e61e-d18f-4e39-9404-027cd3697185 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:59:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9cb83792-8f4e-4574-a78c-9b0e654d6fd6 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:59:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 460230b5-3593-4a25-becb-4c4727892a2e is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:59:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 00:59:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c587e61e-d18f-4e39-9404-027cd3697185 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:59:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9cb83792-8f4e-4574-a78c-9b0e654d6fd6 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:59:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 460230b5-3593-4a25-becb-4c4727892a2e is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:59:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 00:59:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c587e61e-d18f-4e39-9404-027cd3697185 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:59:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9cb83792-8f4e-4574-a78c-9b0e654d6fd6 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:59:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 460230b5-3593-4a25-becb-4c4727892a2e is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:59:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 00:59:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c587e61e-d18f-4e39-9404-027cd3697185 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:59:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9cb83792-8f4e-4574-a78c-9b0e654d6fd6 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:59:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 460230b5-3593-4a25-becb-4c4727892a2e is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:59:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 00:59:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c587e61e-d18f-4e39-9404-027cd3697185 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:59:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9cb83792-8f4e-4574-a78c-9b0e654d6fd6 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:59:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 460230b5-3593-4a25-becb-4c4727892a2e is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:59:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 00:59:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c587e61e-d18f-4e39-9404-027cd3697185 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:59:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9cb83792-8f4e-4574-a78c-9b0e654d6fd6 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:59:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 460230b5-3593-4a25-becb-4c4727892a2e is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 00:59:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 01:00:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c587e61e-d18f-4e39-9404-027cd3697185 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:00:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9cb83792-8f4e-4574-a78c-9b0e654d6fd6 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:00:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 460230b5-3593-4a25-becb-4c4727892a2e is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:00:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 01:00:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c587e61e-d18f-4e39-9404-027cd3697185 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:00:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9cb83792-8f4e-4574-a78c-9b0e654d6fd6 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:00:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 460230b5-3593-4a25-becb-4c4727892a2e is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:00:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 01:00:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c587e61e-d18f-4e39-9404-027cd3697185 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:00:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9cb83792-8f4e-4574-a78c-9b0e654d6fd6 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:00:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 460230b5-3593-4a25-becb-4c4727892a2e is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:00:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 01:00:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c587e61e-d18f-4e39-9404-027cd3697185 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:00:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9cb83792-8f4e-4574-a78c-9b0e654d6fd6 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:00:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 460230b5-3593-4a25-becb-4c4727892a2e is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:00:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 01:00:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c587e61e-d18f-4e39-9404-027cd3697185 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:00:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9cb83792-8f4e-4574-a78c-9b0e654d6fd6 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:00:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 460230b5-3593-4a25-becb-4c4727892a2e is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:00:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 01:00:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c587e61e-d18f-4e39-9404-027cd3697185 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:00:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9cb83792-8f4e-4574-a78c-9b0e654d6fd6 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:00:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 460230b5-3593-4a25-becb-4c4727892a2e is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:00:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 01:00:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c587e61e-d18f-4e39-9404-027cd3697185 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:00:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9cb83792-8f4e-4574-a78c-9b0e654d6fd6 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:00:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 460230b5-3593-4a25-becb-4c4727892a2e is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:00:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 01:00:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c587e61e-d18f-4e39-9404-027cd3697185 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:00:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9cb83792-8f4e-4574-a78c-9b0e654d6fd6 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:00:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 460230b5-3593-4a25-becb-4c4727892a2e is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:00:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 01:00:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c587e61e-d18f-4e39-9404-027cd3697185 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:00:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9cb83792-8f4e-4574-a78c-9b0e654d6fd6 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:00:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 460230b5-3593-4a25-becb-4c4727892a2e is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:00:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 01:00:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c587e61e-d18f-4e39-9404-027cd3697185 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:00:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9cb83792-8f4e-4574-a78c-9b0e654d6fd6 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:00:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 460230b5-3593-4a25-becb-4c4727892a2e is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:00:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 01:00:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c587e61e-d18f-4e39-9404-027cd3697185 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:00:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9cb83792-8f4e-4574-a78c-9b0e654d6fd6 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:00:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 460230b5-3593-4a25-becb-4c4727892a2e is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:00:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 01:00:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c587e61e-d18f-4e39-9404-027cd3697185 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:00:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9cb83792-8f4e-4574-a78c-9b0e654d6fd6 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:00:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 460230b5-3593-4a25-becb-4c4727892a2e is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:00:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 01:00:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c587e61e-d18f-4e39-9404-027cd3697185 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:00:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9cb83792-8f4e-4574-a78c-9b0e654d6fd6 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:00:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 460230b5-3593-4a25-becb-4c4727892a2e is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:00:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 01:00:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c587e61e-d18f-4e39-9404-027cd3697185 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:00:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9cb83792-8f4e-4574-a78c-9b0e654d6fd6 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:00:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 460230b5-3593-4a25-becb-4c4727892a2e is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:00:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 01:00:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c587e61e-d18f-4e39-9404-027cd3697185 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:00:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9cb83792-8f4e-4574-a78c-9b0e654d6fd6 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:00:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 460230b5-3593-4a25-becb-4c4727892a2e is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:00:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 01:00:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c587e61e-d18f-4e39-9404-027cd3697185 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:00:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9cb83792-8f4e-4574-a78c-9b0e654d6fd6 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:00:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 460230b5-3593-4a25-becb-4c4727892a2e is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:00:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 01:00:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c587e61e-d18f-4e39-9404-027cd3697185 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:00:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9cb83792-8f4e-4574-a78c-9b0e654d6fd6 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:00:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 460230b5-3593-4a25-becb-4c4727892a2e is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:00:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 01:00:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c587e61e-d18f-4e39-9404-027cd3697185 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:00:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9cb83792-8f4e-4574-a78c-9b0e654d6fd6 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:00:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 460230b5-3593-4a25-becb-4c4727892a2e is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:00:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 01:00:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c587e61e-d18f-4e39-9404-027cd3697185 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:00:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9cb83792-8f4e-4574-a78c-9b0e654d6fd6 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:00:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 460230b5-3593-4a25-becb-4c4727892a2e is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:00:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 01:00:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c587e61e-d18f-4e39-9404-027cd3697185 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:00:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9cb83792-8f4e-4574-a78c-9b0e654d6fd6 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:00:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 460230b5-3593-4a25-becb-4c4727892a2e is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:00:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 01:01:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c587e61e-d18f-4e39-9404-027cd3697185 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:01:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9cb83792-8f4e-4574-a78c-9b0e654d6fd6 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:01:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 460230b5-3593-4a25-becb-4c4727892a2e is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:01:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 01:01:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c587e61e-d18f-4e39-9404-027cd3697185 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:01:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9cb83792-8f4e-4574-a78c-9b0e654d6fd6 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:01:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 460230b5-3593-4a25-becb-4c4727892a2e is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:01:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 01:01:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c5f0c529-69aa-4268-97af-bde1796abaa2 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:01:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c587e61e-d18f-4e39-9404-027cd3697185 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:01:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9cb83792-8f4e-4574-a78c-9b0e654d6fd6 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:01:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 460230b5-3593-4a25-becb-4c4727892a2e is in state SUCCESS\u001b[0m\n\n\u001b[1;35m[WARNING]: Collection community.general does not support Ansible version\u001b[0m\n\u001b[1;35m2.16.14\u001b[0m\n\nPLAY [Create ceph pools] *******************************************************\n\nTASK [ceph-facts : Include facts.yml] ******************************************\nTuesday 24 February 2026  00:58:46 +0000 (0:00:00.613)       0:00:00.613 ****** \n\u001b[0;36mincluded: /ansible/roles/ceph-facts/tasks/facts.yml for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m\n\nTASK [ceph-facts : Check if it is atomic host] *********************************\nTuesday 24 February 2026  00:58:47 +0000 (0:00:00.684)       0:00:01.297 ****** \n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\nTASK [ceph-facts : Set_fact is_atomic] *****************************************\nTuesday 24 February 2026  00:58:47 +0000 (0:00:00.646)       0:00:01.944 ****** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Check if podman binary is present] **************************\nTuesday 24 February 2026  00:58:48 +0000 (0:00:00.358)       0:00:02.303 ****** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Set_fact container_binary] **********************************\nTuesday 24 February 2026  00:58:49 +0000 (0:00:00.988)       0:00:03.292 ****** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Set_fact ceph_cmd] ******************************************\nTuesday 24 February 2026  00:58:49 +0000 (0:00:00.362)       0:00:03.655 ****** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Set_fact discovered_interpreter_python] *********************\nTuesday 24 February 2026  00:58:50 +0000 (0:00:00.354)       0:00:04.009 ****** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Set_fact discovered_interpreter_python if not previously set] ***\nTuesday 24 February 2026  00:58:50 +0000 (0:00:00.374)       0:00:04.384 ****** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Set_fact ceph_release ceph_stable_release] ******************\nTuesday 24 February 2026  00:58:51 +0000 (0:00:00.604)       0:00:04.988 ****** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Set_fact monitor_name ansible_facts['hostname']] ************\nTuesday 24 February 2026  00:58:51 +0000 (0:00:00.348)       0:00:05.337 ****** \n\u001b[0;32mok: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item=testbed-node-0)\u001b[0m\n\u001b[0;32mok: [testbed-node-3 -> testbed-node-1(192.168.16.11)] => (item=testbed-node-1)\u001b[0m\n\u001b[0;32mok: [testbed-node-3 -> testbed-node-2(192.168.16.12)] => (item=testbed-node-2)\u001b[0m\n\nTASK [ceph-facts : Set_fact container_exec_cmd] ********************************\nTuesday 24 February 2026  00:58:52 +0000 (0:00:00.714)       0:00:06.051 ****** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Find a running mon container] *******************************\nTuesday 24 February 2026  00:58:52 +0000 (0:00:00.494)       0:00:06.546 ****** \n\u001b[0;32mok: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item=testbed-node-0)\u001b[0m\n\u001b[0;32mok: [testbed-node-3 -> testbed-node-1(192.168.16.11)] => (item=testbed-node-1)\u001b[0m\n\u001b[0;32mok: [testbed-node-3 -> testbed-node-2(192.168.16.12)] => (item=testbed-node-2)\u001b[0m\n\nTASK [ceph-facts : Check for a ceph mon socket] ********************************\nTuesday 24 February 2026  00:58:54 +0000 (0:00:02.319)       0:00:08.866 ****** \n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-0) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-1) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-facts : Check if the ceph mon socket is in-use] *********************\nTuesday 24 February 2026  00:58:55 +0000 (0:00:00.895)       0:00:09.762 ****** \n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-0', 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-1', 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-2', 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-facts : Set_fact running_mon - non_container] ***********************\nTuesday 24 February 2026  00:58:56 +0000 (0:00:00.874)       0:00:10.637 ****** \n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': {'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-0', 'ansible_loop_var': 'item'}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': {'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-1', 'ansible_loop_var': 'item'}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': {'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-2', 'ansible_loop_var': 'item'}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-facts : Set_fact running_mon - container] ***************************\nTuesday 24 February 2026  00:58:57 +0000 (0:00:00.402)       0:00:11.039 ****** \n\u001b[0;32mok: [testbed-node-3] => (item={'changed': False, 'stdout': '1cb145726d67', 'stderr': '', 'rc': 0, 'cmd': ['docker', 'ps', '-q', '--filter', 'name=ceph-mon-testbed-node-0'], 'start': '2026-02-24 00:58:53.303404', 'end': '2026-02-24 00:58:53.343080', 'delta': '0:00:00.039676', 'msg': '', 'invocation': {'module_args': {'_raw_params': 'docker ps -q --filter name=ceph-mon-testbed-node-0', '_uses_shell': False, 'expand_argument_vars': True, 'stdin_add_newline': True, 'strip_empty_ends': True, 'argv': None, 'chdir': None, 'executable': None, 'creates': None, 'removes': None, 'stdin': None}}, 'stdout_lines': ['1cb145726d67'], 'stderr_lines': [], 'failed': False, 'failed_when_result': False, 'item': 'testbed-node-0', 'ansible_loop_var': 'item'})\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item={'changed': False, 'stdout': '65c197c778ba', 'stderr': '', 'rc': 0, 'cmd': ['docker', 'ps', '-q', '--filter', 'name=ceph-mon-testbed-node-1'], 'start': '2026-02-24 00:58:54.123679', 'end': '2026-02-24 00:58:54.173430', 'delta': '0:00:00.049751', 'msg': '', 'invocation': {'module_args': {'_raw_params': 'docker ps -q --filter name=ceph-mon-testbed-node-1', '_uses_shell': False, 'expand_argument_vars': True, 'stdin_add_newline': True, 'strip_empty_ends': True, 'argv': None, 'chdir': None, 'executable': None, 'creates': None, 'removes': None, 'stdin': None}}, 'stdout_lines': ['65c197c778ba'], 'stderr_lines': [], 'failed': False, 'failed_when_result': False, 'item': 'testbed-node-1', 'ansible_loop_var': 'item'})\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item={'changed': False, 'stdout': '839d7a58ab74', 'stderr': '', 'rc': 0, 'cmd': ['docker', 'ps', '-q', '--filter', 'name=ceph-mon-testbed-node-2'], 'start': '2026-02-24 00:58:54.681655', 'end': '2026-02-24 00:58:54.723543', 'delta': '0:00:00.041888', 'msg': '', 'invocation': {'module_args': {'_raw_params': 'docker ps -q --filter name=ceph-mon-testbed-node-2', '_uses_shell': False, 'expand_argument_vars': True, 'stdin_add_newline': True, 'strip_empty_ends': True, 'argv': None, 'chdir': None, 'executable': None, 'creates': None, 'removes': None, 'stdin': None}}, 'stdout_lines': ['839d7a58ab74'], 'stderr_lines': [], 'failed': False, 'failed_when_result': False, 'item': 'testbed-node-2', 'ansible_loop_var': 'item'})\u001b[0m\n\nTASK [ceph-facts : Set_fact _container_exec_cmd] *******************************\nTuesday 24 February 2026  00:58:57 +0000 (0:00:00.199)       0:00:11.239 ****** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Get current fsid if cluster is already running] *************\nTuesday 24 February 2026  00:58:57 +0000 (0:00:00.490)       0:00:11.729 ****** \n\u001b[0;32mok: [testbed-node-3 -> testbed-node-2(192.168.16.12)]\u001b[0m\n\nTASK [ceph-facts : Set_fact current_fsid rc 1] *********************************\nTuesday 24 February 2026  00:58:59 +0000 (0:00:01.891)       0:00:13.621 ****** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Get current fsid] *******************************************\nTuesday 24 February 2026  00:58:59 +0000 (0:00:00.310)       0:00:13.931 ****** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Set_fact fsid] **********************************************\nTuesday 24 February 2026  00:59:00 +0000 (0:00:00.379)       0:00:14.310 ****** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Set_fact fsid from current_fsid] ****************************\nTuesday 24 February 2026  00:59:00 +0000 (0:00:00.442)       0:00:14.753 ****** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\nTASK [ceph-facts : Generate cluster fsid] **************************************\nTuesday 24 February 2026  00:59:00 +0000 (0:00:00.126)       0:00:14.880 ****** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-facts : Set_fact fsid] **********************************************\nTuesday 24 February 2026  00:59:01 +0000 (0:00:00.218)       0:00:15.098 ****** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Resolve device link(s)] *************************************\nTuesday 24 February 2026  00:59:01 +0000 (0:00:00.264)       0:00:15.363 ****** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Set_fact build devices from resolved symlinks] **************\nTuesday 24 February 2026  00:59:01 +0000 (0:00:00.283)       0:00:15.646 ****** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Resolve dedicated_device link(s)] ***************************\nTuesday 24 February 2026  00:59:02 +0000 (0:00:00.444)       0:00:16.091 ****** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Set_fact build dedicated_devices from resolved symlinks] ****\nTuesday 24 February 2026  00:59:02 +0000 (0:00:00.328)       0:00:16.419 ****** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Resolve bluestore_wal_device link(s)] ***********************\nTuesday 24 February 2026  00:59:02 +0000 (0:00:00.355)       0:00:16.774 ****** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Set_fact build bluestore_wal_devices from resolved symlinks] ***\nTuesday 24 February 2026  00:59:03 +0000 (0:00:00.331)       0:00:17.106 ****** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Collect existed devices] ************************************\nTuesday 24 February 2026  00:59:03 +0000 (0:00:00.438)       0:00:17.544 ****** \n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'dm-0', 'value': {'holders': [], 'host': '', 'links': {'ids': ['dm-name-ceph--9b92f24c--dac1--5bff--a390--34d9ff9e0f4f-osd--block--9b92f24c--dac1--5bff--a390--34d9ff9e0f4f', 'dm-uuid-LVM-h1BZIKSwq0ZvfBWtXhB9IpbOvYH7db3oH2r78k3xPmfToT2WBghjfHVanEend8Zd'], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': '', 'sectors': 41934848, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'dm-1', 'value': {'holders': [], 'host': '', 'links': {'ids': ['dm-name-ceph--7f6a8fe5--6adc--52b4--b256--edcdbfd9bf4e-osd--block--7f6a8fe5--6adc--52b4--b256--edcdbfd9bf4e', 'dm-uuid-LVM-xRh1zTNtAhB8OweVwVFtzaksaCJoA7fP27eNW4gBNJ6VNIfew4kSrXXLjM4ybDHa'], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': '', 'sectors': 41934848, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'loop0', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'loop1', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'loop2', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'loop3', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'loop4', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'loop5', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'loop6', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'loop7', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'sda', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_f90a4e80-1f36-4d14-b7da-de7b20e9fbca', 'scsi-SQEMU_QEMU_HARDDISK_f90a4e80-1f36-4d14-b7da-de7b20e9fbca'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {'sda1': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_f90a4e80-1f36-4d14-b7da-de7b20e9fbca-part1', 'scsi-SQEMU_QEMU_HARDDISK_f90a4e80-1f36-4d14-b7da-de7b20e9fbca-part1'], 'labels': ['cloudimg-rootfs'], 'masters': [], 'uuids': ['b852d8d2-8460-44aa-8998-23e4f04d73cf']}, 'sectors': 165672927, 'sectorsize': 512, 'size': '79.00 GB', 'start': '2099200', 'uuid': 'b852d8d2-8460-44aa-8998-23e4f04d73cf'}, 'sda14': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_f90a4e80-1f36-4d14-b7da-de7b20e9fbca-part14', 'scsi-SQEMU_QEMU_HARDDISK_f90a4e80-1f36-4d14-b7da-de7b20e9fbca-part14'], 'labels': [], 'masters': [], 'uuids': []}, 'sectors': 8192, 'sectorsize': 512, 'size': '4.00 MB', 'start': '2048', 'uuid': None}, 'sda15': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_f90a4e80-1f36-4d14-b7da-de7b20e9fbca-part15', 'scsi-SQEMU_QEMU_HARDDISK_f90a4e80-1f36-4d14-b7da-de7b20e9fbca-part15'], 'labels': ['UEFI'], 'masters': [], 'uuids': ['5C78-612A']}, 'sectors': 217088, 'sectorsize': 512, 'size': '106.00 MB', 'start': '10240', 'uuid': '5C78-612A'}, 'sda16': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_f90a4e80-1f36-4d14-b7da-de7b20e9fbca-part16', 'scsi-SQEMU_QEMU_HARDDISK_f90a4e80-1f36-4d14-b7da-de7b20e9fbca-part16'], 'labels': ['BOOT'], 'masters': [], 'uuids': ['09d53dc1-1e03-4286-bbb8-2b1796cf92ec']}, 'sectors': 1869825, 'sectorsize': 512, 'size': '913.00 MB', 'start': '227328', 'uuid': '09d53dc1-1e03-4286-bbb8-2b1796cf92ec'}}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 167772160, 'sectorsize': '512', 'size': '80.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'sdb', 'value': {'holders': ['ceph--9b92f24c--dac1--5bff--a390--34d9ff9e0f4f-osd--block--9b92f24c--dac1--5bff--a390--34d9ff9e0f4f'], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['lvm-pv-uuid-mec8nJ-xUip-qiWP-sj8w-uf5B-1Lt0-v8yfvH', 'scsi-0QEMU_QEMU_HARDDISK_9c6b6763-9704-41ec-9b6d-97eb80a5145c', 'scsi-SQEMU_QEMU_HARDDISK_9c6b6763-9704-41ec-9b6d-97eb80a5145c'], 'labels': [], 'masters': ['dm-0'], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'sdc', 'value': {'holders': ['ceph--7f6a8fe5--6adc--52b4--b256--edcdbfd9bf4e-osd--block--7f6a8fe5--6adc--52b4--b256--edcdbfd9bf4e'], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['lvm-pv-uuid-HdfFdZ-81SJ-0Umt-qAzb-Dao2-7tUe-tHuhnG', 'scsi-0QEMU_QEMU_HARDDISK_6644f191-69e9-4f3a-baff-cffe4152595b', 'scsi-SQEMU_QEMU_HARDDISK_6644f191-69e9-4f3a-baff-cffe4152595b'], 'labels': [], 'masters': ['dm-1'], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'sdd', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_4534650b-23a0-46c8-8bf1-44520caee637', 'scsi-SQEMU_QEMU_HARDDISK_4534650b-23a0-46c8-8bf1-44520caee637'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'dm-0', 'value': {'holders': [], 'host': '', 'links': {'ids': ['dm-name-ceph--294dd48f--9d6d--5079--8214--59e60af9e526-osd--block--294dd48f--9d6d--5079--8214--59e60af9e526', 'dm-uuid-LVM-wjENyDIL1ovSEUDphRiOfALWwweDrSZWE8niuN9gyF3lbbybFqRYsEaNkOnZAeUN'], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': '', 'sectors': 41934848, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'sr0', 'value': {'holders': [], 'host': 'IDE interface: Intel Corporation 82371SB PIIX3 IDE [Natoma/Triton II]', 'links': {'ids': ['ata-QEMU_DVD-ROM_QM00001'], 'labels': ['config-2'], 'masters': [], 'uuids': ['2026-02-24-00-01-58-00']}, 'model': 'QEMU DVD-ROM', 'partitions': {}, 'removable': '1', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'mq-deadline', 'sectors': 253, 'sectorsize': '2048', 'size': '506.00 KB', 'support_discard': '0', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'dm-1', 'value': {'holders': [], 'host': '', 'links': {'ids': ['dm-name-ceph--7fbf4d63--7c6f--5c70--887c--2b73c978775b-osd--block--7fbf4d63--7c6f--5c70--887c--2b73c978775b', 'dm-uuid-LVM-aRcKlwv7ZYsrJYaz5OO903mOAEBnwLXI61utfRQTUZQKHwm4ceJ3NApz8UrwnN8y'], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': '', 'sectors': 41934848, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'loop0', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'loop1', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'loop2', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'loop3', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'loop4', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'loop5', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'loop6', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'loop7', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'sda', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_47e600ec-05ca-44a6-93d0-5ca3104e3639', 'scsi-SQEMU_QEMU_HARDDISK_47e600ec-05ca-44a6-93d0-5ca3104e3639'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {'sda1': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_47e600ec-05ca-44a6-93d0-5ca3104e3639-part1', 'scsi-SQEMU_QEMU_HARDDISK_47e600ec-05ca-44a6-93d0-5ca3104e3639-part1'], 'labels': ['cloudimg-rootfs'], 'masters': [], 'uuids': ['b852d8d2-8460-44aa-8998-23e4f04d73cf']}, 'sectors': 165672927, 'sectorsize': 512, 'size': '79.00 GB', 'start': '2099200', 'uuid': 'b852d8d2-8460-44aa-8998-23e4f04d73cf'}, 'sda14': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_47e600ec-05ca-44a6-93d0-5ca3104e3639-part14', 'scsi-SQEMU_QEMU_HARDDISK_47e600ec-05ca-44a6-93d0-5ca3104e3639-part14'], 'labels': [], 'masters': [], 'uuids': []}, 'sectors': 8192, 'sectorsize': 512, 'size': '4.00 MB', 'start': '2048', 'uuid': None}, 'sda15': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_47e600ec-05ca-44a6-93d0-5ca3104e3639-part15', 'scsi-SQEMU_QEMU_HARDDISK_47e600ec-05ca-44a6-93d0-5ca3104e3639-part15'], 'labels': ['UEFI'], 'masters': [], 'uuids': ['5C78-612A']}, 'sectors': 217088, 'sectorsize': 512, 'size': '106.00 MB', 'start': '10240', 'uuid': '5C78-612A'}, 'sda16': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_47e600ec-05ca-44a6-93d0-5ca3104e3639-part16', 'scsi-SQEMU_QEMU_HARDDISK_47e600ec-05ca-44a6-93d0-5ca3104e3639-part16'], 'labels': ['BOOT'], 'masters': [], 'uuids': ['09d53dc1-1e03-4286-bbb8-2b1796cf92ec']}, 'sectors': 1869825, 'sectorsize': 512, 'size': '913.00 MB', 'start': '227328', 'uuid': '09d53dc1-1e03-4286-bbb8-2b1796cf92ec'}}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 167772160, 'sectorsize': '512', 'size': '80.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'sdb', 'value': {'holders': ['ceph--294dd48f--9d6d--5079--8214--59e60af9e526-osd--block--294dd48f--9d6d--5079--8214--59e60af9e526'], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['lvm-pv-uuid-1V6jIs-3p8q-l9Rl-OJ2n-8L4P-5xCz-K0YqYS', 'scsi-0QEMU_QEMU_HARDDISK_f659b61f-c39f-426f-a640-6a753951661d', 'scsi-SQEMU_QEMU_HARDDISK_f659b61f-c39f-426f-a640-6a753951661d'], 'labels': [], 'masters': ['dm-0'], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'sdc', 'value': {'holders': ['ceph--7fbf4d63--7c6f--5c70--887c--2b73c978775b-osd--block--7fbf4d63--7c6f--5c70--887c--2b73c978775b'], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['lvm-pv-uuid-aHCB2h-b0pT-WCWd-loOK-A09H-yFYp-FH2wIv', 'scsi-0QEMU_QEMU_HARDDISK_fb58bf16-a1c6-4abb-925b-29c6a6a6f100', 'scsi-SQEMU_QEMU_HARDDISK_fb58bf16-a1c6-4abb-925b-29c6a6a6f100'], 'labels': [], 'masters': ['dm-1'], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'sdd', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_5660bb62-e2c9-45b2-ae11-8ab0f829b194', 'scsi-SQEMU_QEMU_HARDDISK_5660bb62-e2c9-45b2-ae11-8ab0f829b194'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'sr0', 'value': {'holders': [], 'host': 'IDE interface: Intel Corporation 82371SB PIIX3 IDE [Natoma/Triton II]', 'links': {'ids': ['ata-QEMU_DVD-ROM_QM00001'], 'labels': ['config-2'], 'masters': [], 'uuids': ['2026-02-24-00-01-59-00']}, 'model': 'QEMU DVD-ROM', 'partitions': {}, 'removable': '1', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'mq-deadline', 'sectors': 253, 'sectorsize': '2048', 'size': '506.00 KB', 'support_discard': '0', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'dm-0', 'value': {'holders': [], 'host': '', 'links': {'ids': ['dm-name-ceph--752a9ec4--4bfc--5551--82cd--92b802bdad97-osd--block--752a9ec4--4bfc--5551--82cd--92b802bdad97', 'dm-uuid-LVM-jTt9fg6XlhiXbWhW0I9T02tzYsIeoKIoaf5BxtnvIAUk6EWNKNJMqSjVsZqTXHAa'], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': '', 'sectors': 41934848, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'dm-1', 'value': {'holders': [], 'host': '', 'links': {'ids': ['dm-name-ceph--ba85a0e3--2f57--51d5--a9f1--072f0e9a8deb-osd--block--ba85a0e3--2f57--51d5--a9f1--072f0e9a8deb', 'dm-uuid-LVM-YjmBzAWpiQhdCbdayUM0w3sqNqanvbLDieybFDy5wWcdDcB5LPRFHjaRE0WCOjBW'], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational'\u001b[32m2026-02-24 01:01:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n: '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': '', 'sectors': 41934848, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'loop0', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'loop1', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'loop2', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'loop3', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'loop4', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'loop5', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'loop6', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'loop7', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'sda', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_7a9487cb-ec9e-4508-9c5a-f692700fc5a4', 'scsi-SQEMU_QEMU_HARDDISK_7a9487cb-ec9e-4508-9c5a-f692700fc5a4'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {'sda1': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_7a9487cb-ec9e-4508-9c5a-f692700fc5a4-part1', 'scsi-SQEMU_QEMU_HARDDISK_7a9487cb-ec9e-4508-9c5a-f692700fc5a4-part1'], 'labels': ['cloudimg-rootfs'], 'masters': [], 'uuids': ['b852d8d2-8460-44aa-8998-23e4f04d73cf']}, 'sectors': 165672927, 'sectorsize': 512, 'size': '79.00 GB', 'start': '2099200', 'uuid': 'b852d8d2-8460-44aa-8998-23e4f04d73cf'}, 'sda14': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_7a9487cb-ec9e-4508-9c5a-f692700fc5a4-part14', 'scsi-SQEMU_QEMU_HARDDISK_7a9487cb-ec9e-4508-9c5a-f692700fc5a4-part14'], 'labels': [], 'masters': [], 'uuids': []}, 'sectors': 8192, 'sectorsize': 512, 'size': '4.00 MB', 'start': '2048', 'uuid': None}, 'sda15': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_7a9487cb-ec9e-4508-9c5a-f692700fc5a4-part15', 'scsi-SQEMU_QEMU_HARDDISK_7a9487cb-ec9e-4508-9c5a-f692700fc5a4-part15'], 'labels': ['UEFI'], 'masters': [], 'uuids': ['5C78-612A']}, 'sectors': 217088, 'sectorsize': 512, 'size': '106.00 MB', 'start': '10240', 'uuid': '5C78-612A'}, 'sda16': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_7a9487cb-ec9e-4508-9c5a-f692700fc5a4-part16', 'scsi-SQEMU_QEMU_HARDDISK_7a9487cb-ec9e-4508-9c5a-f692700fc5a4-part16'], 'labels': ['BOOT'], 'masters': [], 'uuids': ['09d53dc1-1e03-4286-bbb8-2b1796cf92ec']}, 'sectors': 1869825, 'sectorsize': 512, 'size': '913.00 MB', 'start': '227328', 'uuid': '09d53dc1-1e03-4286-bbb8-2b1796cf92ec'}}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 167772160, 'sectorsize': '512', 'size': '80.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'sdb', 'value': {'holders': ['ceph--752a9ec4--4bfc--5551--82cd--92b802bdad97-osd--block--752a9ec4--4bfc--5551--82cd--92b802bdad97'], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['lvm-pv-uuid-vfwhPS-pgoq-FNQw-W72Z-epjC-Ziiv-qWNph9', 'scsi-0QEMU_QEMU_HARDDISK_1558935b-4c66-4614-97e5-cda14abd2641', 'scsi-SQEMU_QEMU_HARDDISK_1558935b-4c66-4614-97e5-cda14abd2641'], 'labels': [], 'masters': ['dm-0'], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'sdc', 'value': {'holders': ['ceph--ba85a0e3--2f57--51d5--a9f1--072f0e9a8deb-osd--block--ba85a0e3--2f57--51d5--a9f1--072f0e9a8deb'], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['lvm-pv-uuid-XuBysF-MTWF-hP8t-3CGS-0PEc-QL0q-u9CYq2', 'scsi-0QEMU_QEMU_HARDDISK_6f4643f0-329b-4ae2-97c7-d348648fd970', 'scsi-SQEMU_QEMU_HARDDISK_6f4643f0-329b-4ae2-97c7-d348648fd970'], 'labels': [], 'masters': ['dm-1'], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'sdd', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_5f47789f-2a1a-442a-8a3c-644a4f92e5a8', 'scsi-SQEMU_QEMU_HARDDISK_5f47789f-2a1a-442a-8a3c-644a4f92e5a8'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'sr0', 'value': {'holders': [], 'host': 'IDE interface: Intel Corporation 82371SB PIIX3 IDE [Natoma/Triton II]', 'links': {'ids': ['ata-QEMU_DVD-ROM_QM00001'], 'labels': ['config-2'], 'masters': [], 'uuids': ['2026-02-24-00-02-03-00']}, 'model': 'QEMU DVD-ROM', 'partitions': {}, 'removable': '1', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'mq-deadline', 'sectors': 253, 'sectorsize': '2048', 'size': '506.00 KB', 'support_discard': '0', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Set_fact devices generate device list when osd_auto_discovery] ***\nTuesday 24 February 2026  00:59:04 +0000 (0:00:00.557)       0:00:18.102 ****** \n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'dm-0', 'value': {'holders': [], 'host': '', 'links': {'ids': ['dm-name-ceph--9b92f24c--dac1--5bff--a390--34d9ff9e0f4f-osd--block--9b92f24c--dac1--5bff--a390--34d9ff9e0f4f', 'dm-uuid-LVM-h1BZIKSwq0ZvfBWtXhB9IpbOvYH7db3oH2r78k3xPmfToT2WBghjfHVanEend8Zd'], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': '', 'sectors': 41934848, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'dm-1', 'value': {'holders': [], 'host': '', 'links': {'ids': ['dm-name-ceph--7f6a8fe5--6adc--52b4--b256--edcdbfd9bf4e-osd--block--7f6a8fe5--6adc--52b4--b256--edcdbfd9bf4e', 'dm-uuid-LVM-xRh1zTNtAhB8OweVwVFtzaksaCJoA7fP27eNW4gBNJ6VNIfew4kSrXXLjM4ybDHa'], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': '', 'sectors': 41934848, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop0', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop1', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop2', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop3', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop4', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop5', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop6', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'dm-0', 'value': {'holders': [], 'host': '', 'links': {'ids': ['dm-name-ceph--294dd48f--9d6d--5079--8214--59e60af9e526-osd--block--294dd48f--9d6d--5079--8214--59e60af9e526', 'dm-uuid-LVM-wjENyDIL1ovSEUDphRiOfALWwweDrSZWE8niuN9gyF3lbbybFqRYsEaNkOnZAeUN'], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': '', 'sectors': 41934848, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop7', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'dm-1', 'value': {'holders': [], 'host': '', 'links': {'ids': ['dm-name-ceph--7fbf4d63--7c6f--5c70--887c--2b73c978775b-osd--block--7fbf4d63--7c6f--5c70--887c--2b73c978775b', 'dm-uuid-LVM-aRcKlwv7ZYsrJYaz5OO903mOAEBnwLXI61utfRQTUZQKHwm4ceJ3NApz8UrwnN8y'], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': '', 'sectors': 41934848, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sda', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_f90a4e80-1f36-4d14-b7da-de7b20e9fbca', 'scsi-SQEMU_QEMU_HARDDISK_f90a4e80-1f36-4d14-b7da-de7b20e9fbca'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {'sda1': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_f90a4e80-1f36-4d14-b7da-de7b20e9fbca-part1', 'scsi-SQEMU_QEMU_HARDDISK_f90a4e80-1f36-4d14-b7da-de7b20e9fbca-part1'], 'labels': ['cloudimg-rootfs'], 'masters': [], 'uuids': ['b852d8d2-8460-44aa-8998-23e4f04d73cf']}, 'sectors': 165672927, 'sectorsize': 512, 'size': '79.00 GB', 'start': '2099200', 'uuid': 'b852d8d2-8460-44aa-8998-23e4f04d73cf'}, 'sda14': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_f90a4e80-1f36-4d14-b7da-de7b20e9fbca-part14', 'scsi-SQEMU_QEMU_HARDDISK_f90a4e80-1f36-4d14-b7da-de7b20e9fbca-part14'], 'labels': [], 'masters': [], 'uuids': []}, 'sectors': 8192, 'sectorsize': 512, 'size': '4.00 MB', 'start': '2048', 'uuid': None}, 'sda15': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_f90a4e80-1f36-4d14-b7da-de7b20e9fbca-part15', 'scsi-SQEMU_QEMU_HARDDISK_f90a4e80-1f36-4d14-b7da-de7b20e9fbca-part15'], 'labels': ['UEFI'], 'masters': [], 'uuids': ['5C78-612A']}, 'sectors': 217088, 'sectorsize': 512, 'size': '106.00 MB', 'start': '10240', 'uuid': '5C78-612A'}, 'sda16': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_f90a4e80-1f36-4d14-b7da-de7b20e9fbca-part16', 'scsi-SQEMU_QEMU_HARDDISK_f90a4e80-1f36-4d14-b7da-de7b20e9fbca-part16'], 'labels': ['BOOT'], 'masters': [], 'uuids': ['09d53dc1-1e03-4286-bbb8-2b1796cf92ec']}, 'sectors': 1869825, 'sectorsize': 512, 'size': '913.00 MB', 'start': '227328', 'uuid': '09d53dc1-1e03-4286-bbb8-2b1796cf92ec'}}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 167772160, 'sectorsize': '512', 'size': '80.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop0', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sdb', 'value': {'holders': ['ceph--9b92f24c--dac1--5bff--a390--34d9ff9e0f4f-osd--block--9b92f24c--dac1--5bff--a390--34d9ff9e0f4f'], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['lvm-pv-uuid-mec8nJ-xUip-qiWP-sj8w-uf5B-1Lt0-v8yfvH', 'scsi-0QEMU_QEMU_HARDDISK_9c6b6763-9704-41ec-9b6d-97eb80a5145c', 'scsi-SQEMU_QEMU_HARDDISK_9c6b6763-9704-41ec-9b6d-97eb80a5145c'], 'labels': [], 'masters': ['dm-0'], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop1', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sdc', 'value': {'holders': ['ceph--7f6a8fe5--6adc--52b4--b256--edcdbfd9bf4e-osd--block--7f6a8fe5--6adc--52b4--b256--edcdbfd9bf4e'], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['lvm-pv-uuid-HdfFdZ-81SJ-0Umt-qAzb-Dao2-7tUe-tHuhnG', 'scsi-0QEMU_QEMU_HARDDISK_6644f191-69e9-4f3a-baff-cffe4152595b', 'scsi-SQEMU_QEMU_HARDDISK_6644f191-69e9-4f3a-baff-cffe4152595b'], 'labels': [], 'masters': ['dm-1'], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sdd', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_4534650b-23a0-46c8-8bf1-44520caee637', 'scsi-SQEMU_QEMU_HARDDISK_4534650b-23a0-46c8-8bf1-44520caee637'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop2', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sr0', 'value': {'holders': [], 'host': 'IDE interface: Intel Corporation 82371SB PIIX3 IDE [Natoma/Triton II]', 'links': {'ids': ['ata-QEMU_DVD-ROM_QM00001'], 'labels': ['config-2'], 'masters': [], 'uuids': ['2026-02-24-00-01-58-00']}, 'model': 'QEMU DVD-ROM', 'partitions': {}, 'removable': '1', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'mq-deadline', 'sectors': 253, 'sectorsize': '2048', 'size': '506.00 KB', 'support_discard': '0', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop3', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop4', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop5', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'dm-0', 'value': {'holders': [], 'host': '', 'links': {'ids': ['dm-name-ceph--752a9ec4--4bfc--5551--82cd--92b802bdad97-osd--block--752a9ec4--4bfc--5551--82cd--92b802bdad97', 'dm-uuid-LVM-jTt9fg6XlhiXbWhW0I9T02tzYsIeoKIoaf5BxtnvIAUk6EWNKNJMqSjVsZqTXHAa'], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': '', 'sectors': 41934848, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop6', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'dm-1', 'value': {'holders': [], 'host': '', 'links': {'ids': ['dm-name-ceph--ba85a0e3--2f57--51d5--a9f1--072f0e9a8deb-osd--block--ba85a0e3--2f57--51d5--a9f1--072f0e9a8deb', 'dm-uuid-LVM-YjmBzAWpiQhdCbdayUM0w3sqNqanvbLDieybFDy5wWcdDcB5LPRFHjaRE0WCOjBW'], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': '', 'sectors': 41934848, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop7', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop0', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sda', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_47e600ec-05ca-44a6-93d0-5ca3104e3639', 'scsi-SQEMU_QEMU_HARDDISK_47e600ec-05ca-44a6-93d0-5ca3104e3639'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {'sda1': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_47e600ec-05ca-44a6-93d0-5ca3104e3639-part1', 'scsi-SQEMU_QEMU_HARDDISK_47e600ec-05ca-44a6-93d0-5ca3104e3639-part1'], 'labels': ['cloudimg-rootfs'], 'masters': [], 'uuids': ['b852d8d2-8460-44aa-8998-23e4f04d73cf']}, 'sectors': 165672927, 'sectorsize': 512, 'size': '79.00 GB', 'start': '2099200', 'uuid': 'b852d8d2-8460-44aa-8998-23e4f04d73cf'}, 'sda14': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_47e600ec-05ca-44a6-93d0-5ca3104e3639-part14', 'scsi-SQEMU_QEMU_HARDDISK_47e600ec-05ca-44a6-93d0-5ca3104e3639-part14'], 'labels': [], 'masters': [], 'uuids': []}, 'sectors': 8192, 'sectorsize': 512, 'size': '4.00 MB', 'start': '2048', 'uuid': None}, 'sda15': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_47e600ec-05ca-44a6-93d0-5ca3104e3639-part15', 'scsi-SQEMU_QEMU_HARDDISK_47e600ec-05ca-44a6-93d0-5ca3104e3639-part15'], 'labels': ['UEFI'], 'masters': [], 'uuids': ['5C78-612A']}, 'sectors': 217088, 'sectorsize': 512, 'size': '106.00 MB', 'start': '10240', 'uuid': '5C78-612A'}, 'sda16': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_47e600ec-05ca-44a6-93d0-5ca3104e3639-part16', 'scsi-SQEMU_QEMU_HARDDISK_47e600ec-05ca-44a6-93d0-5ca3104e3639-part16'], 'labels': ['BOOT'], 'masters': [], 'uuids': ['09d53dc1-1e03-4286-bbb8-2b1796cf92ec']}, 'sectors': 1869825, 'sectorsize': 512, 'size': '913.00 MB', 'start': '227328', 'uuid': '09d53dc1-1e03-4286-bbb8-2b1796cf92ec'}}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 167772160, 'sectorsize': '512', 'size': '80.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop1', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sdb', 'value': {'holders': ['ceph--294dd48f--9d6d--5079--8214--59e60af9e526-osd--block--294dd48f--9d6d--5079--8214--59e60af9e526'], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['lvm-pv-uuid-1V6jIs-3p8q-l9Rl-OJ2n-8L4P-5xCz-K0YqYS', 'scsi-0QEMU_QEMU_HARDDISK_f659b61f-c39f-426f-a640-6a753951661d', 'scsi-SQEMU_QEMU_HARDDISK_f659b61f-c39f-426f-a640-6a753951661d'], 'labels': [], 'masters': ['dm-0'], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop2', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sdc', 'value': {'holders': ['ceph--7fbf4d63--7c6f--5c70--887c--2b73c978775b-osd--block--7fbf4d63--7c6f--5c70--887c--2b73c978775b'], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['lvm-pv-uuid-aHCB2h-b0pT-WCWd-loOK-A09H-yFYp-FH2wIv', 'scsi-0QEMU_QEMU_HARDDISK_fb58bf16-a1c6-4abb-925b-29c6a6a6f100', 'scsi-SQEMU_QEMU_HARDDISK_fb58bf16-a1c6-4abb-925b-29c6a6a6f100'], 'labels': [], 'masters': ['dm-1'], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop3', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop4', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sdd', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_5660bb62-e2c9-45b2-ae11-8ab0f829b194', 'scsi-SQEMU_QEMU_HARDDISK_5660bb62-e2c9-45b2-ae11-8ab0f829b194'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop5', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sr0', 'value': {'holders': [], 'host': 'IDE interface: Intel Corporation 82371SB PIIX3 IDE [Natoma/Triton II]', 'links': {'ids': ['ata-QEMU_DVD-ROM_QM00001'], 'labels': ['config-2'], 'masters': [], 'uuids': ['2026-02-24-00-01-59-00']}, 'model': 'QEMU DVD-ROM', 'partitions': {}, 'removable': '1', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'mq-deadline', 'sectors': 253, 'sectorsize': '2048', 'size': '506.00 KB', 'support_discard': '0', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop6', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop7', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sda', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_7a9487cb-ec9e-4508-9c5a-f692700fc5a4', 'scsi-SQEMU_QEMU_HARDDISK_7a9487cb-ec9e-4508-9c5a-f692700fc5a4'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {'sda1': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_7a9487cb-ec9e-4508-9c5a-f692700fc5a4-part1', 'scsi-SQEMU_QEMU_HARDDISK_7a9487cb-ec9e-4508-9c5a-f692700fc5a4-part1'], 'labels': ['cloudimg-rootfs'], 'masters': [], 'uuids': ['b852d8d2-8460-44aa-8998-23e4f04d73cf']}, 'sectors': 165672927, 'sectorsize': 512, 'size': '79.00 GB', 'start': '2099200', 'uuid': 'b852d8d2-8460-44aa-8998-23e4f04d73cf'}, 'sda14': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_7a9487cb-ec9e-4508-9c5a-f692700fc5a4-part14', 'scsi-SQEMU_QEMU_HARDDISK_7a9487cb-ec9e-4508-9c5a-f692700fc5a4-part14'], 'labels': [], 'masters': [], 'uuids': []}, 'sectors': 8192, 'sectorsize': 512, 'size': '4.00 MB', 'start': '2048', 'uuid': None}, 'sda15': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_7a9487cb-ec9e-4508-9c5a-f692700fc5a4-part15', 'scsi-SQEMU_QEMU_HARDDISK_7a9487cb-ec9e-4508-9c5a-f692700fc5a4-part15'], 'labels': ['UEFI'], 'masters': [], 'uuids': ['5C78-612A']}, 'sectors': 217088, 'sectorsize': 512, 'size': '106.00 MB', 'start': '10240', 'uuid': '5C78-612A'}, 'sda16': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_7a9487cb-ec9e-4508-9c5a-f692700fc5a4-part16', 'scsi-SQEMU_QEMU_HARDDISK_7a9487cb-ec9e-4508-9c5a-f692700fc5a4-part16'], 'labels': ['BOOT'], 'masters': [], 'uuids': ['09d53dc1-1e03-4286-bbb8-2b1796cf92ec']}, 'sectors': 1869825, 'sectorsize': 512, 'size': '913.00 MB', 'start': '227328', 'uuid': '09d53dc1-1e03-4286-bbb8-2b1796cf92ec'}}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 167772160, 'sectorsize': '512', 'size': '80.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sdb', 'value': {'holders': ['ceph--752a9ec4--4bfc--5551--82cd--92b802bdad97-osd--block--752a9ec4--4bfc--5551--82cd--92b802bdad97'], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['lvm-pv-uuid-vfwhPS-pgoq-FNQw-W72Z-epjC-Ziiv-qWNph9', 'scsi-0QEMU_QEMU_HARDDISK_1558935b-4c66-4614-97e5-cda14abd2641', 'scsi-SQEMU_QEMU_HARDDISK_1558935b-4c66-4614-97e5-cda14abd2641'], 'labels': [], 'masters': ['dm-0'], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sdc', 'value': {'holders': ['ceph--ba85a0e3--2f57--51d5--a9f1--072f0e9a8deb-osd--block--ba85a0e3--2f57--51d5--a9f1--072f0e9a8deb'], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['lvm-pv-uuid-XuBysF-MTWF-hP8t-3CGS-0PEc-QL0q-u9CYq2', 'scsi-0QEMU_QEMU_HARDDISK_6f4643f0-329b-4ae2-97c7-d348648fd970', 'scsi-SQEMU_QEMU_HARDDISK_6f4643f0-329b-4ae2-97c7-d348648fd970'], 'labels': [], 'masters': ['dm-1'], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sdd', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_5f47789f-2a1a-442a-8a3c-644a4f92e5a8', 'scsi-SQEMU_QEMU_HARDDISK_5f47789f-2a1a-442a-8a3c-644a4f92e5a8'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sr0', 'value': {'holders': [], 'host': 'IDE interface: Intel Corporation 82371SB PIIX3 IDE [Natoma/Triton II]', 'links': {'ids': ['ata-QEMU_DVD-ROM_QM00001'], 'labels': ['config-2'], 'masters': [], 'uuids': ['2026-02-24-00-02-03-00']}, 'model': 'QEMU DVD-ROM', 'partitions': {}, 'removable': '1', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'mq-deadline', 'sectors': 253, 'sectorsize': '2048', 'size': '506.00 KB', 'support_discard': '0', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Check if the ceph conf exists] ******************************\nTuesday 24 February 2026  00:59:04 +0000 (0:00:00.571)       0:00:18.674 ****** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Set default osd_pool_default_crush_rule fact] ***************\nTuesday 24 February 2026  00:59:05 +0000 (0:00:00.756)       0:00:19.430 ****** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Read osd pool default crush rule] ***************************\nTuesday 24 February 2026  00:59:05 +0000 (0:00:00.436)       0:00:19.866 ****** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\nTASK [ceph-facts : Set osd_pool_default_crush_rule fact] ***********************\nTuesday 24 February 2026  00:59:06 +0000 (0:00:00.652)       0:00:20.518 ****** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Read osd pool default crush rule] ***************************\nTuesday 24 February 2026  00:59:06 +0000 (0:00:00.293)       0:00:20.812 ****** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Set osd_pool_default_crush_rule fact] ***********************\nTuesday 24 February 2026  00:59:07 +0000 (0:00:00.391)       0:00:21.204 ****** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Set_fact _monitor_addresses - ipv4] *************************\nTuesday 24 February 2026  00:59:07 +0000 (0:00:00.490)       0:00:21.694 ****** \n\u001b[0;32mok: [testbed-node-3] => (item=testbed-node-0)\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item=testbed-node-0)\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item=testbed-node-1)\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item=testbed-node-0)\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item=testbed-node-1)\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item=testbed-node-2)\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item=testbed-node-1)\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item=testbed-node-2)\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item=testbed-node-2)\u001b[0m\n\nTASK [ceph-facts : Set_fact _monitor_addresses - ipv6] *************************\nTuesday 24 February 2026  00:59:08 +0000 (0:00:00.854)       0:00:22.549 ****** \n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-0) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-1) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=testbed-node-0) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=testbed-node-1) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=testbed-node-2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=testbed-node-0) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=testbed-node-1) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=testbed-node-2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Import_tasks set_radosgw_address.yml] ***********************\nTuesday 24 February 2026  00:59:08 +0000 (0:00:00.378)       0:00:22.927 ****** \n\u001b[0;36mincluded: /ansible/roles/ceph-facts/tasks/set_radosgw_address.yml for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m\n\nTASK [ceph-facts : Set current radosgw_address_block, radosgw_address, radosgw_interface  from node \"{{ ceph_dashboard_call_item }}\"] ***\nTuesday 24 February 2026  00:59:09 +0000 (0:00:00.672)       0:00:23.599 ****** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Set_fact _radosgw_address to radosgw_address_block ipv4] ****\nTuesday 24 February 2026  00:59:09 +0000 (0:00:00.322)       0:00:23.922 ****** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Set_fact _radosgw_address to radosgw_address_block ipv6] ****\nTuesday 24 February 2026  00:59:10 +0000 (0:00:00.365)       0:00:24.287 ****** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Set_fact _radosgw_address to radosgw_address] ***************\nTuesday 24 February 2026  00:59:10 +0000 (0:00:00.336)       0:00:24.623 ****** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Set_fact _interface] ****************************************\nTuesday 24 February 2026  00:59:11 +0000 (0:00:00.845)       0:00:25.468 ****** \n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-3) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-4) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-5) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-facts : Set_fact _radosgw_address to radosgw_interface - ipv4] ******\nTuesday 24 February 2026  00:59:11 +0000 (0:00:00.375)       0:00:25.844 ****** \n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-3) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-4) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-5) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-facts : Set_fact _radosgw_address to radosgw_interface - ipv6] ******\nTuesday 24 February 2026  00:59:12 +0000 (0:00:00.427)       0:00:26.272 ****** \n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-3) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-4) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-5) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [ceph-facts : Reset rgw_instances (workaround)] ***************************\nTuesday 24 February 2026  00:59:12 +0000 (0:00:00.369)       0:00:26.642 ****** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [ceph-facts : Set_fact rgw_instances] *************************************\nTuesday 24 February 2026  00:59:13 +0000 (0:00:00.332)       0:00:26.975 ****** \n\u001b[0;32mok: [testbed-node-3] => (item=0)\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item=0)\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item=0)\u001b[0m\n\nTASK [ceph-facts : Set_fact ceph_run_cmd] **************************************\nTuesday 24 February 2026  00:59:13 +0000 (0:00:00.465)       0:00:27.440 ****** \n\u001b[0;32mok: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item=testbed-node-0)\u001b[0m\n\u001b[0;32mok: [testbed-node-3 -> testbed-node-1(192.168.16.11)] => (item=testbed-node-1)\u001b[0m\n\u001b[0;32mok: [testbed-node-3 -> testbed-node-2(192.168.16.12)] => (item=testbed-node-2)\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item=testbed-node-3)\u001b[0m\n\u001b[0;32mok: [testbed-node-3 -> testbed-node-4(192.168.16.14)] => (item=testbed-node-4)\u001b[0m\n\u001b[0;32mok: [testbed-node-3 -> testbed-node-5(192.168.16.15)] => (item=testbed-node-5)\u001b[0m\n\u001b[0;32mok: [testbed-node-3 -> testbed-manager(192.168.16.5)] => (item=testbed-manager)\u001b[0m\n\nTASK [ceph-facts : Set_fact ceph_admin_command] ********************************\nTuesday 24 February 2026  00:59:14 +0000 (0:00:01.084)       0:00:28.525 ****** \n\u001b[0;32mok: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item=testbed-node-0)\u001b[0m\n\u001b[0;32mok: [testbed-node-3 -> testbed-node-1(192.168.16.11)] => (item=testbed-node-1)\u001b[0m\n\u001b[0;32mok: [testbed-node-3 -> testbed-node-2(192.168.16.12)] => (item=testbed-node-2)\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item=testbed-node-3)\u001b[0m\n\u001b[0;32mok: [testbed-node-3 -> testbed-node-4(192.168.16.14)] => (item=testbed-node-4)\u001b[0m\n\u001b[0;32mok: [testbed-node-3 -> testbed-node-5(192.168.16.15)] => (item=testbed-node-5)\u001b[0m\n\u001b[0;32mok: [testbed-node-3 -> testbed-manager(192.168.16.5)] => (item=testbed-manager)\u001b[0m\n\nTASK [Include tasks from the ceph-osd role] ************************************\nTuesday 24 February 2026  00:59:16 +0000 (0:00:02.041)       0:00:30.567 ****** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mincluded: /ansible/tasks/openstack_config.yml for testbed-node-5\u001b[0m\n\nTASK [create openstack pool(s)] ************************************************\nTuesday 24 February 2026  00:59:16 +0000 (0:00:00.368)       0:00:30.936 ****** \n\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item={'application': 'rbd', 'erasure_profile': '', 'expected_num_objects': '', 'min_size': 0, 'name': 'backups', 'pg_autoscale_mode': False, 'pg_num': 32, 'pgp_num': 32, 'rule_name': 'replicated_rule', 'size': 3, 'type': 1})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item={'application': 'rbd', 'erasure_profile': '', 'expected_num_objects': '', 'min_size': 0, 'name': 'volumes', 'pg_autoscale_mode': False, 'pg_num': 32, 'pgp_num': 32, 'rule_name': 'replicated_rule', 'size': 3, 'type': 1})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item={'application': 'rbd', 'erasure_profile': '', 'expected_num_objects': '', 'min_size': 0, 'name': 'images', 'pg_autoscale_mode': False, 'pg_num': 32, 'pgp_num': 32, 'rule_name': 'replicated_rule', 'size': 3, 'type': 1})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item={'application': 'rbd', 'erasure_profile': '', 'expected_num_objects': '', 'min_size': 0, 'name': 'metrics', 'pg_autoscale_mode': False, 'pg_num': 32, 'pgp_num': 32, 'rule_name': 'replicated_rule', 'size': 3, 'type': 1})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item={'application': 'rbd', 'erasure_profile': '', 'expected_num_objects': '', 'min_size': 0, 'name': 'vms', 'pg_autoscale_mode': False, 'pg_num': 32, 'pgp_num': 32, 'rule_name': 'replicated_rule', 'size': 3, 'type': 1})\u001b[0m\n\nTASK [generate keys] ***********************************************************\nTuesday 24 February 2026  01:00:05 +0000 (0:00:48.587)       0:01:19.524 ****** \n\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5 -> {{ groups[mon_group_name][0] }}]\u001b[0m\n\nTASK [get keys from monitors] **************************************************\nTuesday 24 February 2026  01:00:32 +0000 (0:00:27.173)       0:01:46.698 ****** \n\u001b[0;32mok: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m\n\u001b[0;32mok: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m\n\u001b[0;32mok: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m\n\u001b[0;32mok: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m\n\u001b[0;32mok: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m\n\u001b[0;32mok: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m\n\u001b[0;32mok: [testbed-node-5 -> {{ groups.get(mon_group_name)[0] }}]\u001b[0m\n\nTASK [copy ceph key(s) if needed] **********************************************\nTuesday 24 February 2026  01:00:46 +0000 (0:00:13.807)       0:02:00.505 ****** \n\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-1(192.168.16.11)] => (item=None)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-2(192.168.16.12)] => (item=None)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-1(192.168.16.11)] => (item=None)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-2(192.168.16.12)] => (item=None)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-1(192.168.16.11)] => (item=None)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-2(192.168.16.12)] => (item=None)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-1(192.168.16.11)] => (item=None)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-2(192.168.16.12)] => (item=None)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-1(192.168.16.11)] => (item=None)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-2(192.168.16.12)] => (item=None)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-1(192.168.16.11)] => (item=None)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-2(192.168.16.12)] => (item=None)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5 -> {{ item.1 }}]\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;32mtestbed-node-3\u001b[0m             : \u001b[0;32mok=25  \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=28  \u001b[0m rescued=0    ignored=0   \n\u001b[0;32mtestbed-node-4\u001b[0m             : \u001b[0;32mok=18  \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=21  \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-5\u001b[0m             : \u001b[0;32mok=23  \u001b[0m \u001b[0;33mchanged=3   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=20  \u001b[0m rescued=0    ignored=0   \n\n\n\nTASKS RECAP ********************************************************************\nTuesday 24 February 2026  01:01:05 +0000 (0:00:19.375)       0:02:19.881 ****** \n=============================================================================== \ncreate openstack pool(s) ----------------------------------------------- 48.59s\ngenerate keys ---------------------------------------------------------- 27.17s\ncopy ceph key(s) if needed --------------------------------------------- 19.38s\nget keys from monitors ------------------------------------------------- 13.81s\nceph-facts : Find a running mon container ------------------------------- 2.32s\nceph-facts : Set_fact ceph_admin_command -------------------------------- 2.04s\nceph-facts : Get current fsid if cluster is already running ------------- 1.89s\nceph-facts : Set_fact ceph_run_cmd -------------------------------------- 1.08s\nceph-facts : Check if podman binary is present -------------------------- 0.99s\nceph-facts : Check for a ceph mon socket -------------------------------- 0.90s\nceph-facts : Check if the ceph mon socket is in-use --------------------- 0.87s\nceph-facts : Set_fact _monitor_addresses - ipv4 ------------------------- 0.86s\nceph-facts : Set_fact _radosgw_address to radosgw_address --------------- 0.85s\nceph-facts : Check if the ceph conf exists ------------------------------ 0.76s\nceph-facts : Set_fact monitor_name ansible_facts['hostname'] ------------ 0.71s\nceph-facts : Include facts.yml ------------------------------------------ 0.68s\nceph-facts : Import_tasks set_radosgw_address.yml ----------------------- 0.67s\nceph-facts : Read osd pool default crush rule --------------------------- 0.65s\nceph-facts : Check if it is atomic host --------------------------------- 0.65s\nceph-facts : Set_fact discovered_interpreter_python if not previously set --- 0.60s\n\u001b[32m2026-02-24 01:01:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c5f0c529-69aa-4268-97af-bde1796abaa2 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:01:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c587e61e-d18f-4e39-9404-027cd3697185 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:01:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9cb83792-8f4e-4574-a78c-9b0e654d6fd6 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:01:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 01:01:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c5f0c529-69aa-4268-97af-bde1796abaa2 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:01:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c587e61e-d18f-4e39-9404-027cd3697185 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:01:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9cb83792-8f4e-4574-a78c-9b0e654d6fd6 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:01:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 01:01:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c5f0c529-69aa-4268-97af-bde1796abaa2 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:01:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c587e61e-d18f-4e39-9404-027cd3697185 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:01:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9cb83792-8f4e-4574-a78c-9b0e654d6fd6 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:01:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 01:01:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c5f0c529-69aa-4268-97af-bde1796abaa2 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:01:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c587e61e-d18f-4e39-9404-027cd3697185 is in state SUCCESS\u001b[0m\n\n\nPLAY [Group hosts based on configuration] **************************************\n\nTASK [Group hosts based on Kolla action] ***************************************\nTuesday 24 February 2026  00:59:34 +0000 (0:00:00.275)       0:00:00.275 ****** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [Group hosts based on enabled services] ***********************************\nTuesday 24 February 2026  00:59:35 +0000 (0:00:00.303)       0:00:00.578 ****** \n\u001b[0;32mok: [testbed-node-0] => (item=enable_horizon_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item=enable_horizon_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item=enable_horizon_True)\u001b[0m\n\nPLAY [Apply role horizon] ******************************************************\n\nTASK [horizon : include_tasks] *************************************************\nTuesday 24 February 2026  00:59:35 +0000 (0:00:00.399)       0:00:00.978 ****** \n\u001b[0;36mincluded: /ansible/roles/horizon/tasks/deploy.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [horizon : Ensuring config directories exist] *****************************\nTuesday 24 February 2026  00:59:36 +0000 (0:00:00.503)       0:00:01.481 ****** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'horizon', 'value': {'container_name': 'horizon', 'group': 'horizon', 'enabled': True, 'image': 'registry.osism.tech/kolla/horizon:2024.2', 'environment': {'ENABLE_BLAZAR': 'no', 'ENABLE_CLOUDKITTY': 'no', 'ENABLE_DESIGNATE': 'yes', 'ENABLE_FWAAS': 'no', 'ENABLE_HEAT': 'no', 'ENABLE_IRONIC': 'no', 'ENABLE_MAGNUM': 'yes', 'ENABLE_MANILA': 'yes', 'ENABLE_MASAKARI': 'no', 'ENABLE_MISTRAL': 'no', 'ENABLE_NEUTRON_VPNAAS': 'no', 'ENABLE_OCTAVIA': 'yes', 'ENABLE_TACKER': 'no', 'ENABLE_TROVE': 'no', 'ENABLE_WATCHER': 'no', 'ENABLE_ZUN': 'no', 'FORCE_GENERATE': 'no'}, 'volumes': ['/etc/kolla/horizon/:/var/lib/kolla/config_files/:ro', '', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:80'], 'timeout': '30'}, 'haproxy': {'horizon': {'enabled': True, 'mode': 'http', 'external': False, 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_redirect': {'enabled': True, 'mode': 'redirect', 'external': False, 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'horizon_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_external_redirect': {'enabled': True, 'mode': 'redirect', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'acme_client': {'enabled': True, 'with_frontend': False, 'custom_member_list': []}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'horizon', 'value': {'container_name': 'horizon', 'group': 'horizon', 'enabled': True, 'image': 'registry.osism.tech/kolla/horizon:2024.2', 'environment': {'ENABLE_BLAZAR': 'no', 'ENABLE_CLOUDKITTY': 'no', 'ENABLE_DESIGNATE': 'yes', 'ENABLE_FWAAS': 'no', 'ENABLE_HEAT': 'no', 'ENABLE_IRONIC': 'no', 'ENABLE_MAGNUM': 'yes', 'ENABLE_MANILA': 'yes', 'ENABLE_MASAKARI': 'no', 'ENABLE_MISTRAL': 'no', 'ENABLE_NEUTRON_VPNAAS': 'no', 'ENABLE_OCTAVIA': 'yes', 'ENABLE_TACKER': 'no', 'ENABLE_TROVE': 'no', 'ENABLE_WATCHER': 'no', 'ENABLE_ZUN': 'no', 'FORCE_GENERATE': 'no'}, 'volumes': ['/etc/kolla/horizon/:/var/lib/kolla/config_files/:ro', '', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:80'], 'timeout': '30'}, 'haproxy': {'horizon': {'enabled': True, 'mode': 'http', 'external': False, 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_redirect': {'enabled': True, 'mode': 'redirect', 'external': False, 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'horizon_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_external_redirect': {'enabled': True, 'mode': 'redirect', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'acme_client': {'enabled': True, 'with_frontend': False, 'custom_member_list': []}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'horizon', 'value': {'container_name': 'horizon', 'group': 'horizon', 'enabled': True, 'image': 'registry.osism.tech/kolla/horizon:2024.2', 'environment': {'ENABLE_BLAZAR': 'no', 'ENABLE_CLOUDKITTY': 'no', 'ENABLE_DESIGNATE': 'yes', 'ENABLE_FWAAS': 'no', 'ENABLE_HEAT': 'no', 'ENABLE_IRONIC': 'no', 'ENABLE_MAGNUM': 'yes', 'ENABLE_MANILA': 'yes', 'ENABLE_MASAKARI': 'no', 'ENABLE_MISTRAL': 'no', 'ENABLE_NEUTRON_VPNAAS': 'no', 'ENABLE_OCTAVIA': 'yes', 'ENABLE_TACKER': 'no', 'ENABLE_TROVE': 'no', 'ENABLE_WATCHER': 'no', 'ENABLE_ZUN': 'no', 'FORCE_GENERATE': 'no'}, 'volumes': ['/etc/kolla/horizon/:/var/lib/kolla/config_files/:ro', '', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:80'], 'timeout': '30'}, 'haproxy': {'horizon': {'enabled': True, 'mode': 'http', 'external': False, 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_redirect': {'enabled': True, 'mode': 'redirect', 'external': False, 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'horizon_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_external_redirect': {'enabled': True, 'mode': 'redirect', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'acme_client': {'enabled': True, 'with_frontend': False, 'custom_member_list': []}}}})\u001b[0m\n\nTASK [horizon : Set empty custom policy] ***************************************\nTuesday 24 February 2026  00:59:37 +0000 (0:00:01.200)       0:00:02.681 ****** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [horizon : include_tasks] *************************************************\nTuesday 24 February 2026  00:59:37 +0000 (0:00:00.423)       0:00:03.105 ****** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'name': 'cloudkitty', 'enabled': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'name': 'heat', 'enabled': 'no'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'name': 'ironic', 'enabled': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'name': 'masakari', 'enabled': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'name': 'mistral', 'enabled': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'name': 'tacker', 'enabled': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'name': 'trove', 'enabled': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'name': 'watcher', 'enabled': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'name': 'cloudkitty', 'enabled': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'name': 'heat', 'enabled': 'no'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'name': 'ironic', 'enabled': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'name': 'masakari', 'enabled': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'name': 'mistral', 'enabled': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'name': 'tacker', 'enabled': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'name': 'trove', 'enabled': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'name': 'watcher', 'enabled': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'name': 'cloudkitty', 'enabled': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'name': 'heat', 'enabled': 'no'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'name': 'ironic', 'enabled': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'name': 'masakari', 'enabled': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'name': 'mistral', 'enabled': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'name': 'tacker', 'enabled': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'name': 'trove', 'enabled': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'name': 'watcher', 'enabled': False}) \u001b[0m\n\u001b[0;36mincluded: /ansible/roles/horizon/tasks/policy_item.yml for testbed-node-0, testbed-node-1, testbed-node-2 => (item={'name': 'ceilometer', 'enabled': 'yes'})\u001b[0m\n\u001b[0;36mincluded: /ansible/roles/horizon/tasks/policy_item.yml for testbed-node-0, testbed-node-1, testbed-node-2 => (item={'name': 'cinder', 'enabled': 'yes'})\u001b[0m\n\u001b[0;36mincluded: /ansible/roles/horizon/tasks/policy_item.yml for testbed-node-0, testbed-node-1, testbed-node-2 => (item={'name': 'designate', 'enabled': True})\u001b[0m\n\u001b[0;36mincluded: /ansible/roles/horizon/tasks/policy_item.yml for testbed-node-0, testbed-node-1, testbed-node-2 => (item={'name': 'glance', 'enabled': True})\u001b[0m\n\u001b[0;36mincluded: /ansible/roles/horizon/tasks/policy_item.yml for testbed-node-0, testbed-node-1, testbed-node-2 => (item={'name': 'keystone', 'enabled': True})\u001b[0m\n\u001b[0;36mincluded: /ansible/roles/horizon/tasks/policy_item.yml for testbed-node-0, testbed-node-1, testbed-node-2 => (item={'name': 'magnum', 'enabled': True})\u001b[0m\n\u001b[0;36mincluded: /ansible/roles/horizon/tasks/policy_item.yml for testbed-node-0, testbed-node-1, testbed-node-2 => (item={'name': 'manila', 'enabled': True})\u001b[0m\n\u001b[0;36mincluded: /ansible/roles/horizon/tasks/policy_item.yml for testbed-node-0, testbed-node-1, testbed-node-2 => (item={'name': 'neutron', 'enabled': True})\u001b[0m\n\u001b[0;36mincluded: /ansible/roles/horizon/tasks/policy_item.yml for testbed-node-0, testbed-node-1, testbed-node-2 => (item={'name': 'nova', 'enabled': True})\u001b[0m\n\u001b[0;36mincluded: /ansible/roles/horizon/tasks/policy_item.yml for testbed-node-0, testbed-node-1, testbed-node-2 => (item={'name': 'octavia', 'enabled': True})\u001b[0m\n\nTASK [horizon : Update policy file name] ***************************************\nTuesday 24 February 2026  00:59:38 +0000 (0:00:00.782)       0:00:03.887 ****** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [horizon : Check if policies shall be overwritten] ************************\nTuesday 24 February 2026  00:59:38 +0000 (0:00:00.328)       0:00:04.215 ****** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [horizon : Update custom policy file name] ********************************\nTuesday 24 February 2026  00:59:38 +0000 (0:00:00.135)       0:00:04.351 ****** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [horizon : Update policy file name] ***************************************\nTuesday 24 February 2026  00:59:39 +0000 (0:00:00.463)       0:00:04.814 ****** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [horizon : Check if policies shall be overwritten] ************************\nTuesday 24 February 2026  00:59:39 +0000 (0:00:00.345)       0:00:05.160 ****** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [horizon : Update custom policy file name] ********************************\nTuesday 24 February 2026  00:59:39 +0000 (0:00:00.114)       0:00:05.275 ****** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [horizon : Update policy file name] ***************************************\nTuesday 24 February 2026  00:59:40 +0000 (0:00:00.349)       0:00:05.624 ****** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [horizon : Check if policies shall be overwritten] ************************\nTuesday 24 February 2026  00:59:40 +0000 (0:00:00.366)       0:00:05.991 ****** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [horizon : Update custom policy file name] ********************************\nTuesday 24 February 2026  00:59:40 +0000 (0:00:00.289)       0:00:06.281 ****** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [horizon : Update policy file name] ***************************************\nTuesday 24 February 2026  00:59:41 +0000 (0:00:00.277)       0:00:06.558 ****** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [horizon : Check if policies shall be overwritten] ************************\nTuesday 24 February 2026  00:59:41 +0000 (0:00:00.322)       0:00:06.881 ****** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [horizon : Update custom policy file name] ********************************\nTuesday 24 February 2026  00:59:41 +0000 (0:00:00.124)       0:00:07.005 ****** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [horizon : Update policy file name] ***************************************\nTuesday 24 February 2026  00:59:41 +0000 (0:00:00.265)       0:00:07.271 ****** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [horizon : Check if policies shall be overwritten] ************************\nTuesday 24 February 2026  00:59:42 +0000 (0:00:00.475)       0:00:07.747 ****** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [horizon : Update custom policy file name] ********************************\nTuesday 24 February 2026  00:59:42 +0000 (0:00:00.133)       0:00:07.881 ****** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [horizon : Update policy file name] ***************************************\nTuesday 24 February 2026  00:59:42 +0000 (0:00:00.274)       0:00:08.155 ****** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [horizon : Check if policies shall be overwritten] ************************\nTuesday 24 February 2026  00:59:43 +0000 (0:00:00.307)       0:00:08.462 ****** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [horizon : Update custom policy file name] ********************************\nTuesday 24 February 2026  00:59:43 +0000 (0:00:00.155)       0:00:08.617 ****** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [horizon : Update policy file name] ***************************************\nTuesday 24 February 2026  00:59:43 +0000 (0:00:00.293)       0:00:08.911 ****** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [horizon : Check if policies shall be overwritten] ************************\nTuesday 24 February 2026  00:59:43 +0000 (0:00:00.466)       0:00:09.377 ****** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [horizon : Update custom policy file name] ********************************\nTuesday 24 February 2026  00:59:44 +0000 (0:00:00.163)       0:00:09.541 ****** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [horizon : Update policy file name] ***************************************\nTuesday 24 February 2026  00:59:44 +0000 (0:00:00.277)       0:00:09.818 ****** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [horizon : Check if policies shall be overwritten] ************************\nTuesday 24 February 2026  00:59:44 +0000 (0:00:00.314)       0:00:10.133 ****** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [horizon : Update custom policy file name] ********************************\nTuesday 24 February 2026  00:59:44 +0000 (0:00:00.155)       0:00:10.289 ****** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [horizon : Update policy file name] ***************************************\nTuesday 24 February 2026  00:59:45 +0000 (0:00:00.484)       0:00:10.773 ****** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [horizon : Check if policies shall be overwritten] ************************\nTuesday 24 February 2026  00:59:45 +0000 (0:00:00.354)       0:00:11.128 ****** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [horizon : Update custom policy file name] ********************************\nTuesday 24 February 2026  00:59:45 +0000 (0:00:00.131)       0:00:11.259 ****** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [horizon : Update policy file name] ***************************************\nTuesday 24 February 2026  00:59:46 +0000 (0:00:00.296)       0:00:11.556 ****** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [horizon : Check if policies shall be overwritten] ************************\nTuesday 24 February 2026  00:59:46 +0000 (0:00:00.309)       0:00:11.865 ****** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [horizon : Update custom policy file name] ********************************\nTuesday 24 February 2026  00:59:46 +0000 (0:00:00.148)       0:00:12.014 ****** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [horizon : Copying over config.json files for services] *******************\nTuesday 24 February 2026  00:59:47 +0000 (0:00:00.529)       0:00:12.544 ****** \n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [horizon : Copying over horizon.conf] *************************************\nTuesday 24 February 2026  00:59:48 +0000 (0:00:01.616)       0:00:14.160 ****** \n\u001b[0;33mchanged: [testbed-node-0] => (item=/ansible/roles/horizon/templates/horizon.conf.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=/ansible/roles/horizon/templates/horizon.conf.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=/ansible/roles/horizon/templates/horizon.conf.j2)\u001b[0m\n\nTASK [horizon : Copying over kolla-settings.py] ********************************\nTuesday 24 February 2026  00:59:50 +0000 (0:00:02.210)       0:00:16.371 ****** \n\u001b[0;33mchanged: [testbed-node-0] => (item=/ansible/roles/horizon/templates/_9998-kolla-settings.py.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=/ansible/roles/horizon/templates/_9998-kolla-settings.py.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=/ansible/roles/horizon/templates/_9998-kolla-settings.py.j2)\u001b[0m\n\nTASK [horizon : Copying over custom-settings.py] *******************************\nTuesday 24 February 2026  00:59:53 +0000 (0:00:02.339)       0:00:18.710 ****** \n\u001b[0;33mchanged: [testbed-node-0] => (item=/ansible/roles/horizon/templates/_9999-custom-settings.py.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=/ansible/roles/horizon/templates/_9999-custom-settings.py.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=/ansible/roles/horizon/templates/_9999-custom-settings.py.j2)\u001b[0m\n\nTASK [horizon : Copying over existing policy file] *****************************\nTuesday 24 February 2026  00:59:55 +0000 (0:00:02.371)       0:00:21.082 ****** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [horizon : Copying over custom themes] ************************************\nTuesday 24 February 2026  00:59:56 +0000 (0:00:00.382)       0:00:21.464 ****** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [horizon : include_tasks] *************************************************\nTuesday 24 February 2026  00:59:56 +0000 (0:00:00.427)       0:00:21.892 ****** \n\u001b[0;36mincluded: /ansible/roles/horizon/tasks/copy-certs.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [service-cert-copy : horizon | Copying over extra CA certificates] ********\nTuesday 24 February 2026  00:59:57 +0000 (0:00:00.918)       0:00:22.810 ****** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'horizon', 'value': {'container_name': 'horizon', 'group': 'horizon', 'enabled': True, 'image': 'registry.osism.tech/kolla/horizon:2024.2', 'environment': {'ENABLE_BLAZAR': 'no', 'ENABLE_CLOUDKITTY': 'no', 'ENABLE_DESIGNATE': 'yes', 'ENABLE_FWAAS': 'no', 'ENABLE_HEAT': 'no', 'ENABLE_IRONIC': 'no', 'ENABLE_MAGNUM': 'yes', 'ENABLE_MANILA': 'yes', 'ENABLE_MASAKARI': 'no', 'ENABLE_MISTRAL': 'no', 'ENABLE_NEUTRON_VPNAAS': 'no', 'ENABLE_OCTAVIA': 'yes', 'ENABLE_TACKER': 'no', 'ENABLE_TROVE': 'no', 'ENABLE_WATCHER': 'no', 'ENABLE_ZUN': 'no', 'FORCE_GENERATE': 'no'}, 'volumes': ['/etc/kolla/horizon/:/var/lib/kolla/config_files/:ro', '', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:80'], 'timeout': '30'}, 'haproxy': {'horizon': {'enabled': True, 'mode': 'http', 'external': False, 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_redirect': {'enabled': True, 'mode': 'redirect', 'external': False, 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'horizon_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_external_redirect': {'enabled': True, 'mode': 'redirect', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'acme_client': {'enabled': True, 'with_frontend': False, 'custom_member_list': []}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'horizon', 'value': {'container_name': 'horizon', 'group': 'horizon', 'enabled': True, 'image': 'registry.osism.tech/kolla/horizon:2024.2', 'environment': {'ENABLE_BLAZAR': 'no', 'ENABLE_CLOUDKITTY': 'no', 'ENABLE_DESIGNATE': 'yes', 'ENABLE_FWAAS': 'no', 'ENABLE_HEAT': 'no', 'ENABLE_IRONIC': 'no', 'ENABLE_MAGNUM': 'yes', 'ENABLE_MANILA': 'yes', 'ENABLE_MASAKARI': 'no', 'ENABLE_MISTRAL': 'no', 'ENABLE_NEUTRON_VPNAAS': 'no', 'ENABLE_OCTAVIA': 'yes', 'ENABLE_TACKER': 'no', 'ENABLE_TROVE': 'no', 'ENABLE_WATCHER': 'no', 'ENABLE_ZUN': 'no', 'FORCE_GENERATE': 'no'}, 'volumes': ['/etc/kolla/horizon/:/var/lib/kolla/config_files/:ro', '', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:80'], 'timeout': '30'}, 'haproxy': {'horizon': {'enabled': True, 'mode': 'http', 'external': False, 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_redirect': {'enabled': True, 'mode': 'redirect', 'external': False, 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'horizon_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_external_redirect': {'enabled': True, 'mode': 'redirect', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'acme_client': {'enabled': True, 'with_frontend': False, 'custom_member_list': []}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'horizon', 'value': {'container_name': 'horizon', 'group': 'horizon', 'enabled': True, 'image': 'registry.osism.tech/kolla/horizon:2024.2', 'environment': {'ENABLE_BLAZAR': 'no', 'ENABLE_CLOUDKITTY': 'no', 'ENABLE_DESIGNATE': 'yes', 'ENABLE_FWAAS': 'no', 'ENABLE_HEAT': 'no', 'ENABLE_IRONIC': 'no', 'ENABLE_MAGNUM': 'yes', 'ENABLE_MANILA': 'yes', 'ENABLE_MASAKARI': 'no', 'ENABLE_MISTRAL': 'no', 'ENABLE_NEUTRON_VPNAAS': 'no', 'ENABLE_OCTAVIA': 'yes', 'ENABLE_TACKER': 'no', 'ENABLE_TROVE': 'no', 'ENABLE_WATCHER': 'no', 'ENABLE_ZUN': 'no', 'FORCE_GENERATE': 'no'}, 'volumes': ['/etc/kolla/horizon/:/var/lib/kolla/config_files/:ro', '', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:80'], 'timeout': '30'}, 'haproxy': {'horizon': {'enabled': True, 'mode': 'http', 'external': False, 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_redirect': {'enabled': True, 'mode': 'redirect', 'external': False, 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'horizon_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_external_redirect': {'enabled': True, 'mode': 'redirect', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'acme_client': {'enabled': True, 'with_frontend': False, 'custom_member_list': []}}}})\u001b[0m\n\nTASK [service-cert-copy : horizon | Copying over backend internal TLS certificate] ***\nTuesday 24 February 2026  00:59:59 +0000 (0:00:01.716)       0:00:24.526 ****** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'horizon', 'value': {'container_name': 'horizon', 'group': 'horizon', 'enabled': True, 'image': 'registry.osism.tech/kolla/horizon:2024.2', 'environment': {'ENABLE_BLAZAR': 'no', 'ENABLE_CLOUDKITTY': 'no', 'ENABLE_DESIGNATE': 'yes', 'ENABLE_FWAAS': 'no', 'ENABLE_HEAT': 'no', 'ENABLE_IRONIC': 'no', 'ENABLE_MAGNUM': 'yes', 'ENABLE_MANILA': 'yes', 'ENABLE_MASAKARI': 'no', 'ENABLE_MISTRAL': 'no', 'ENABLE_NEUTRON_VPNAAS': 'no', 'ENABLE_OCTAVIA': 'yes', 'ENABLE_TACKER': 'no', 'ENABLE_TROVE': 'no', 'ENABLE_WATCHER': 'no', 'ENABLE_ZUN': 'no', 'FORCE_GENERATE': 'no'}, 'volumes': ['/etc/kolla/horizon/:/var/lib/kolla/config_files/:ro', '', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:80'], 'timeout': '30'}, 'haproxy': {'horizon': {'enabled': True, 'mode': 'http', 'external': False, 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_redirect': {'enabled': True, 'mode': 'redirect', 'external': False, 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'horizon_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_external_redirect': {'enabled': True, 'mode': 'redirect', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'acme_client': {'enabled': True, 'with_frontend': False, 'custom_member_list': []}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'horizon', 'value': {'container_name': 'horizon', 'group': 'horizon', 'enabled': True, 'image': 'registry.osism.tech/kolla/horizon:2024.2', 'environment': {'ENABLE_BLAZAR': 'no', 'ENABLE_CLOUDKITTY': 'no', 'ENABLE_DESIGNATE': 'yes', 'ENABLE_FWAAS': 'no', 'ENABLE_HEAT': 'no', 'ENABLE_IRONIC': 'no', 'ENABLE_MAGNUM': 'yes', 'ENABLE_MANILA': 'yes', 'ENABLE_MASAKARI': 'no', 'ENABLE_MISTRAL': 'no', 'ENABLE_NEUTRON_VPNAAS': 'no', 'ENABLE_OCTAVIA': 'yes', 'ENABLE_TACKER': 'no', 'ENABLE_TROVE': 'no', 'ENABLE_WATCHER': 'no', 'ENABLE_ZUN': 'no', 'FORCE_GENERATE': 'no'}, 'volumes': ['/etc/kolla/horizon/:/var/lib/kolla/config_files/:ro', '', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:80'], 'timeout': '30'}, 'haproxy': {'horizon': {'enabled': True, 'mode': 'http', 'external': False, 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_redirect': {'enabled': True, 'mode': 'redirect', 'external': False, 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'horizon_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_external_redirect': {'enabled': True, 'mode': 'redirect', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'acme_client': {'enabled': True, 'with_frontend': False, 'custom_member_list': []}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'horizon', 'value': {'container_name': 'horizon', 'group': 'horizon', 'enabled': True, 'image': 'registry.osism.tech/kolla/horizon:2024.2', 'environment': {'ENABLE_BLAZAR': 'no', 'ENABLE_CLOUDKITTY': 'no', 'ENABLE_DESIGNATE': 'yes', 'ENABLE_FWAAS': 'no', 'ENABLE_HEAT': 'no', 'ENABLE_IRONIC': 'no', 'ENABLE_MAGNUM': 'yes', 'ENABLE_MANILA': 'yes', 'ENABLE_MASAKARI': 'no', 'ENABLE_MISTRAL': 'no', 'ENABLE_NEUTRON_VPNAAS': 'no', 'ENABLE_OCTAVIA': 'yes', 'ENABLE_TACKER': 'no', 'ENABLE_TROVE': 'no', 'ENABLE_WATCHER': 'no', 'ENABLE_ZUN': 'no', 'FORCE_GENERATE': 'no'}, 'volumes': ['/etc/kolla/horizon/:/var/lib/kolla/config_files/:ro', '', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:80'], 'timeout': '30'}, 'haproxy': {'horizon': {'enabled': True, 'mode': 'http', 'external': False, 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_redirect': {'enabled': True, 'mode': 'redirect', 'external': False, 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'horizon_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_external_redirect': {'enabled': True, 'mode': 'redirect', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'acme_client': {'enabled': True, 'with_frontend': False, 'custom_member_list': []}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [service-cert-copy : horizon | Copying over backend internal TLS key] *****\nTuesday 24 February 2026  00:59:59 +0000 (0:00:00.855)       0:00:25.381 ****** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'horizon', 'value': {'container_name': 'horizon', 'group': 'horizon', 'enabled': True, 'image': 'registry.osism.tech/kolla/horizon:2024.2', 'environment': {'ENABLE_BLAZAR': 'no', 'ENABLE_CLOUDKITTY': 'no', 'ENABLE_DESIGNATE': 'yes', 'ENABLE_FWAAS': 'no', 'ENABLE_HEAT': 'no', 'ENABLE_IRONIC': 'no', 'ENABLE_MAGNUM': 'yes', 'ENABLE_MANILA': 'yes', 'ENABLE_MASAKARI': 'no', 'ENABLE_MISTRAL': 'no', 'ENABLE_NEUTRON_VPNAAS': 'no', 'ENABLE_OCTAVIA': 'yes', 'ENABLE_TACKER': 'no', 'ENABLE_TROVE': 'no', 'ENABLE_WATCHER': 'no', 'ENABLE_ZUN': 'no', 'FORCE_GENERATE': 'no'}, 'volumes': ['/etc/kolla/horizon/:/var/lib/kolla/config_files/:ro', '', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:80'], 'timeout': '30'}, 'haproxy': {'horizon': {'enabled': True, 'mode': 'http', 'external': False, 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_redirect': {'enabled': True, 'mode': 'redirect', 'external': False, 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'horizon_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_external_redirect': {'enabled': True, 'mode': 'redirect', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'acme_client': {'enabled': True, 'with_frontend': False, 'custom_member_list': []}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'horizon', 'value': {'container_name': 'horizon', 'group': 'horizon', 'enabled': True, 'image': 'registry.osism.tech/kolla/horizon:2024.2', 'environment': {'ENABLE_BLAZAR': 'no', 'ENABLE_CLOUDKITTY': 'no', 'ENABLE_DESIGNATE': 'yes', 'ENABLE_FWAAS': 'no', 'ENABLE_HEAT': 'no', 'ENABLE_IRONIC': 'no', 'ENABLE_MAGNUM': 'yes', 'ENABLE_MANILA': 'yes', 'ENABLE_MASAKARI': 'no', 'ENABLE_MISTRAL': 'no', 'ENABLE_NEUTRON_VPNAAS': 'no', 'ENABLE_OCTAVIA': 'yes', 'ENABLE_TACKER': 'no', 'ENABLE_TROVE': 'no', 'ENABLE_WATCHER': 'no', 'ENABLE_ZUN': 'no', 'FORCE_GENERATE': 'no'}, 'volumes': ['/etc/kolla/horizon/:/var/lib/kolla/config_files/:ro', '', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:80'], 'timeout': '30'}, 'haproxy': {'horizon': {'enabled': True, 'mode': 'http', 'external': False, 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_redirect': {'enabled': True, 'mode': 'redirect', 'external': False, 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'horizon_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_external_redirect': {'enabled': True, 'mode': 'redirect', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'acme_client': {'enabled': True, 'with_frontend': False, 'custom_member_list': []}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'horizon', 'value': {'container_name': 'horizon', 'group': 'horizon', 'enabled': True, 'image': 'registry.osism.tech/kolla/horizon:2024.2', 'environment': {'ENABLE_BLAZAR': 'no', 'ENABLE_CLOUDKITTY': 'no', 'ENABLE_DESIGNATE': 'yes', 'ENABLE_FWAAS': 'no', 'ENABLE_HEAT': 'no', 'ENABLE_IRONIC': 'no', 'ENABLE_MAGNUM': 'yes', 'ENABLE_MANILA': 'yes', 'ENABLE_MASAKARI': 'no', 'ENABLE_MISTRAL': 'no', 'ENABLE_NEUTRON_VPNAAS': 'no', 'ENABLE_OCTAVIA': 'yes', 'ENABLE_TACKER': 'no', 'ENABLE_TROVE': 'no', 'ENABLE_WATCHER': 'no', 'ENABLE_ZUN': 'no', 'FORCE_GENERATE': 'no'}, 'volumes': ['/etc/kolla/horizon/:/var/lib/kolla/config_files/:ro', '', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:80'], 'timeout': '30'}, 'haproxy': {'horizon': {'enabled': True, 'mode': 'http', 'external': False, 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_redirect': {'enabled': True, 'mode': 'redirect', 'external': False, 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'horizon_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_external_redirect': {'enabled': True, 'mode': 'redirect', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'acme_client': {'enabled': True, 'with_frontend': False, 'custom_member_list': []}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [horizon : Deploy horizon container] **************************************\nTuesday 24 February 2026  01:00:01 +0000 (0:00:01.038)       0:00:26.420 ****** \n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'horizon', 'value': {'container_name': 'horizon', 'group': 'horizon', 'enabled': True, 'image': 'registry.osism.tech/kolla/horizon:2024.2', 'environment': {'ENABLE_BLAZAR': 'no', 'ENABLE_CLOUDKITTY': 'no', 'ENABLE_DESIGNATE': 'yes', 'ENABLE_FWAAS': 'no', 'ENABLE_HEAT': 'no', 'ENABLE_IRONIC': 'no', 'ENABLE_MAGNUM': 'yes', 'ENABLE_MANILA': 'yes', 'ENABLE_MASAKARI': 'no', 'ENABLE_MISTRAL': 'no', 'ENABLE_NEUTRON_VPNAAS': 'no', 'ENABLE_OCTAVIA': 'yes', 'ENABLE_TACKER': 'no', 'ENABLE_TROVE': 'no', 'ENABLE_WATCHER': 'no', 'ENABLE_ZUN': 'no', 'FORCE_GENERATE': 'no'}, 'volumes': ['/etc/kolla/horizon/:/var/lib/kolla/config_files/:ro', '', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:80'], 'timeout': '30'}, 'haproxy': {'horizon': {'enabled': True, 'mode': 'http', 'external': False, 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_redirect': {'enabled': True, 'mode': 'redirect', 'external': False, 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'horizon_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_external_redirect': {'enabled': True, 'mode': 'redirect', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'acme_client': {'enabled': True, 'with_frontend': False, 'custom_member_list': []}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'horizon', 'value': {'container_name': 'horizon', 'group': 'horizon', 'enabled': True, 'image': 'registry.osism.tech/kolla/horizon:2024.2', 'environment': {'ENABLE_BLAZAR': 'no', 'ENABLE_CLOUDKITTY': 'no', 'ENABLE_DESIGNATE': 'yes', 'ENABLE_FWAAS': 'no', 'ENABLE_HEAT': 'no', 'ENABLE_IRONIC': 'no', 'ENABLE_MAGNUM': 'yes', 'ENABLE_MANILA': 'yes', 'ENABLE_MASAKARI': 'no', 'ENABLE_MISTRAL': 'no', 'ENABLE_NEUTRON_VPNAAS': 'no', 'ENABLE_OCTAVIA': 'yes', 'ENABLE_TACKER': 'no', 'ENABLE_TROVE': 'no', 'ENABLE_WATCHER': 'no', 'ENABLE_ZUN': 'no', 'FORCE_GENERATE': 'no'}, 'volumes': ['/etc/kolla/horizon/:/var/lib/kolla/config_files/:ro', '', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:80'], 'timeout': '30'}, 'haproxy': {'horizon': {'enabled': True, 'mode': 'http', 'external': False, 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_redirect': {'enabled': True, 'mode': 'redirect', 'external': False, 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'horizon_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_external_redirect': {'enabled': True, 'mode': 'redirect', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'acme_client': {'enabled': True, 'with_frontend': False, 'custom_member_list': []}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'horizon', 'value': {'container_name': 'horizon', 'group': 'horizon', 'enabled': True, 'image': 'registry.osism.tech/kolla/horizon:2024.2', 'environment': {'ENABLE_BLAZAR': 'no', 'ENABLE_CLOUDKITTY': 'no', 'ENABLE_DESIGNATE': 'yes', 'ENABLE_FWAAS': 'no', 'ENABLE_HEAT': 'no', 'ENABLE_IRONIC': 'no', 'ENABLE_MAGNUM': 'yes', 'ENABLE_MANILA': 'yes', 'ENABLE_MASAKARI': 'no', 'ENABLE_MISTRAL': 'no', 'ENABLE_NEUTRON_VPNAAS': 'no', 'ENABLE_OCTAVIA': 'yes', 'ENABLE_TACKER': 'no', 'ENABLE_TROVE': 'no', 'ENABLE_WATCHER': 'no', 'ENABLE_ZUN': 'no', 'FORCE_GENERATE': 'no'}, 'volumes': ['/etc/kolla/horizon/:/var/lib/kolla/config_files/:ro', '', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:80'], 'timeout': '30'}, 'haproxy': {'horizon': {'enabled': True, 'mode': 'http', 'external': False, 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_redirect': {'enabled': True, 'mode': 'redirect', 'external': False, 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'horizon_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_external_redirect': {'enabled': True, 'mode': 'redirect', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'acme_client': {'enabled': True, 'with_frontend': False, 'custom_member_list': []}}}})\u001b[0m\n\nTASK [horizon : include_tasks] *************************************************\nTuesday 24 February 2026  01:00:02 +0000 (0:00:01.756)       0:00:28.176 ****** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [horizon : include_tasks] *************************************************\nTuesday 24 February 2026  01:00:03 +0000 (0:00:00.295)       0:00:28.471 ****** \n\u001b[0;36mincluded: /ansible/roles/horizon/tasks/bootstrap.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [horizon : Creating Horizon database] *************************************\nTuesday 24 February 2026  01:00:03 +0000 (0:00:00.586)       0:00:29.058 ****** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [horizon : Creating Horizon database user and setting permissions] ********\nTuesday 24 February 2026  01:00:06 +0000 (0:00:02.981)       0:00:32.040 ****** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [horizon : Running Horizon bootstrap container] ***************************\nTuesday 24 February 2026  01:00:09 +0000 (0:00:03.089)       0:00:35.130 ****** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [horizon : Flush handlers] ************************************************\nTuesday 24 February 2026  01:00:28 +0000 (0:00:18.547)       0:00:53.677 ****** \n\nTASK [horizon : Flush handlers] ************************************************\nTuesday 24 February 2026  01:00:28 +0000 (0:00:00.129)       0:00:53.807 ****** \n\nTASK [horizon : Flush handlers] ************************************************\nTuesday 24 February 2026  01:00:28 +0000 (0:00:00.102)       0:00:53.910 ****** \n\nRUNNING HANDLER [horizon : Restart horizon container] **************************\nTuesday 24 February 2026  01:00:28 +0000 (0:00:00.074)       0:00:53.984 ****** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=37  \u001b[0m \u001b[0;33mchanged=11  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=25  \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=34  \u001b[0m \u001b[0;33mchanged=8   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=15  \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=34  \u001b[0m \u001b[0;33mchanged=8   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=15  \u001b[0m rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nTuesday 24 February 2026  01:01:19 +0000 (0:00:50.936)       0:01:44.921 ****** \n=============================================================================== \nhorizon : Restart horizon container ------------------------------------ 50.94s\nhorizon : Running Horizon bootstrap container -------------------------- 18.55s\nhorizon : Creating Horizon database user and setting permissions -------- 3.09s\nhorizon : Creating Horizon database ------------------------------------- 2.98s\nhorizon : Copying over custom-settings.py ------------------------------- 2.37s\nhorizon : Copying over kolla-settings.py -------------------------------- 2.34s\nhorizon : Copying over horizon.conf ------------------------------------- 2.21s\nhorizon : Deploy horizon container -------------------------------------- 1.76s\nservice-cert-copy : horizon | Copying over extra CA certificates -------- 1.72s\nhorizon : Copying over config.json files for services ------------------- 1.62s\nhorizon : Ensuring config directories exist ----------------------------- 1.20s\nservice-cert-copy : horizon | Copying over backend internal TLS key ----- 1.04s\nhorizon : include_tasks ------------------------------------------------- 0.92s\nservice-cert-copy : horizon | Copying over backend internal TLS certificate --- 0.86s\nhorizon : include_tasks ------------------------------------------------- 0.78s\nhorizon : include_tasks ------------------------------------------------- 0.59s\nhorizon : Update custom policy file name -------------------------------- 0.53s\nhorizon : include_tasks ------------------------------------------------- 0.50s\nhorizon : Update custom policy file name -------------------------------- 0.48s\nhorizon : Update policy file name --------------------------------------- 0.48s\n\u001b[32m2026-02-24 01:01:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9cb83792-8f4e-4574-a78c-9b0e654d6fd6 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:01:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 01:01:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c5f0c529-69aa-4268-97af-bde1796abaa2 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:01:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9cb83792-8f4e-4574-a78c-9b0e654d6fd6 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:01:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 01:01:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c5f0c529-69aa-4268-97af-bde1796abaa2 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:01:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9cb83792-8f4e-4574-a78c-9b0e654d6fd6 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:01:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 01:01:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c5f0c529-69aa-4268-97af-bde1796abaa2 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:01:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9cb83792-8f4e-4574-a78c-9b0e654d6fd6 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:01:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 01:01:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c5f0c529-69aa-4268-97af-bde1796abaa2 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:01:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9cb83792-8f4e-4574-a78c-9b0e654d6fd6 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:01:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 01:01:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c5f0c529-69aa-4268-97af-bde1796abaa2 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:01:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9cb83792-8f4e-4574-a78c-9b0e654d6fd6 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:01:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 01:01:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c5f0c529-69aa-4268-97af-bde1796abaa2 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:01:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9cb83792-8f4e-4574-a78c-9b0e654d6fd6 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:01:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 01:01:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c5f0c529-69aa-4268-97af-bde1796abaa2 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:01:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9cb83792-8f4e-4574-a78c-9b0e654d6fd6 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:01:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 01:01:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c5f0c529-69aa-4268-97af-bde1796abaa2 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:01:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9cb83792-8f4e-4574-a78c-9b0e654d6fd6 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:01:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 01:01:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c5f0c529-69aa-4268-97af-bde1796abaa2 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:01:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9cb83792-8f4e-4574-a78c-9b0e654d6fd6 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:01:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 01:01:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c5f0c529-69aa-4268-97af-bde1796abaa2 is in state SUCCESS\u001b[0m\n\u001b[32m2026-02-24 01:01:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9cb83792-8f4e-4574-a78c-9b0e654d6fd6 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:01:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 81556691-758e-4eda-9a1d-3b717f8d6221 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:01:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 01:01:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9cb83792-8f4e-4574-a78c-9b0e654d6fd6 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:01:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 81556691-758e-4eda-9a1d-3b717f8d6221 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:01:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 01:01:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9cb83792-8f4e-4574-a78c-9b0e654d6fd6 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:01:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 81556691-758e-4eda-9a1d-3b717f8d6221 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:01:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 01:02:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9cb83792-8f4e-4574-a78c-9b0e654d6fd6 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:02:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 81556691-758e-4eda-9a1d-3b717f8d6221 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:02:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 01:02:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9cb83792-8f4e-4574-a78c-9b0e654d6fd6 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:02:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 81556691-758e-4eda-9a1d-3b717f8d6221 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:02:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 01:02:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9cb83792-8f4e-4574-a78c-9b0e654d6fd6 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:02:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 81556691-758e-4eda-9a1d-3b717f8d6221 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:02:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 01:02:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9cb83792-8f4e-4574-a78c-9b0e654d6fd6 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:02:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 81556691-758e-4eda-9a1d-3b717f8d6221 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:02:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 01:02:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9cb83792-8f4e-4574-a78c-9b0e654d6fd6 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:02:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 81556691-758e-4eda-9a1d-3b717f8d6221 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:02:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 01:02:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9cb83792-8f4e-4574-a78c-9b0e654d6fd6 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:02:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 81556691-758e-4eda-9a1d-3b717f8d6221 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:02:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 01:02:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9cb83792-8f4e-4574-a78c-9b0e654d6fd6 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:02:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 81556691-758e-4eda-9a1d-3b717f8d6221 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:02:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 01:02:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9cb83792-8f4e-4574-a78c-9b0e654d6fd6 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:02:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 81556691-758e-4eda-9a1d-3b717f8d6221 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:02:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 01:02:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9cb83792-8f4e-4574-a78c-9b0e654d6fd6 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:02:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 81556691-758e-4eda-9a1d-3b717f8d6221 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:02:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 01:02:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9cb83792-8f4e-4574-a78c-9b0e654d6fd6 is in state SUCCESS\u001b[0m\n\n\nPLAY [Copy ceph keys to the configuration repository] **************************\n\nTASK [Check if ceph keys exist] ************************************************\nTuesday 24 February 2026  01:01:11 +0000 (0:00:00.182)       0:00:00.182 ****** \n\u001b[0;32mok: [testbed-manager -> testbed-node-0(192.168.16.10)] => (item=ceph.client.admin.keyring)\u001b[0m\n\u001b[0;32mok: [testbed-manager -> testbed-node-0(192.168.16.10)] => (item=ceph.client.cinder.keyring)\u001b[0m\n\u001b[0;32mok: [testbed-manager -> testbed-node-0(192.168.16.10)] => (item=ceph.client.cinder.keyring)\u001b[0m\n\u001b[0;32mok: [testbed-manager -> testbed-node-0(192.168.16.10)] => (item=ceph.client.cinder-backup.keyring)\u001b[0m\n\u001b[0;32mok: [testbed-manager -> testbed-node-0(192.168.16.10)] => (item=ceph.client.cinder.keyring)\u001b[0m\n\u001b[0;32mok: [testbed-manager -> testbed-node-0(192.168.16.10)] => (item=ceph.client.nova.keyring)\u001b[0m\n\u001b[0;32mok: [testbed-manager -> testbed-node-0(192.168.16.10)] => (item=ceph.client.glance.keyring)\u001b[0m\n\u001b[0;32mok: [testbed-manager -> testbed-node-0(192.168.16.10)] => (item=ceph.client.gnocchi.keyring)\u001b[0m\n\u001b[0;32mok: [testbed-manager -> testbed-node-0(192.168.16.10)] => (item=ceph.client.manila.keyring)\u001b[0m\n\nTASK [Fetch all ceph keys] *****************************************************\nTuesday 24 February 2026  01:01:17 +0000 (0:00:05.357)       0:00:05.540 ****** \n\u001b[0;32mok: [testbed-manager -> testbed-node-0(192.168.16.10)] => (item=ceph.client.admin.keyring)\u001b[0m\n\u001b[0;32mok: [testbed-manager -> testbed-node-0(192.168.16.10)] => (item=ceph.client.cinder.keyring)\u001b[0m\n\u001b[0;32mok: [testbed-manager -> testbed-node-0(192.168.16.10)] => (item=ceph.client.cinder.keyring)\u001b[0m\n\u001b[0;32mok: [testbed-manager -> testbed-node-0(192.168.16.10)] => (item=ceph.client.cinder-backup.keyring)\u001b[0m\n\u001b[0;32mok: [testbed-manager -> testbed-node-0(192.168.16.10)] => (item=ceph.client.cinder.keyring)\u001b[0m\n\u001b[0;32mok: [testbed-manager -> testbed-node-0(192.168.16.10)] => (item=ceph.client.nova.keyring)\u001b[0m\n\u001b[0;32mok: [testbed-manager -> testbed-node-0(192.168.16.10)] => (item=ceph.client.glance.keyring)\u001b[0m\n\u001b[0;32mok: [testbed-manager -> testbed-node-0(192.168.16.10)] => (item=ceph.client.gnocchi.keyring)\u001b[0m\n\u001b[0;32mok: [testbed-manager -> testbed-node-0(192.168.16.10)] => (item=ceph.client.manila.keyring)\u001b[0m\n\nTASK [Create share directory] **************************************************\nTuesday 24 February 2026  01:01:22 +0000 (0:00:04.972)       0:00:10.512 ****** \n\u001b[0;33mchanged: [testbed-manager -> localhost]\u001b[0m\n\nTASK [Write ceph keys to the share directory] **********************************\nTuesday 24 February 2026  01:01:23 +0000 (0:00:01.118)       0:00:11.631 ****** \n\u001b[0;33mchanged: [testbed-manager -> localhost] => (item=ceph.client.admin.keyring)\u001b[0m\n\u001b[0;33mchanged: [testbed-manager -> localhost] => (item=ceph.client.cinder.keyring)\u001b[0m\n\u001b[0;32mok: [testbed-manager -> localhost] => (item=ceph.client.cinder.keyring)\u001b[0m\n\u001b[0;33mchanged: [testbed-manager -> localhost] => (item=ceph.client.cinder-backup.keyring)\u001b[0m\n\u001b[0;32mok: [testbed-manager -> localhost] => (item=ceph.client.cinder.keyring)\u001b[0m\n\u001b[0;33mchanged: [testbed-manager -> localhost] => (item=ceph.client.nova.keyring)\u001b[0m\n\u001b[0;33mchanged: [testbed-manager -> localhost] => (item=ceph.client.glance.keyring)\u001b[0m\n\u001b[0;33mchanged: [testbed-manager -> localhost] => (item=ceph.client.gnocchi.keyring)\u001b[0m\n\u001b[0;33mchanged: [testbed-manager -> localhost] => (item=ceph.client.manila.keyring)\u001b[0m\n\nTASK [Check if target directories exist] ***************************************\nTuesday 24 February 2026  01:01:38 +0000 (0:00:15.235)       0:00:26.866 ****** \n\u001b[0;32mok: [testbed-manager] => (item=/opt/configuration/environments/infrastructure/files/ceph)\u001b[0m\n\u001b[0;32mok: [testbed-manager] => (item=/opt/configuration/environments/kolla/files/overlays/cinder/cinder-volume)\u001b[0m\n\u001b[0;32mok: [testbed-manager] => (item=/opt/configuration/environments/kolla/files/overlays/cinder/cinder-backup)\u001b[0m\n\u001b[0;32mok: [testbed-manager] => (item=/opt/configuration/environments/kolla/files/overlays/cinder/cinder-backup)\u001b[0m\n\u001b[0;32mok: [testbed-manager] => (item=/opt/configuration/environments/kolla/files/overlays/nova)\u001b[0m\n\u001b[0;32mok: [testbed-manager] => (item=/opt/configuration/environments/kolla/files/overlays/nova)\u001b[0m\n\u001b[0;32mok: [testbed-manager] => (item=/opt/configuration/environments/kolla/files/overlays/glance)\u001b[0m\n\u001b[0;32mok: [testbed-manager] => (item=/opt/configuration/environments/kolla/files/overlays/gnocchi)\u001b[0m\n\u001b[0;32mok: [testbed-manager] => (item=/opt/configuration/environments/kolla/files/overlays/manila)\u001b[0m\n\nTASK [Write ceph keys to the configuration directory] **************************\nTuesday 24 February 2026  01:01:41 +0000 (0:00:03.514)       0:00:30.380 ****** \n\u001b[0;33mchanged: [testbed-manager] => (item=ceph.client.admin.keyring)\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item=ceph.client.cinder.keyring)\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item=ceph.client.cinder.keyring)\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item=ceph.client.cinder-backup.keyring)\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item=ceph.client.cinder.keyring)\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item=ceph.client.nova.keyring)\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item=ceph.client.glance.keyring)\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item=ceph.client.gnocchi.keyring)\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item=ceph.client.manila.keyring)\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;33mtestbed-manager\u001b[0m            : \u001b[0;32mok=6   \u001b[0m \u001b[0;33mchanged=3   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nTuesday 24 February 2026  01:01:49 +0000 (0:00:07.411)       0:00:37.792 ****** \n=============================================================================== \nWrite ceph keys to the share directory --------------------------------- 15.24s\nWrite ceph keys to the configuration directory -------------------------- 7.41s\nCheck if ceph keys exist ------------------------------------------------ 5.36s\nFetch all ceph keys ----------------------------------------------------- 4.97s\nCheck if target directories exist --------------------------------------- 3.51s\nCreate share directory -------------------------------------------------- 1.12s\n\n\nPLAY [Group hosts based on configuration] **************************************\n\nTASK [Group hosts based on Kolla action] ***************************************\nTuesday 24 February 2026  00:59:34 +0000 (0:00:00.251)       0:00:00.251 ****** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [Group hosts based on enabled services] ***********************************\nTuesday 24 February 2026  00:59:35 +0000 (0:00:00.346)       0:00:00.598 ****** \n\u001b[0;32mok: [testbed-node-0] => (item=enable_keystone_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item=enable_keystone_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item=enable_keystone_True)\u001b[0m\n\nPLAY [Apply role keystone] *****************************************************\n\nTASK [keystone : include_tasks] ************************************************\nTuesday 24 February 2026  00:59:35 +0000 (0:00:00.438)       0:00:01.037 ****** \n\u001b[0;36mincluded: /ansible/roles/keystone/tasks/deploy.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [keystone : Ensuring config directories exist] ****************************\nTuesday 24 February 2026  00:59:36 +0000 (0:00:00.577)       0:00:01.614 ****** \n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone:2024.2', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone:2024.2', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone:2024.2', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-ssh:2024.2', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-ssh:2024.2', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-ssh:2024.2', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-fernet:2024.2', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-fernet:2024.2', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-fernet:2024.2', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}})\u001b[0m\n\nTASK [keystone : Check if policies shall be overwritten] ***********************\nTuesday 24 February 2026  00:59:38 +0000 (0:00:01.908)       0:00:03.523 ****** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [keystone : Set keystone policy file] *************************************\nTuesday 24 February 2026  00:59:38 +0000 (0:00:00.123)       0:00:03.646 ****** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [keystone : Check if Keystone domain-specific config is supplied] *********\nTuesday 24 February 2026  00:59:38 +0000 (0:00:00.466)       0:00:04.113 ****** \n\u001b[0;32mok: [testbed-node-0 -> localhost]\u001b[0m\n\nTASK [keystone : include_tasks] ************************************************\nTuesday 24 February 2026  00:59:39 +0000 (0:00:00.878)       0:00:04.991 ****** \n\u001b[0;36mincluded: /ansible/roles/keystone/tasks/copy-certs.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [service-cert-copy : keystone | Copying over extra CA certificates] *******\nTuesday 24 February 2026  00:59:40 +0000 (0:00:00.576)       0:00:05.568 ****** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone:2024.2', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone:2024.2', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone:2024.2', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-ssh:2024.2', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-ssh:2024.2', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-ssh:2024.2', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-fernet:2024.2', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-fernet:2024.2', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-fernet:2024.2', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}})\u001b[0m\n\nTASK [service-cert-copy : keystone | Copying over backend internal TLS certificate] ***\nTuesday 24 February 2026  00:59:43 +0000 (0:00:03.540)       0:00:09.109 ****** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone:2024.2', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-ssh:2024.2', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-fernet:2024.2', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone:2024.2', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-ssh:2024.2', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-fernet:2024.2', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone:2024.2', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-ssh:2024.2', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-fernet:2024.2', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [service-cert-copy : keystone | Copying over backend internal TLS key] ****\nTuesday 24 February 2026  00:59:44 +0000 (0:00:00.638)       0:00:09.747 ****** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone:2024.2', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-ssh:2024.2', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-fernet:2024.2', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone:2024.2', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-ssh:2024.2', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-fernet:2024.2', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone:2024.2', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-ssh:2024.2', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-fernet:2024.2', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [keystone : Copying over config.json files for services] ******************\nTuesday 24 February 2026  00:59:45 +0000 (0:00:00.763)       0:00:10.511 ****** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone:2024.2', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone:2024.2', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone:2024.2', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-ssh:2024.2', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-ssh:2024.2', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-ssh:2024.2', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-fernet:2024.2', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-fernet:2024.2', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-fernet:2024.2', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}})\u001b[0m\n\nTASK [keystone : Copying over keystone.conf] ***********************************\nTuesday 24 February 2026  00:59:48 +0000 (0:00:03.411)       0:00:13.923 ****** \n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone:2024.2', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone:2024.2', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-ssh:2024.2', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-ssh:2024.2', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone:2024.2', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-ssh:2024.2', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-fernet:2024.2', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-fernet:2024.2', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-fernet:2024.2', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}})\u001b[0m\n\nTASK [keystone : Copying keystone-startup script for keystone] *****************\nTuesday 24 February 2026  00:59:54 +0000 (0:00:06.096)       0:00:20.019 ****** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [keystone : Create Keystone domain-specific config directory] *************\nTuesday 24 February 2026  00:59:56 +0000 (0:00:01.571)       0:00:21.590 ****** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [keystone : Get file list in custom domains folder] ***********************\nTuesday 24 February 2026  00:59:56 +0000 (0:00:00.655)       0:00:22.246 ****** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [keystone : Copying Keystone Domain specific settings] ********************\nTuesday 24 February 2026  00:59:57 +0000 (0:00:00.339)       0:00:22.585 ****** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [keystone : Copying over existing policy file] ****************************\nTuesday 24 February 2026  00:59:57 +0000 (0:00:00.662)       0:00:23.247 ****** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone:2024.2', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-ssh:2024.2', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-fernet:2024.2', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone:2024.2', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-ssh:2024.2', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-fernet:2024.2', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone:2024.2', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-ssh:2024.2', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-fernet:2024.2', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [keystone : include_tasks] ************************************************\nTuesday 24 February 2026  00:59:58 +0000 (0:00:00.645)       0:00:23.893 ****** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [keystone : Copying over wsgi-keystone.conf] ******************************\nTuesday 24 February 2026  00:59:58 +0000 (0:00:00.378)       0:00:24.272 ****** \n\u001b[0;33mchanged: [testbed-node-0] => (item=/ansible/roles/keystone/templates/wsgi-keystone.conf.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=/ansible/roles/keystone/templates/wsgi-keystone.conf.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=/ansible/roles/keystone/templates/wsgi-keystone.conf.j2)\u001b[0m\n\nTASK [keystone : Checking whether keystone-paste.ini file exists] **************\nTuesday 24 February 2026  01:00:00 +0000 (0:00:01.724)       0:00:25.997 ****** \n\u001b[0;32mok: [testbed-node-0 -> localhost]\u001b[0m\n\nTASK [keystone : Copying over keystone-paste.ini] ******************************\nTuesday 24 February 2026  01:00:01 +0000 (0:00:01.124)       0:00:27.122 ****** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [keystone : Generate the required cron jobs for the node] *****************\nTuesday 24 February 2026  01:00:02 +0000 (0:00:01.015)       0:00:28.137 ****** \n\u001b[0;32mok: [testbed-node-1 -> localhost]\u001b[0m\n\u001b[0;32mok: [testbed-node-0 -> localhost]\u001b[0m\n\u001b[0;32mok: [testbed-node-2 -> localhost]\u001b[0m\n\nTASK [keystone : Set fact with the generated cron jobs for building the crontab later] ***\nTuesday 24 February 2026  01:00:03 +0000 (0:00:01.051)       0:00:29.188 ****** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [keystone : Copying files for keystone-fernet] ****************************\nTuesday 24 February 2026  01:00:04 +0000 (0:00:00.296)       0:00:29.485 ****** \n\u001b[0;33mchanged: [testbed-node-1] => (item={'src': 'crontab.j2', 'dest': 'crontab'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'src': 'crontab.j2', 'dest': 'crontab'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'src': 'crontab.j2', 'dest': 'crontab'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'src': 'fernet-rotate.sh.j2', 'dest': 'fernet-rotate.sh'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'src': 'fernet-rotate.sh.j2', 'dest': 'fernet-rotate.sh'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'src': 'fernet-rotate.sh.j2', 'dest': 'fernet-rotate.sh'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'src': 'fernet-node-sync.sh.j2', 'dest': 'fernet-node-sync.sh'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'src': 'fernet-node-sync.sh.j2', 'dest': 'fernet-node-sync.sh'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'src': 'fernet-node-sync.sh.j2', 'dest': 'fernet-node-sync.sh'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'src': 'fernet-push.sh.j2', 'dest': 'fernet-push.sh'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'src': 'fernet-push.sh.j2', 'dest': 'fernet-push.sh'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'src': 'fernet-push.sh.j2', 'dest': 'fernet-push.sh'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'src': 'fernet-healthcheck.sh.j2', 'dest': 'fernet-healthcheck.sh'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'src': 'fernet-healthcheck.sh.j2', 'dest': 'fernet-healthcheck.sh'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'src': 'fernet-healthcheck.sh.j2', 'dest': 'fernet-healthcheck.sh'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'src': 'id_rsa', 'dest': 'id_rsa'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'src': 'id_rsa', 'dest': 'id_rsa'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'src': 'id_rsa', 'dest': 'id_rsa'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'src': 'ssh_config.j2', 'dest': 'ssh_config'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'src': 'ssh_config.j2', 'dest': 'ssh_config'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'src': 'ssh_config.j2', 'dest': 'ssh_config'})\u001b[0m\n\nTASK [keystone : Copying files for keystone-ssh] *******************************\nTuesday 24 February 2026  01:00:13 +0000 (0:00:09.426)       0:00:38.912 ****** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'src': 'sshd_config.j2', 'dest': 'sshd_config'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'src': 'sshd_config.j2', 'dest': 'sshd_config'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'src': 'sshd_config.j2', 'dest': 'sshd_config'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'src': 'id_rsa.pub', 'dest': 'id_rsa.pub'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'src': 'id_rsa.pub', 'dest': 'id_rsa.pub'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'src': 'id_rsa.pub', 'dest': 'id_rsa.pub'})\u001b[0m\n\nTASK [keystone : Check keystone containers] ************************************\nTuesday 24 February 2026  01:00:16 +0000 (0:00:02.888)       0:00:41.800 ****** \n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone:2024.2', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone:2024.2', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone:2024.2', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-ssh:2024.2', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-ssh:2024.2', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-ssh:2024.2', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-fernet:2024.2', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-fernet:2024.2', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-fernet:2024.2', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}})\u001b[0m\n\nTASK [keystone : include_tasks] ************************************************\nTuesday 24 February 2026  01:00:19 +0000 (0:00:02.779)       0:00:44.579 ****** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [keystone : Creating keystone database] ***********************************\nTuesday 24 February 2026  01:00:19 +0000 (0:00:00.387)       0:00:44.967 ****** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [keystone : Creating Keystone database user and setting permissions] ******\nTuesday 24 February 2026  01:00:22 +0000 (0:00:02.942)       0:00:47.910 ****** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [keystone : Checking for any running keystone_fernet containers] **********\nTuesday 24 February 2026  01:00:25 +0000 (0:00:02.689)       0:00:50.600 ****** \n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [keystone : Group nodes where keystone_fernet is running] *****************\nTuesday 24 February 2026  01:00:26 +0000 (0:00:01.140)       0:00:51.741 ****** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [keystone : Fail if any hosts need bootstrapping and not all hosts targeted] ***\nTuesday 24 February 2026  01:00:26 +0000 (0:00:00.334)       0:00:52.075 ****** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [keystone : Running Keystone bootstrap container] *************************\nTuesday 24 February 2026  01:00:27 +0000 (0:00:00.344)       0:00:52.419 ****** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [keystone : Running Keystone fernet bootstrap container] ******************\nTuesday 24 February 2026  01:00:44 +0000 (0:00:17.780)       0:01:10.199 ****** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [keystone : Flush handlers] ***********************************************\nTuesday 24 February 2026  01:00:57 +0000 (0:00:12.957)       0:01:23.156 ****** \n\nTASK [keystone : Flush handlers] ***********************************************\nTuesday 24 February 2026  01:00:57 +0000 (0:00:00.070)       0:01:23.227 ****** \n\nTASK [keystone : Flush handlers] ***********************************************\nTuesday 24 February 2026  01:00:57 +0000 (0:00:00.068)       0:01:23.296 ****** \n\nRUNNING HANDLER [keystone : Restart keystone-ssh container] ********************\nTuesday 24 February 2026  01:00:57 +0000 (0:00:00.071)       0:01:23.367 ****** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [keystone : Restart keystone-fernet container] *****************\nTuesday 24 February 2026  01:01:10 +0000 (0:00:12.578)       0:01:35.947 ****** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [keystone : Restart keystone container] ************************\nTuesday 24 February 2026  01:01:15 +0000 (0:00:05.419)       0:01:41.366 ****** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [keystone : include_tasks] ************************************************\nTuesday 24 February 2026  01:01:27 +0000 (0:00:11.600)       0:01:52.967 ****** \n\u001b[0;36mincluded: /ansible/roles/keystone/tasks/distribute_fernet.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [keystone : Waiting for Keystone SSH port to be UP] ***********************\nTuesday 24 February 2026  01:01:28 +0000 (0:00:00.824)       0:01:53.792 ****** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [keystone : Run key distribution] *****************************************\nTuesday 24 February 2026  01:01:29 +0000 (0:00:00.893)       0:01:54.686 ****** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [keystone : Creating admin project, user, role, service, and endpoint] ****\nTuesday 24 February 2026  01:01:31 +0000 (0:00:01.872)       0:01:56.559 ****** \n\u001b[0;33mchanged: [testbed-node-0] => (item=RegionOne)\u001b[0m\n\nTASK [service-ks-register : keystone | Creating services] **********************\nTuesday 24 February 2026  01:01:45 +0000 (0:00:14.151)       0:02:10.710 ****** \n\u001b[0;33mchanged: [testbed-node-0] => (item=keystone (identity))\u001b[0m\n\nTASK [service-ks-register : keystone | Creating endpoints] *********************\nTuesday 24 February 2026  01:02:13 +0000 (0:00:28.261)       0:02:38.972 ****** \n\u001b[0;32mok: [testbed-node-0] => (item=keystone -> https://api-int.testbed.osism.xyz:5000 -> internal)\u001b[0m\n\u001b[0;32mok: [testbed-node-0] => (item=keystone -> https://api.testbed.osism.xyz:5000 -> public)\u001b[0m\n\nTASK [service-ks-register : keystone | Creating projects] **********************\nTuesday 24 February 2026  01:02:21 +0000 (0:00:08.079)       0:02:47.051 ****** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [service-ks-register : keystone | Creating users] *************************\nTuesday 24 February 2026  01:02:21 +0000 (0:00:00.144)       0:02:47.196 ****** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [service-ks-register : keystone | Creating roles] *************************\nTuesday 24 February 2026  01:02:21 +0000 (0:00:00.124)       0:02:47.320 ****** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [service-ks-register : keystone | Granting user roles] ********************\nTuesday 24 February 2026  01:02:22 +0000 (0:00:00.131)       0:02:47.452 ****** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [keystone : Creating default user role] ***********************************\nTuesday 24 February 2026  01:02:22 +0000 (0:00:00.620)       0:02:48.072 ****** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [keystone : include_tasks] ************************************************\nTuesday 24 February 2026  01:02:26 +0000 (0:00:03.925)       0:02:51.997 ****** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=33  \u001b[0m \u001b[0;33mchanged=19  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=17  \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=22  \u001b[0m \u001b[0;33mchanged=12  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=12  \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=22  \u001b[0m \u001b[0;33mchanged=12  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=12  \u001b[0m rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nTuesday 24 February 2026  01:02:27 +0000 (0:00:00.461)       0:02:52.459 ****** \n=============================================================================== \nservice-ks-register : keystone | Creating services --------------------- 28.26s\nkeystone : Running Keystone bootstrap container ------------------------ 17.78s\nkeystone : Creating admin project, user, role, service, and endpoint --- 14.15s\nkeystone : Running Keystone fernet bootstrap container ----------------- 12.96s\nkeystone : Restart keystone-ssh container ------------------------------ 12.58s\nkeystone : Restart keystone container ---------------------------------- 11.60s\nkeystone : Copying files for keystone-fernet ---------------------------- 9.43s\nservice-ks-register : keystone | Creating endpoints --------------------- 8.08s\nkeystone : Copying over keystone.conf ----------------------------------- 6.10s\nkeystone : Restart keystone-fernet container ---------------------------- 5.42s\nkeystone : Creating default user role ----------------------------------- 3.93s\nservice-cert-copy : keystone | Copying over extra CA certificates ------- 3.54s\nkeystone : Copying over config.json files for services ------------------ 3.41s\nkeystone : Creating keystone database ----------------------------------- 2.94s\nkeystone : Copying files for keystone-ssh ------------------------------- 2.89s\nkeystone : Check keystone containers ------------------------------------ 2.78s\nkeystone : Creating Keystone database user and setting permissions ------ 2.69s\nkeystone : Ensuring config directories exist ---------------------------- 1.91s\nkeystone : Run key distribution ----------------------------------------- 1.87s\nkeystone : Copying over wsgi-keystone.conf ------------------------------ 1.73s\n\u001b[32m2026-02-24 01:02:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 81556691-758e-4eda-9a1d-3b717f8d6221 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:02:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 01:02:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f714ea82-7678-4e20-85bd-ee1502bb5463 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:02:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask beffba80-d510-4944-9e2b-b4ff3c4fc669 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:02:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ac05741a-1939-415f-aeff-a7eed21049c9 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:02:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 81556691-758e-4eda-9a1d-3b717f8d6221 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:02:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44aa611b-f898-4352-b174-ada638d8f758 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:02:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 01:02:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f714ea82-7678-4e20-85bd-ee1502bb5463 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:02:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask beffba80-d510-4944-9e2b-b4ff3c4fc669 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:02:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ac05741a-1939-415f-aeff-a7eed21049c9 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:02:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 81556691-758e-4eda-9a1d-3b717f8d6221 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:02:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44aa611b-f898-4352-b174-ada638d8f758 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:02:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 01:02:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f714ea82-7678-4e20-85bd-ee1502bb5463 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:02:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask beffba80-d510-4944-9e2b-b4ff3c4fc669 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:02:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ac05741a-1939-415f-aeff-a7eed21049c9 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:02:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 81556691-758e-4eda-9a1d-3b717f8d6221 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:02:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44aa611b-f898-4352-b174-ada638d8f758 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:02:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 01:02:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f714ea82-7678-4e20-85bd-ee1502bb5463 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:02:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask beffba80-d510-4944-9e2b-b4ff3c4fc669 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:02:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ac05741a-1939-415f-aeff-a7eed21049c9 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:02:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 81556691-758e-4eda-9a1d-3b717f8d6221 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:02:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44aa611b-f898-4352-b174-ada638d8f758 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:02:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 01:02:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f714ea82-7678-4e20-85bd-ee1502bb5463 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:02:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask beffba80-d510-4944-9e2b-b4ff3c4fc669 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:02:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ac05741a-1939-415f-aeff-a7eed21049c9 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:02:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 81556691-758e-4eda-9a1d-3b717f8d6221 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:02:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44aa611b-f898-4352-b174-ada638d8f758 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:02:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 01:02:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f714ea82-7678-4e20-85bd-ee1502bb5463 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:02:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask beffba80-d510-4944-9e2b-b4ff3c4fc669 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:02:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ac05741a-1939-415f-aeff-a7eed21049c9 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:02:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 81556691-758e-4eda-9a1d-3b717f8d6221 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:02:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44aa611b-f898-4352-b174-ada638d8f758 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:02:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 01:02:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f714ea82-7678-4e20-85bd-ee1502bb5463 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:02:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask beffba80-d510-4944-9e2b-b4ff3c4fc669 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:02:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ac05741a-1939-415f-aeff-a7eed21049c9 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:02:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 81556691-758e-4eda-9a1d-3b717f8d6221 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:02:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44aa611b-f898-4352-b174-ada638d8f758 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:02:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 01:02:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f714ea82-7678-4e20-85bd-ee1502bb5463 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:02:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask beffba80-d510-4944-9e2b-b4ff3c4fc669 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:02:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ac05741a-1939-415f-aeff-a7eed21049c9 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:02:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 81556691-758e-4eda-9a1d-3b717f8d6221 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:02:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44aa611b-f898-4352-b174-ada638d8f758 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:02:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 01:02:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f714ea82-7678-4e20-85bd-ee1502bb5463 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:02:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask beffba80-d510-4944-9e2b-b4ff3c4fc669 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:02:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ac05741a-1939-415f-aeff-a7eed21049c9 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:02:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 81556691-758e-4eda-9a1d-3b717f8d6221 is in state SUCCESS\u001b[0m\n\u001b[32m2026-02-24 01:02:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6fd9e63e-df8b-4c9f-86ca-d2316078b744 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:02:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44aa611b-f898-4352-b174-ada638d8f758 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:02:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 01:02:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f714ea82-7678-4e20-85bd-ee1502bb5463 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:02:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask beffba80-d510-4944-9e2b-b4ff3c4fc669 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:02:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ac05741a-1939-415f-aeff-a7eed21049c9 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:02:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6fd9e63e-df8b-4c9f-86ca-d2316078b744 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:02:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44aa611b-f898-4352-b174-ada638d8f758 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:02:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 01:03:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f714ea82-7678-4e20-85bd-ee1502bb5463 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:03:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask beffba80-d510-4944-9e2b-b4ff3c4fc669 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:03:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ac05741a-1939-415f-aeff-a7eed21049c9 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:03:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6fd9e63e-df8b-4c9f-86ca-d2316078b744 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:03:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44aa611b-f898-4352-b174-ada638d8f758 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:03:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 01:03:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f714ea82-7678-4e20-85bd-ee1502bb5463 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:03:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask beffba80-d510-4944-9e2b-b4ff3c4fc669 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:03:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ac05741a-1939-415f-aeff-a7eed21049c9 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:03:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6fd9e63e-df8b-4c9f-86ca-d2316078b744 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:03:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44aa611b-f898-4352-b174-ada638d8f758 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:03:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 01:03:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f714ea82-7678-4e20-85bd-ee1502bb5463 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:03:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask beffba80-d510-4944-9e2b-b4ff3c4fc669 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:03:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ac05741a-1939-415f-aeff-a7eed21049c9 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:03:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6fd9e63e-df8b-4c9f-86ca-d2316078b744 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:03:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44aa611b-f898-4352-b174-ada638d8f758 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:03:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 01:03:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f714ea82-7678-4e20-85bd-ee1502bb5463 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:03:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask beffba80-d510-4944-9e2b-b4ff3c4fc669 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:03:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ac05741a-1939-415f-aeff-a7eed21049c9 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:03:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6fd9e63e-df8b-4c9f-86ca-d2316078b744 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:03:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44aa611b-f898-4352-b174-ada638d8f758 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:03:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 01:03:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f714ea82-7678-4e20-85bd-ee1502bb5463 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:03:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask beffba80-d510-4944-9e2b-b4ff3c4fc669 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:03:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ac05741a-1939-415f-aeff-a7eed21049c9 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:03:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6fd9e63e-df8b-4c9f-86ca-d2316078b744 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:03:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44aa611b-f898-4352-b174-ada638d8f758 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:03:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 01:03:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f714ea82-7678-4e20-85bd-ee1502bb5463 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:03:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask beffba80-d510-4944-9e2b-b4ff3c4fc669 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:03:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ac05741a-1939-415f-aeff-a7eed21049c9 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:03:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6fd9e63e-df8b-4c9f-86ca-d2316078b744 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:03:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44aa611b-f898-4352-b174-ada638d8f758 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:03:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 01:03:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f714ea82-7678-4e20-85bd-ee1502bb5463 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:03:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask beffba80-d510-4944-9e2b-b4ff3c4fc669 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:03:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ac05741a-1939-415f-aeff-a7eed21049c9 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:03:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6fd9e63e-df8b-4c9f-86ca-d2316078b744 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:03:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44aa611b-f898-4352-b174-ada638d8f758 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:03:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 01:03:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f714ea82-7678-4e20-85bd-ee1502bb5463 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:03:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask beffba80-d510-4944-9e2b-b4ff3c4fc669 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:03:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ac05741a-1939-415f-aeff-a7eed21049c9 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:03:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6fd9e63e-df8b-4c9f-86ca-d2316078b744 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:03:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44aa611b-f898-4352-b174-ada638d8f758 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:03:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 01:03:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f714ea82-7678-4e20-85bd-ee1502bb5463 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:03:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask beffba80-d510-4944-9e2b-b4ff3c4fc669 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:03:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ac05741a-1939-415f-aeff-a7eed21049c9 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:03:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6fd9e63e-df8b-4c9f-86ca-d2316078b744 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:03:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44aa611b-f898-4352-b174-ada638d8f758 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:03:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 01:03:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f714ea82-7678-4e20-85bd-ee1502bb5463 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:03:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask beffba80-d510-4944-9e2b-b4ff3c4fc669 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:03:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ac05741a-1939-415f-aeff-a7eed21049c9 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:03:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6fd9e63e-df8b-4c9f-86ca-d2316078b744 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:03:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44aa611b-f898-4352-b174-ada638d8f758 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:03:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 01:03:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f714ea82-7678-4e20-85bd-ee1502bb5463 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:03:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask beffba80-d510-4944-9e2b-b4ff3c4fc669 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:03:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ac05741a-1939-415f-aeff-a7eed21049c9 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:03:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6fd9e63e-df8b-4c9f-86ca-d2316078b744 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:03:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44aa611b-f898-4352-b174-ada638d8f758 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:03:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 01:03:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f714ea82-7678-4e20-85bd-ee1502bb5463 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:03:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask beffba80-d510-4944-9e2b-b4ff3c4fc669 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:03:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ac05741a-1939-415f-aeff-a7eed21049c9 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:03:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6fd9e63e-df8b-4c9f-86ca-d2316078b744 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:03:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44aa611b-f898-4352-b174-ada638d8f758 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:03:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 01:03:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f714ea82-7678-4e20-85bd-ee1502bb5463 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:03:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask beffba80-d510-4944-9e2b-b4ff3c4fc669 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:03:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ac05741a-1939-415f-aeff-a7eed21049c9 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:03:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6fd9e63e-df8b-4c9f-86ca-d2316078b744 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:03:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44aa611b-f898-4352-b174-ada638d8f758 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:03:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 01:03:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f714ea82-7678-4e20-85bd-ee1502bb5463 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:03:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask beffba80-d510-4944-9e2b-b4ff3c4fc669 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:03:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ac05741a-1939-415f-aeff-a7eed21049c9 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:03:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6fd9e63e-df8b-4c9f-86ca-d2316078b744 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:03:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44aa611b-f898-4352-b174-ada638d8f758 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:03:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 01:03:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f714ea82-7678-4e20-85bd-ee1502bb5463 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:03:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask beffba80-d510-4944-9e2b-b4ff3c4fc669 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:03:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ac05741a-1939-415f-aeff-a7eed21049c9 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:03:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6fd9e63e-df8b-4c9f-86ca-d2316078b744 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:03:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44aa611b-f898-4352-b174-ada638d8f758 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:03:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 01:03:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f714ea82-7678-4e20-85bd-ee1502bb5463 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:03:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask beffba80-d510-4944-9e2b-b4ff3c4fc669 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:03:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ac05741a-1939-415f-aeff-a7eed21049c9 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:03:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6fd9e63e-df8b-4c9f-86ca-d2316078b744 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:03:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44aa611b-f898-4352-b174-ada638d8f758 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:03:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 01:03:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f714ea82-7678-4e20-85bd-ee1502bb5463 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:03:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask beffba80-d510-4944-9e2b-b4ff3c4fc669 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:03:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ac05741a-1939-415f-aeff-a7eed21049c9 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:03:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6fd9e63e-df8b-4c9f-86ca-d2316078b744 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:03:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44aa611b-f898-4352-b174-ada638d8f758 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:03:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 01:03:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f714ea82-7678-4e20-85bd-ee1502bb5463 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:03:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask beffba80-d510-4944-9e2b-b4ff3c4fc669 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:03:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ac05741a-1939-415f-aeff-a7eed21049c9 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:03:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6fd9e63e-df8b-4c9f-86ca-d2316078b744 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:03:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44aa611b-f898-4352-b174-ada638d8f758 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:03:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 01:03:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f714ea82-7678-4e20-85bd-ee1502bb5463 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:03:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask beffba80-d510-4944-9e2b-b4ff3c4fc669 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:03:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ac05741a-1939-415f-aeff-a7eed21049c9 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:03:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6fd9e63e-df8b-4c9f-86ca-d2316078b744 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:03:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44aa611b-f898-4352-b174-ada638d8f758 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:03:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 01:03:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f714ea82-7678-4e20-85bd-ee1502bb5463 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:03:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask beffba80-d510-4944-9e2b-b4ff3c4fc669 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:03:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ac05741a-1939-415f-aeff-a7eed21049c9 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:03:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6fd9e63e-df8b-4c9f-86ca-d2316078b744 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:03:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44aa611b-f898-4352-b174-ada638d8f758 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:03:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 01:04:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f714ea82-7678-4e20-85bd-ee1502bb5463 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:04:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask beffba80-d510-4944-9e2b-b4ff3c4fc669 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:04:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ac05741a-1939-415f-aeff-a7eed21049c9 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:04:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6fd9e63e-df8b-4c9f-86ca-d2316078b744 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:04:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44aa611b-f898-4352-b174-ada638d8f758 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:04:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 01:04:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f714ea82-7678-4e20-85bd-ee1502bb5463 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:04:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask beffba80-d510-4944-9e2b-b4ff3c4fc669 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:04:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ac05741a-1939-415f-aeff-a7eed21049c9 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:04:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6fd9e63e-df8b-4c9f-86ca-d2316078b744 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:04:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44aa611b-f898-4352-b174-ada638d8f758 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:04:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 01:04:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f714ea82-7678-4e20-85bd-ee1502bb5463 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:04:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask beffba80-d510-4944-9e2b-b4ff3c4fc669 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:04:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ac05741a-1939-415f-aeff-a7eed21049c9 is in state SUCCESS\u001b[0m\n\u001b[32m2026-02-24 01:04:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6fd9e63e-df8b-4c9f-86ca-d2316078b744 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:04:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44aa611b-f898-4352-b174-ada638d8f758 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:04:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3046662f-74e4-408a-bb14-4c7f7919dd4b is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:04:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 01:04:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f714ea82-7678-4e20-85bd-ee1502bb5463 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:04:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask beffba80-d510-4944-9e2b-b4ff3c4fc669 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:04:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6fd9e63e-df8b-4c9f-86ca-d2316078b744 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:04:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44aa611b-f898-4352-b174-ada638d8f758 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:04:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3046662f-74e4-408a-bb14-4c7f7919dd4b is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:04:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 01:04:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f714ea82-7678-4e20-85bd-ee1502bb5463 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:04:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask beffba80-d510-4944-9e2b-b4ff3c4fc669 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:04:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6fd9e63e-df8b-4c9f-86ca-d2316078b744 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:04:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44aa611b-f898-4352-b174-ada638d8f758 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:04:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3046662f-74e4-408a-bb14-4c7f7919dd4b is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:04:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 01:04:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f714ea82-7678-4e20-85bd-ee1502bb5463 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:04:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask beffba80-d510-4944-9e2b-b4ff3c4fc669 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:04:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6fd9e63e-df8b-4c9f-86ca-d2316078b744 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:04:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44aa611b-f898-4352-b174-ada638d8f758 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:04:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3046662f-74e4-408a-bb14-4c7f7919dd4b is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:04:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 01:04:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f714ea82-7678-4e20-85bd-ee1502bb5463 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:04:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask beffba80-d510-4944-9e2b-b4ff3c4fc669 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:04:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6fd9e63e-df8b-4c9f-86ca-d2316078b744 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:04:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44aa611b-f898-4352-b174-ada638d8f758 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:04:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3046662f-74e4-408a-bb14-4c7f7919dd4b is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:04:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 01:04:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f714ea82-7678-4e20-85bd-ee1502bb5463 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:04:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask beffba80-d510-4944-9e2b-b4ff3c4fc669 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:04:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6fd9e63e-df8b-4c9f-86ca-d2316078b744 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:04:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44aa611b-f898-4352-b174-ada638d8f758 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:04:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3046662f-74e4-408a-bb14-4c7f7919dd4b is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:04:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 01:04:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f714ea82-7678-4e20-85bd-ee1502bb5463 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:04:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask beffba80-d510-4944-9e2b-b4ff3c4fc669 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:04:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6fd9e63e-df8b-4c9f-86ca-d2316078b744 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:04:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44aa611b-f898-4352-b174-ada638d8f758 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:04:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3046662f-74e4-408a-bb14-4c7f7919dd4b is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:04:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 01:04:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f714ea82-7678-4e20-85bd-ee1502bb5463 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:04:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask beffba80-d510-4944-9e2b-b4ff3c4fc669 is in state STARTED\u001b[0m\n\n\nPLAY [Apply role cephclient] ***************************************************\n\nTASK [osism.services.cephclient : Include container tasks] *********************\nTuesday 24 February 2026  01:01:54 +0000 (0:00:00.271)       0:00:00.271 ****** \n\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/services/roles/cephclient/tasks/container.yml for testbed-manager\u001b[0m\n\nTASK [osism.services.cephclient : Create required directories] *****************\nTuesday 24 February 2026  01:01:54 +0000 (0:00:00.238)       0:00:00.510 ****** \n\u001b[0;33mchanged: [testbed-manager] => (item=/opt/cephclient/configuration)\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item=/opt/cephclient/data)\u001b[0m\n\u001b[0;32mok: [testbed-manager] => (item=/opt/cephclient)\u001b[0m\n\nTASK [osism.services.cephclient : Copy configuration files] ********************\nTuesday 24 February 2026  01:01:56 +0000 (0:00:01.377)       0:00:01.887 ****** \n\u001b[0;33mchanged: [testbed-manager] => (item={'src': 'ceph.conf.j2', 'dest': '/opt/cephclient/configuration/ceph.conf'})\u001b[0m\n\nTASK [osism.services.cephclient : Copy keyring file] ***************************\nTuesday 24 February 2026  01:01:57 +0000 (0:00:01.635)       0:00:03.523 ****** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nTASK [osism.services.cephclient : Copy docker-compose.yml file] ****************\nTuesday 24 February 2026  01:01:58 +0000 (0:00:00.993)       0:00:04.516 ****** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nTASK [osism.services.cephclient : Manage cephclient service] *******************\nTuesday 24 February 2026  01:01:59 +0000 (0:00:01.028)       0:00:05.545 ****** \n\u001b[1;30mFAILED - RETRYING: [testbed-manager]: Manage cephclient service (10 retries left).\u001b[0m\n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\nTASK [osism.services.cephclient : Copy wrapper scripts] ************************\nTuesday 24 February 2026  01:02:42 +0000 (0:00:42.879)       0:00:48.424 ****** \n\u001b[0;33mchanged: [testbed-manager] => (item=ceph)\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item=ceph-authtool)\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item=rados)\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item=radosgw-admin)\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item=rbd)\u001b[0m\n\nTASK [osism.services.cephclient : Remove old wrapper scripts] ******************\nTuesday 24 February 2026  01:02:47 +0000 (0:00:04.646)       0:00:53.071 ****** \n\u001b[0;32mok: [testbed-manager] => (item=crushtool)\u001b[0m\n\nTASK [osism.services.cephclient : Include package tasks] ***********************\nTuesday 24 February 2026  01:02:47 +0000 (0:00:00.511)       0:00:53.583 ****** \n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\nTASK [osism.services.cephclient : Include rook task] ***************************\nTuesday 24 February 2026  01:02:47 +0000 (0:00:00.144)       0:00:53.727 ****** \n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\nRUNNING HANDLER [osism.services.cephclient : Restart cephclient service] *******\nTuesday 24 February 2026  01:02:48 +0000 (0:00:00.596)       0:00:54.324 ****** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nRUNNING HANDLER [osism.services.cephclient : Ensure that all containers are up] ***\nTuesday 24 February 2026  01:02:50 +0000 (0:00:01.671)       0:00:55.996 ****** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nRUNNING HANDLER [osism.services.cephclient : Wait for an healthy service] ******\nTuesday 24 February 2026  01:02:51 +0000 (0:00:00.842)       0:00:56.838 ****** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nRUNNING HANDLER [osism.services.cephclient : Copy bash completion scripts] *****\nTuesday 24 February 2026  01:02:51 +0000 (0:00:00.712)       0:00:57.551 ****** \n\u001b[0;32mok: [testbed-manager] => (item=ceph)\u001b[0m\n\u001b[0;32mok: [testbed-manager] => (item=rados)\u001b[0m\n\u001b[0;32mok: [testbed-manager] => (item=radosgw-admin)\u001b[0m\n\u001b[0;32mok: [testbed-manager] => (item=rbd)\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;33mtestbed-manager\u001b[0m            : \u001b[0;32mok=12  \u001b[0m \u001b[0;33mchanged=8   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nTuesday 24 February 2026  01:02:53 +0000 (0:00:01.707)       0:00:59.258 ****** \n=============================================================================== \nosism.services.cephclient : Manage cephclient service ------------------ 42.88s\nosism.services.cephclient : Copy wrapper scripts ------------------------ 4.65s\nosism.services.cephclient : Copy bash completion scripts ---------------- 1.71s\nosism.services.cephclient : Restart cephclient service ------------------ 1.67s\nosism.services.cephclient : Copy configuration files -------------------- 1.64s\nosism.services.cephclient : Create required directories ----------------- 1.38s\nosism.services.cephclient : Copy docker-compose.yml file ---------------- 1.03s\nosism.services.cephclient : Copy keyring file --------------------------- 0.99s\nosism.services.cephclient : Ensure that all containers are up ----------- 0.84s\nosism.services.cephclient : Wait for an healthy service ----------------- 0.71s\nosism.services.cephclient : Include rook task --------------------------- 0.60s\nosism.services.cephclient : Remove old wrapper scripts ------------------ 0.51s\nosism.services.cephclient : Include container tasks --------------------- 0.24s\nosism.services.cephclient : Include package tasks ----------------------- 0.14s\n\n\nPLAY [Download ironic ipa images] **********************************************\n\nTASK [Ensure the destination directory exists] *********************************\nTuesday 24 February 2026  01:02:32 +0000 (0:00:00.086)       0:00:00.086 ****** \n\u001b[0;33mchanged: [localhost]\u001b[0m\n\nTASK [Download ironic-agent initramfs] *****************************************\nTuesday 24 February 2026  01:02:32 +0000 (0:00:00.851)       0:00:00.938 ****** \n\u001b[1;30mFAILED - RETRYING: [localhost]: Download ironic-agent initramfs (3 retries left).\u001b[0m\n\u001b[1;30mFAILED - RETRYING: [localhost]: Download ironic-agent initramfs (2 retries left).\u001b[0m\n\u001b[1;30mFAILED - RETRYING: [localhost]: Download ironic-agent initramfs (1 retries left).\u001b[0m\n\u001b[0;31mfatal: [localhost]: FAILED! => {\"attempts\": 3, \"changed\": false, \"dest\": \"/share/ironic/ironic/ironic-agent.initramfs\", \"elapsed\": 10, \"msg\": \"Request failed: <urlopen error [Errno 101] Network is unreachable>\", \"url\": \"https://tarballs.opendev.org/openstack/ironic-python-agent/dib/files/ipa-centos9-stable-2024.2.initramfs.sha256\"}\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;31mlocalhost\u001b[0m                  : \u001b[0;32mok=1   \u001b[0m \u001b[0;33mchanged=1   \u001b[0m unreachable=0    \u001b[0;31mfailed=1   \u001b[0m skipped=0    rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nTuesday 24 February 2026  01:04:07 +0000 (0:01:34.415)       0:01:35.354 ****** \n=============================================================================== \nDownload ironic-agent initramfs ---------------------------------------- 94.42s\nEnsure the destination directory exists --------------------------------- 0.85s\n\n\u001b[1;35m[WARNING]: Collection community.general does not support Ansible version\u001b[0m\n\u001b[1;35m2.16.14\u001b[0m\n\nPLAY [Bootstraph ceph dashboard] ***********************************************\n\nTASK [Disable the ceph dashboard] **********************************************\nTuesday 24 February 2026  01:02:58 +0000 (0:00:00.279)       0:00:00.279 ****** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nTASK [Set mgr/dashboard/ssl to false] ******************************************\nTuesday 24 February 2026  01:03:00 +0000 (0:00:01.654)       0:00:01.934 ****** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nTASK [Set mgr/dashboard/server_port to 7000] ***********************************\nTuesday 24 February 2026  01:03:01 +0000 (0:00:01.214)       0:00:03.149 ****** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nTASK [Set mgr/dashboard/server_addr to 0.0.0.0] ********************************\nTuesday 24 February 2026  01:03:02 +0000 (0:00:01.129)       0:00:04.279 ****** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nTASK [Set mgr/dashboard/standby_behaviour to error] ****************************\nTuesday 24 February 2026  01:03:04 +0000 (0:00:01.368)       0:00:05.647 ****** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nTASK [Set mgr/dashboard/standby_error_status_code to 404] **********************\nTuesday 24 February 2026  01:03:05 +0000 (0:00:01.191)       0:00:06.839 ****** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nTASK [Enable the ceph dashboard] ***********************************************\nTuesday 24 February 2026  01:03:06 +0000 (0:00:01.198)       0:00:08.038 ****** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nTASK [Write ceph_dashboard_password to temporary file] *************************\nTuesday 24 February 2026  01:03:08 +0000 (0:00:02.110)       0:00:10.148 ****** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nTASK [Create admin user] *******************************************************\nTuesday 24 February 2026  01:03:10 +0000 (0:00:01.431)       0:00:11.580 ****** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nTASK [Remove temporary file for ceph_dashboard_password] ***********************\nTuesday 24 February 2026  01:04:05 +0000 (0:00:54.942)       0:01:06.522 ****** \n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\nPLAY [Restart ceph manager services] *******************************************\n\nTASK [Restart ceph manager service] ********************************************\nTuesday 24 February 2026  01:04:05 +0000 (0:00:00.157)       0:01:06.680 ****** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nPLAY [Restart ceph manager services] *******************************************\n\nTASK [Restart ceph manager service] ********************************************\nTuesday 24 February 2026  01:04:17 +0000 (0:00:11.863)       0:01:18.544 ****** \n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\nPLAY [Restart ceph manager services] *******************************************\n\nTASK [Restart ceph manager service] ********************************************\nTuesday 24 February 2026  01:04:28 +0000 (0:00:11.288)       0:01:29.832 ****** \n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;33mtestbed-manager\u001b[0m            : \u001b[0;32mok=9   \u001b[0m \u001b[0;33mchanged=9   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=1   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=1   \u001b[0m \u001b[0;33mchanged=1   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=1   \u001b[0m \u001b[0;33mchanged=1   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=1   \u001b[0m \u001b[0;33mchanged=1   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\n\n\nTASKS RECAP ********************************************************************\nTuesday 24 February 2026  01:04:29 +0000 (0:00:01.125)       0:01:30.957 ****** \n=============================================================================== \nCreate admin user ------------------------------------------------------ 54.94s\nRestart ceph manager service ------------------------------------------- 24.28s\nEnable the ceph dashboard ----------------------------------------------- 2.11s\nDisable the ceph dashboard ---------------------------------------------- 1.65s\nWrite ceph_dashboard_password to temporary file ------------------------- 1.43s\nSet mgr/dashboard/server_addr to 0.0.0.0 -------------------------------- 1.37s\nSet mgr/dashboard/ssl to false ------------------------------------------ 1.21s\nSet mgr/dashboard/standby_error_status_code to 404 ---------------------- 1.20s\nSet mgr/dashboard/standby_behaviour to error ---------------------------- 1.19s\nSet mgr/dashboard/server_port to 7000 ----------------------------------- 1.13s\nRemove temporary file for ceph_dashboard_password ----------------------- 0.16s\n\u001b[32m2026-02-24 01:04:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6fd9e63e-df8b-4c9f-86ca-d2316078b744 is in state SUCCESS\u001b[0m\n\u001b[32m2026-02-24 01:04:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44aa611b-f898-4352-b174-ada638d8f758 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:04:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3046662f-74e4-408a-bb14-4c7f7919dd4b is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:04:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 01:04:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f714ea82-7678-4e20-85bd-ee1502bb5463 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:04:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask beffba80-d510-4944-9e2b-b4ff3c4fc669 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:04:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44aa611b-f898-4352-b174-ada638d8f758 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:04:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3046662f-74e4-408a-bb14-4c7f7919dd4b is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:04:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 01:04:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f714ea82-7678-4e20-85bd-ee1502bb5463 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:04:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask beffba80-d510-4944-9e2b-b4ff3c4fc669 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:04:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44aa611b-f898-4352-b174-ada638d8f758 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:04:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3046662f-74e4-408a-bb14-4c7f7919dd4b is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:04:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 01:04:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f714ea82-7678-4e20-85bd-ee1502bb5463 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:04:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask beffba80-d510-4944-9e2b-b4ff3c4fc669 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:04:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44aa611b-f898-4352-b174-ada638d8f758 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:04:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3046662f-74e4-408a-bb14-4c7f7919dd4b is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:04:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 01:04:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f714ea82-7678-4e20-85bd-ee1502bb5463 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:04:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask beffba80-d510-4944-9e2b-b4ff3c4fc669 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:04:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44aa611b-f898-4352-b174-ada638d8f758 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:04:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3046662f-74e4-408a-bb14-4c7f7919dd4b is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:04:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 01:04:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f714ea82-7678-4e20-85bd-ee1502bb5463 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:04:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask beffba80-d510-4944-9e2b-b4ff3c4fc669 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:04:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44aa611b-f898-4352-b174-ada638d8f758 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:04:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3046662f-74e4-408a-bb14-4c7f7919dd4b is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:04:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 01:04:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f714ea82-7678-4e20-85bd-ee1502bb5463 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:04:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask beffba80-d510-4944-9e2b-b4ff3c4fc669 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:04:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44aa611b-f898-4352-b174-ada638d8f758 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:04:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3046662f-74e4-408a-bb14-4c7f7919dd4b is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:04:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 01:04:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f714ea82-7678-4e20-85bd-ee1502bb5463 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:04:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask beffba80-d510-4944-9e2b-b4ff3c4fc669 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:04:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44aa611b-f898-4352-b174-ada638d8f758 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:04:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3046662f-74e4-408a-bb14-4c7f7919dd4b is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:04:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 01:04:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f714ea82-7678-4e20-85bd-ee1502bb5463 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:04:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask beffba80-d510-4944-9e2b-b4ff3c4fc669 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:04:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44aa611b-f898-4352-b174-ada638d8f758 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:04:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3046662f-74e4-408a-bb14-4c7f7919dd4b is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:04:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 01:04:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f714ea82-7678-4e20-85bd-ee1502bb5463 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:04:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask beffba80-d510-4944-9e2b-b4ff3c4fc669 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:04:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44aa611b-f898-4352-b174-ada638d8f758 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:04:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3046662f-74e4-408a-bb14-4c7f7919dd4b is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:04:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 01:05:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f714ea82-7678-4e20-85bd-ee1502bb5463 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:05:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask beffba80-d510-4944-9e2b-b4ff3c4fc669 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:05:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44aa611b-f898-4352-b174-ada638d8f758 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:05:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3046662f-74e4-408a-bb14-4c7f7919dd4b is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:05:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 01:05:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f714ea82-7678-4e20-85bd-ee1502bb5463 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:05:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask beffba80-d510-4944-9e2b-b4ff3c4fc669 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:05:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44aa611b-f898-4352-b174-ada638d8f758 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:05:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3046662f-74e4-408a-bb14-4c7f7919dd4b is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:05:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 01:05:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f714ea82-7678-4e20-85bd-ee1502bb5463 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:05:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask beffba80-d510-4944-9e2b-b4ff3c4fc669 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:05:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44aa611b-f898-4352-b174-ada638d8f758 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:05:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3046662f-74e4-408a-bb14-4c7f7919dd4b is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:05:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 01:05:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f714ea82-7678-4e20-85bd-ee1502bb5463 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:05:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask beffba80-d510-4944-9e2b-b4ff3c4fc669 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:05:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44aa611b-f898-4352-b174-ada638d8f758 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:05:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3046662f-74e4-408a-bb14-4c7f7919dd4b is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:05:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 01:05:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f714ea82-7678-4e20-85bd-ee1502bb5463 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:05:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask beffba80-d510-4944-9e2b-b4ff3c4fc669 is in state SUCCESS\u001b[0m\n\n\nPLAY [Group hosts based on configuration] **************************************\n\nTASK [Group hosts based on Kolla action] ***************************************\nTuesday 24 February 2026  01:02:32 +0000 (0:00:00.378)       0:00:00.378 ****** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [Group hosts based on enabled services] ***********************************\nTuesday 24 February 2026  01:02:32 +0000 (0:00:00.320)       0:00:00.698 ****** \n\u001b[0;32mok: [testbed-node-0] => (item=enable_barbican_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item=enable_barbican_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item=enable_barbican_True)\u001b[0m\n\nPLAY [Apply role barbican] *****************************************************\n\nTASK [barbican : include_tasks] ************************************************\nTuesday 24 February 2026  01:02:33 +0000 (0:00:00.503)       0:00:01.202 ****** \n\u001b[0;36mincluded: /ansible/roles/barbican/tasks/deploy.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [service-ks-register : barbican | Creating services] **********************\nTuesday 24 February 2026  01:02:33 +0000 (0:00:00.522)       0:00:01.725 ****** \n\u001b[0;33mchanged: [testbed-node-0] => (item=barbican (key-manager))\u001b[0m\n\nTASK [service-ks-register : barbican | Creating endpoints] *********************\nTuesday 24 February 2026  01:02:37 +0000 (0:00:03.782)       0:00:05.507 ****** \n\u001b[0;33mchanged: [testbed-node-0] => (item=barbican -> https://api-int.testbed.osism.xyz:9311 -> internal)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=barbican -> https://api.testbed.osism.xyz:9311 -> public)\u001b[0m\n\nTASK [service-ks-register : barbican | Creating projects] **********************\nTuesday 24 February 2026  01:02:46 +0000 (0:00:08.751)       0:00:14.258 ****** \n\u001b[1;30mFAILED - RETRYING: [testbed-node-0]: barbican | Creating projects (5 retries left).\u001b[0m\n\u001b[0;32mok: [testbed-node-0] => (item=service)\u001b[0m\n\nTASK [service-ks-register : barbican | Creating users] *************************\nTuesday 24 February 2026  01:03:04 +0000 (0:00:18.156)       0:00:32.415 ****** \n\u001b[0;33mchanged: [testbed-node-0] => (item=barbican -> service)\u001b[0m\n\u001b[1;35m[WARNING]: Module did not set no_log for update_password\u001b[0m\n\nTASK [service-ks-register : barbican | Creating roles] *************************\nTuesday 24 February 2026  01:03:09 +0000 (0:00:04.862)       0:00:37.278 ****** \n\u001b[0;32mok: [testbed-node-0] => (item=admin)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=key-manager:service-admin)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=creator)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=observer)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=audit)\u001b[0m\n\nTASK [service-ks-register : barbican | Granting user roles] ********************\nTuesday 24 February 2026  01:03:26 +0000 (0:00:16.594)       0:00:53.873 ****** \n\u001b[0;33mchanged: [testbed-node-0] => (item=barbican -> service -> admin)\u001b[0m\n\nTASK [barbican : Ensuring config directories exist] ****************************\nTuesday 24 February 2026  01:03:31 +0000 (0:00:05.327)       0:00:59.200 ****** \n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-api:2024.2', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-api:2024.2', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-api:2024.2', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-keystone-listener:2024.2', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-keystone-listener:2024.2', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-keystone-listener:2024.2', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-worker:2024.2', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-worker:2024.2', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-worker:2024.2', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}})\u001b[0m\n\nTASK [barbican : Ensuring vassals config directories exist] ********************\nTuesday 24 February 2026  01:03:34 +0000 (0:00:03.309)       0:01:02.510 ****** \n\u001b[0;33mchanged: [testbed-node-0] => (item=barbican-api/vassals)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=barbican-api/vassals)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=barbican-api/vassals)\u001b[0m\n\nTASK [barbican : Check if policies shall be overwritten] ***********************\nTuesday 24 February 2026  01:03:36 +0000 (0:00:01.719)       0:01:04.229 ****** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [barbican : Set barbican policy file] *************************************\nTuesday 24 February 2026  01:03:36 +0000 (0:00:00.472)       0:01:04.702 ****** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [barbican : include_tasks] ************************************************\nTuesday 24 February 2026  01:03:38 +0000 (0:00:01.433)       0:01:06.135 ****** \n\u001b[0;36mincluded: /ansible/roles/barbican/tasks/copy-certs.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [service-cert-copy : barbican | Copying over extra CA certificates] *******\nTuesday 24 February 2026  01:03:38 +0000 (0:00:00.646)       0:01:06.782 ****** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-api:2024.2', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-api:2024.2', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-api:2024.2', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-keystone-listener:2024.2', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-keystone-listener:2024.2', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-keystone-listener:2024.2', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-worker:2024.2', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-worker:2024.2', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-worker:2024.2', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}})\u001b[0m\n\nTASK [service-cert-copy : barbican | Copying over backend internal TLS certificate] ***\nTuesday 24 February 2026  01:03:43 +0000 (0:00:04.227)       0:01:11.009 ****** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-api:2024.2', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-keystone-listener:2024.2', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-worker:2024.2', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-api:2024.2', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-keystone-listener:2024.2', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-worker:2024.2', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-api:2024.2', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-keystone-listener:2024.2', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-worker:2024.2', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [service-cert-copy : barbican | Copying over backend internal TLS key] ****\nTuesday 24 February 2026  01:03:44 +0000 (0:00:01.494)       0:01:12.503 ****** \n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-api:2024.2', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-keystone-listener:2024.2', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-worker:2024.2', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-api:2024.2', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-keystone-listener:2024.2', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-worker:2024.2', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-api:2024.2', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-keystone-listener:2024.2', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-worker:2024.2', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [barbican : Copying over config.json files for services] ******************\nTuesday 24 February 2026  01:03:46 +0000 (0:00:02.048)       0:01:14.551 ****** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-api:2024.2', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-api:2024.2', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-api:2024.2', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-keystone-listener:2024.2', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-keystone-listener:2024.2', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-keystone-listener:2024.2', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-worker:2024.2', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-worker:2024.2', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-worker:2024.2', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}})\u001b[0m\n\nTASK [barbican : Copying over barbican-api.ini] ********************************\nTuesday 24 February 2026  01:03:50 +0000 (0:00:03.899)       0:01:18.451 ****** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [barbican : Checking whether barbican-api-paste.ini file exists] **********\nTuesday 24 February 2026  01:03:54 +0000 (0:00:03.470)       0:01:21.922 ****** \n\u001b[0;32mok: [testbed-node-0 -> localhost]\u001b[0m\n\nTASK [barbican : Copying over barbican-api-paste.ini] **************************\nTuesday 24 February 2026  01:03:56 +0000 (0:00:02.912)       0:01:24.835 ****** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [barbican : Copying over barbican.conf] ***********************************\nTuesday 24 February 2026  01:03:58 +0000 (0:00:01.175)       0:01:26.010 ****** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-api:2024.2', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-api:2024.2', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-api:2024.2', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-keystone-listener:2024.2', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-keystone-listener:2024.2', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-keystone-listener:2024.2', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-worker:2024.2', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-worker:2024.2', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-worker:2024.2', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}})\u001b[0m\n\nTASK [barbican : Copying over existing policy file] ****************************\nTuesday 24 February 2026  01:04:07 +0000 (0:00:09.742)       0:01:35.752 ****** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-api:2024.2', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-keystone-listener:2024.2', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-worker:2024.2', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-api:2024.2', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-keystone-listener:2024.2', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-worker:2024.2', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-api:2024.2', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-keystone-listener:2024.2', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-worker:2024.2', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [barbican : Check barbican containers] ************************************\nTuesday 24 February 2026  01:04:10 +0000 (0:00:02.225)       0:01:37.978 ****** \n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-api:2024.2', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-api:2024.2', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-api:2024.2', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-keystone-listener:2024.2', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-keystone-listener:2024.2', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-keystone-listener:2024.2', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-worker:2024.2', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-worker:2024.2', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-worker:2024.2', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}})\u001b[0m\n\nTASK [barbican : include_tasks] ************************************************\nTuesday 24 February 2026  01:04:16 +0000 (0:00:06.884)       0:01:44.863 ****** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [barbican : Creating barbican database] ***********************************\nTuesday 24 February 2026  01:04:18 +0000 (0:00:01.027)       0:01:45.890 ****** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [barbican : Creating barbican database user and setting permissions] ******\nTuesday 24 February 2026  01:04:20 +0000 (0:00:02.964)       0:01:48.855 ****** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [barbican : Running barbican bootstrap container] *************************\nTuesday 24 February 2026  01:04:24 +0000 (0:00:03.481)       0:01:52.337 ****** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [barbican : Flush handlers] ***********************************************\nTuesday 24 February 2026  01:04:37 +0000 (0:00:12.888)       0:02:05.225 ****** \n\nTASK [barbican : Flush handlers] ***********************************************\nTuesday 24 February 2026  01:04:37 +0000 (0:00:00.152)       0:02:05.378 ****** \n\nTASK [barbican : Flush handlers] ***********************************************\nTuesday 24 February 2026  01:04:37 +0000 (0:00:00.125)       0:02:05.503 ****** \n\nRUNNING HANDLER [barbican : Restart barbican-api container] ********************\nTuesday 24 February 2026  01:04:37 +0000 (0:00:00.067)       0:02:05.570 ****** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [barbican : Restart barbican-keystone-listener container] ******\nTuesday 24 February 2026  01:04:51 +0000 (0:00:13.707)       0:02:19.278 ****** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\nRUNNING HANDLER [barbican : Restart barbican-worker container] *****************\nTuesday 24 February 2026  01:05:01 +0000 (0:00:10.325)       0:02:29.604 ****** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=24  \u001b[0m \u001b[0;33mchanged=18  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=7   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=14  \u001b[0m \u001b[0;33mchanged=10  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=6   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=14  \u001b[0m \u001b[0;33mchanged=10  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=6   \u001b[0m rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nTuesday 24 February 2026  01:05:10 +0000 (0:00:08.933)       0:02:38.537 ****** \n=============================================================================== \nservice-ks-register : barbican | Creating projects --------------------- 18.16s\nservice-ks-register : barbican | Creating roles ------------------------ 16.59s\nbarbican : Restart barbican-api container ------------------------------ 13.71s\nbarbican : Running barbican bootstrap container ------------------------ 12.89s\nbarbican : Restart barbican-keystone-listener container ---------------- 10.33s\nbarbican : Copying over barbican.conf ----------------------------------- 9.74s\nbarbican : Restart barbican-worker container ---------------------------- 8.93s\nservice-ks-register : barbican | Creating endpoints --------------------- 8.75s\nbarbican : Check barbican containers ------------------------------------ 6.88s\nservice-ks-register : barbican | Granting user roles -------------------- 5.33s\nservice-ks-register : barbican | Creating users ------------------------- 4.86s\nservice-cert-copy : barbican | Copying over extra CA certificates ------- 4.23s\nbarbican : Copying over config.json files for services ------------------ 3.90s\nservice-ks-register : barbican | Creating services ---------------------- 3.78s\nbarbican : Creating barbican database user and setting permissions ------ 3.48s\nbarbican : Copying over barbican-api.ini -------------------------------- 3.47s\nbarbican : Ensuring config directories exist ---------------------------- 3.31s\nbarbican : Creating barbican database ----------------------------------- 2.96s\nbarbican : Checking whether barbican-api-paste.ini file exists ---------- 2.91s\nbarbican : Copying over existing policy file ---------------------------- 2.23s\n\u001b[32m2026-02-24 01:05:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a8ebcda5-cdc0-4611-9f70-ca910ead0481 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:05:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44aa611b-f898-4352-b174-ada638d8f758 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:05:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3046662f-74e4-408a-bb14-4c7f7919dd4b is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:05:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 01:05:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f714ea82-7678-4e20-85bd-ee1502bb5463 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:05:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a8ebcda5-cdc0-4611-9f70-ca910ead0481 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:05:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44aa611b-f898-4352-b174-ada638d8f758 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:05:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3046662f-74e4-408a-bb14-4c7f7919dd4b is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:05:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 01:05:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f714ea82-7678-4e20-85bd-ee1502bb5463 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:05:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a8ebcda5-cdc0-4611-9f70-ca910ead0481 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:05:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44aa611b-f898-4352-b174-ada638d8f758 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:05:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3046662f-74e4-408a-bb14-4c7f7919dd4b is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:05:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 01:05:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f714ea82-7678-4e20-85bd-ee1502bb5463 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:05:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a8ebcda5-cdc0-4611-9f70-ca910ead0481 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:05:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44aa611b-f898-4352-b174-ada638d8f758 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:05:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3046662f-74e4-408a-bb14-4c7f7919dd4b is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:05:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 01:05:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f714ea82-7678-4e20-85bd-ee1502bb5463 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:05:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a8ebcda5-cdc0-4611-9f70-ca910ead0481 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:05:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44aa611b-f898-4352-b174-ada638d8f758 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:05:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3046662f-74e4-408a-bb14-4c7f7919dd4b is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:05:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 01:05:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f714ea82-7678-4e20-85bd-ee1502bb5463 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:05:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a8ebcda5-cdc0-4611-9f70-ca910ead0481 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:05:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44aa611b-f898-4352-b174-ada638d8f758 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:05:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3046662f-74e4-408a-bb14-4c7f7919dd4b is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:05:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 01:05:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f714ea82-7678-4e20-85bd-ee1502bb5463 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:05:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a8ebcda5-cdc0-4611-9f70-ca910ead0481 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:05:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44aa611b-f898-4352-b174-ada638d8f758 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:05:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3046662f-74e4-408a-bb14-4c7f7919dd4b is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:05:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 01:05:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f714ea82-7678-4e20-85bd-ee1502bb5463 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:05:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a8ebcda5-cdc0-4611-9f70-ca910ead0481 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:05:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44aa611b-f898-4352-b174-ada638d8f758 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:05:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3046662f-74e4-408a-bb14-4c7f7919dd4b is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:05:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 01:05:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f714ea82-7678-4e20-85bd-ee1502bb5463 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:05:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a8ebcda5-cdc0-4611-9f70-ca910ead0481 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:05:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44aa611b-f898-4352-b174-ada638d8f758 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:05:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3046662f-74e4-408a-bb14-4c7f7919dd4b is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:05:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 01:05:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f714ea82-7678-4e20-85bd-ee1502bb5463 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:05:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a8ebcda5-cdc0-4611-9f70-ca910ead0481 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:05:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44aa611b-f898-4352-b174-ada638d8f758 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:05:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3046662f-74e4-408a-bb14-4c7f7919dd4b is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:05:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 01:05:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f714ea82-7678-4e20-85bd-ee1502bb5463 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:05:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a8ebcda5-cdc0-4611-9f70-ca910ead0481 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:05:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44aa611b-f898-4352-b174-ada638d8f758 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:05:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3046662f-74e4-408a-bb14-4c7f7919dd4b is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:05:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 01:05:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f714ea82-7678-4e20-85bd-ee1502bb5463 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:05:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a8ebcda5-cdc0-4611-9f70-ca910ead0481 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:05:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44aa611b-f898-4352-b174-ada638d8f758 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:05:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3046662f-74e4-408a-bb14-4c7f7919dd4b is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:05:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 01:05:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f714ea82-7678-4e20-85bd-ee1502bb5463 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:05:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a8ebcda5-cdc0-4611-9f70-ca910ead0481 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:05:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44aa611b-f898-4352-b174-ada638d8f758 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:05:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3046662f-74e4-408a-bb14-4c7f7919dd4b is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:05:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 01:05:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f714ea82-7678-4e20-85bd-ee1502bb5463 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:05:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a8ebcda5-cdc0-4611-9f70-ca910ead0481 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:05:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44aa611b-f898-4352-b174-ada638d8f758 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:05:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3046662f-74e4-408a-bb14-4c7f7919dd4b is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:05:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 01:05:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f714ea82-7678-4e20-85bd-ee1502bb5463 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:05:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a8ebcda5-cdc0-4611-9f70-ca910ead0481 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:05:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 47bc2a2f-3681-4d97-beda-759966d298ca is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:05:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44aa611b-f898-4352-b174-ada638d8f758 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:05:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3046662f-74e4-408a-bb14-4c7f7919dd4b is in state SUCCESS\u001b[0m\n\u001b[32m2026-02-24 01:05:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\n\nPLAY [Group hosts based on configuration] **************************************\n\nTASK [Group hosts based on Kolla action] ***************************************\nTuesday 24 February 2026  01:04:19 +0000 (0:00:00.339)       0:00:00.339 ****** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [Group hosts based on enabled services] ***********************************\nTuesday 24 February 2026  01:04:19 +0000 (0:00:00.566)       0:00:00.906 ****** \n\u001b[0;32mok: [testbed-node-0] => (item=enable_placement_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item=enable_placement_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item=enable_placement_True)\u001b[0m\n\nPLAY [Apply role placement] ****************************************************\n\nTASK [placement : include_tasks] ***********************************************\nTuesday 24 February 2026  01:04:21 +0000 (0:00:01.650)       0:00:02.556 ****** \n\u001b[0;36mincluded: /ansible/roles/placement/tasks/deploy.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [service-ks-register : placement | Creating services] *********************\nTuesday 24 February 2026  01:04:22 +0000 (0:00:00.951)       0:00:03.508 ****** \n\u001b[0;33mchanged: [testbed-node-0] => (item=placement (placement))\u001b[0m\n\nTASK [service-ks-register : placement | Creating endpoints] ********************\nTuesday 24 February 2026  01:04:26 +0000 (0:00:04.198)       0:00:07.707 ****** \n\u001b[0;33mchanged: [testbed-node-0] => (item=placement -> https://api-int.testbed.osism.xyz:8780 -> internal)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=placement -> https://api.testbed.osism.xyz:8780 -> public)\u001b[0m\n\nTASK [service-ks-register : placement | Creating projects] *********************\nTuesday 24 February 2026  01:04:34 +0000 (0:00:07.633)       0:00:15.340 ****** \n\u001b[0;32mok: [testbed-node-0] => (item=service)\u001b[0m\n\nTASK [service-ks-register : placement | Creating users] ************************\nTuesday 24 February 2026  01:04:38 +0000 (0:00:03.948)       0:00:19.288 ****** \n\u001b[0;33mchanged: [testbed-node-0] => (item=placement -> service)\u001b[0m\n\u001b[1;35m[WARNING]: Module did not set no_log for update_password\u001b[0m\n\nTASK [service-ks-register : placement | Creating roles] ************************\nTuesday 24 February 2026  01:04:42 +0000 (0:00:04.837)       0:00:24.126 ****** \n\u001b[0;32mok: [testbed-node-0] => (item=admin)\u001b[0m\n\nTASK [service-ks-register : placement | Granting user roles] *******************\nTuesday 24 February 2026  01:04:47 +0000 (0:00:04.433)       0:00:28.559 ****** \n\u001b[0;33mchanged: [testbed-node-0] => (item=placement -> service -> admin)\u001b[0m\n\nTASK [placement : include_tasks] ***********************************************\nTuesday 24 February 2026  01:04:52 +0000 (0:00:04.969)       0:00:33.529 ****** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [placement : Ensuring config directories exist] ***************************\nTuesday 24 February 2026  01:04:52 +0000 (0:00:00.429)       0:00:33.958 ****** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/placement-api:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8780'], 'timeout': '30'}, 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/placement-api:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8780'], 'timeout': '30'}, 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/placement-api:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8780'], 'timeout': '30'}, 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}}})\u001b[0m\n\nTASK [placement : Check if policies shall be overwritten] **********************\nTuesday 24 February 2026  01:04:54 +0000 (0:00:01.711)       0:00:35.670 ****** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [placement : Set placement policy file] ***********************************\nTuesday 24 February 2026  01:04:54 +0000 (0:00:00.289)       0:00:35.960 ****** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [placement : include_tasks] ***********************************************\nTuesday 24 February 2026  01:04:55 +0000 (0:00:00.872)       0:00:36.832 ****** \n\u001b[0;36mincluded: /ansible/roles/placement/tasks/copy-certs.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [service-cert-copy : placement | Copying over extra CA certificates] ******\nTuesday 24 February 2026  01:04:56 +0000 (0:00:00.627)       0:00:37.459 ****** \n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/placement-api:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8780'], 'timeout': '30'}, 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/placement-api:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8780'], 'timeout': '30'}, 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/placement-api:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8780'], 'timeout': '30'}, 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}}})\u001b[0m\n\nTASK [service-cert-copy : placement | Copying over backend internal TLS certificate] ***\nTuesday 24 February 2026  01:04:58 +0000 (0:00:02.017)       0:00:39.477 ****** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/placement-api:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8780'], 'timeout': '30'}, 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/placement-api:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8780'], 'timeout': '30'}, 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/placement-api:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8780'], 'timeout': '30'}, 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [service-cert-copy : placement | Copying over backend internal TLS key] ***\nTuesday 24 February 2026  01:04:59 +0000 (0:00:00.918)       0:00:40.395 ****** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/placement-api:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8780'], 'timeout': '30'}, 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/placement-api:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8780'], 'timeout': '30'}, 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/placement-api:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8780'], 'timeout': '30'}, 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [placement : Copying over config.json files for services] *****************\nTuesday 24 February 2026  01:05:00 +0000 (0:00:01.788)       0:00:42.184 ****** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/placement-api:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8780'], 'timeout': '30'}, 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/placement-api:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8780'], 'timeout': '30'}, 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/placement-api:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8780'], 'timeout': '30'}, 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}}})\u001b[0m\n\nTASK [placement : Copying over placement.conf] *********************************\nTuesday 24 February 2026  01:05:03 +0000 (0:00:02.110)       0:00:44.294 ****** \n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/placement-api:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8780'], 'timeout': '30'}, 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/placement-api:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8780'], 'timeout': '30'}, 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/placement-api:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8780'], 'timeout': '30'}, 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}}})\u001b[0m\n\nTASK [placement : Copying over placement-api wsgi configuration] ***************\nTuesday 24 February 2026  01:05:09 +0000 (0:00:06.241)       0:00:50.536 ****** \n\u001b[0;33mchanged: [testbed-node-0] => (item=/ansible/roles/placement/templates/placement-api-wsgi.conf.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=/ansible/roles/placement/templates/placement-api-wsgi.conf.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=/ansible/roles/placement/templates/placement-api-wsgi.conf.j2)\u001b[0m\n\nTASK [placement : Copying over migrate-db.rc.j2 configuration] *****************\nTuesday 24 February 2026  01:05:11 +0000 (0:00:02.441)       0:00:52.978 ****** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [placement : Copying over existing policy file] ***************************\nTuesday 24 February 2026  01:05:13 +0000 (0:00:02.040)       0:00:55.018 ****** \n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/placement-api:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8780'], 'timeout': '30'}, 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/placement-api:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8780'], 'timeout': '30'}, 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/placement-api:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8780'], 'timeout': '30'}, 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [placement : Check placement containers] **********************************\nTuesday 24 February 2026  01:05:14 +0000 (0:00:00.676)       0:00:55.695 ****** \n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/placement-api:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8780'], 'timeout': '30'}, 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/placement-api:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8780'], 'timeout': '30'}, 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/placement-api:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8780'], 'timeout': '30'}, 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}}})\u001b[0m\n\nTASK [placement : Creating placement databases] ********************************\nTuesday 24 February 2026  01:05:16 +0000 (0:00:01.584)       0:00:57.280 ****** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [placement : Creating placement databases user and setting permissions] ***\nTuesday 24 February 2026  01:05:21 +0000 (0:00:05.082)       0:01:02.363 ****** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [placement : Running placement bootstrap container] ***********************\nTuesday 24 February 2026  01:05:24 +0000 (0:00:03.089)       0:01:05.452 ****** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [placement : Flush handlers] **********************************************\nTuesday 24 February 2026  01:05:40 +0000 (0:00:16.111)       0:01:21.563 ****** \n\nTASK [placement : Flush handlers] **********************************************\nTuesday 24 February 2026  01:05:40 +0000 (0:00:00.073)       0:01:21.637 ****** \n\nTASK [placement : Flush handlers] **********************************************\nTuesday 24 February 2026  01:05:40 +0000 (0:00:00.118)       0:01:21.756 ****** \n\nRUNNING HANDLER [placement : Restart placement-api container] ******************\nTuesday 24 February 2026  01:05:40 +0000 (0:00:00.079)       0:01:21.835 ****** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=21  \u001b[0m \u001b[0;33mchanged=15  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=6   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=12  \u001b[0m \u001b[0;33mchanged=8   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=5   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=12  \u001b[0m \u001b[0;33mchanged=8   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=5   \u001b[0m rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nTuesday 24 February 2026  01:05:52 +0000 (0:00:12.133)       0:01:33.968 ****** \n=============================================================================== \nplacement : Running placement bootstrap container ---------------------- 16.11s\nplacement : Restart placement-api container ---------------------------- 12.13s\nservice-ks-register : placement | Creating endpoints -------------------- 7.63s\nplacement : Copying over placement.conf --------------------------------- 6.24s\nplacement : Creating placement databases -------------------------------- 5.08s\nservice-ks-register : placement | Granting user roles ------------------- 4.97s\nservice-ks-register : placement | Creating users ------------------------ 4.84s\nservice-ks-register : placement | Creating roles ------------------------ 4.43s\nservice-ks-register : placement | Creating services --------------------- 4.20s\nservice-ks-register : placement | Creating projects --------------------- 3.95s\nplacement : Creating placement databases user and setting permissions --- 3.09s\nplacement : Copying over placement-api wsgi configuration --------------- 2.44s\nplacement : Copying over config.json files for services ----------------- 2.11s\nplacement : Copying over migrate-db.rc.j2 configuration ----------------- 2.04s\nservice-cert-copy : placement | Copying over extra CA certificates ------ 2.02s\nservice-cert-copy : placement | Copying over backend internal TLS key --- 1.79s\nplacement : Ensuring config directories exist --------------------------- 1.71s\nGroup hosts based on enabled services ----------------------------------- 1.65s\nplacement : Check placement containers ---------------------------------- 1.58s\nplacement : include_tasks ----------------------------------------------- 0.95s\n\u001b[32m2026-02-24 01:05:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f714ea82-7678-4e20-85bd-ee1502bb5463 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:05:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a8ebcda5-cdc0-4611-9f70-ca910ead0481 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:05:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 47bc2a2f-3681-4d97-beda-759966d298ca is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:05:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44aa611b-f898-4352-b174-ada638d8f758 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:05:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 01:06:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f714ea82-7678-4e20-85bd-ee1502bb5463 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:06:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a8ebcda5-cdc0-4611-9f70-ca910ead0481 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:06:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a29489db-95ad-418c-bc67-f87a0737c252 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:06:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 47bc2a2f-3681-4d97-beda-759966d298ca is in state SUCCESS\u001b[0m\n\u001b[32m2026-02-24 01:06:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44aa611b-f898-4352-b174-ada638d8f758 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:06:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 01:06:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f714ea82-7678-4e20-85bd-ee1502bb5463 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:06:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a8ebcda5-cdc0-4611-9f70-ca910ead0481 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:06:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a29489db-95ad-418c-bc67-f87a0737c252 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:06:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44aa611b-f898-4352-b174-ada638d8f758 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:06:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 01:06:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f714ea82-7678-4e20-85bd-ee1502bb5463 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:06:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a8ebcda5-cdc0-4611-9f70-ca910ead0481 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:06:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a29489db-95ad-418c-bc67-f87a0737c252 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:06:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44aa611b-f898-4352-b174-ada638d8f758 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:06:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 01:06:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f714ea82-7678-4e20-85bd-ee1502bb5463 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:06:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a8ebcda5-cdc0-4611-9f70-ca910ead0481 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:06:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a29489db-95ad-418c-bc67-f87a0737c252 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:06:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44aa611b-f898-4352-b174-ada638d8f758 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:06:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 01:06:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f714ea82-7678-4e20-85bd-ee1502bb5463 is in state SUCCESS\u001b[0m\n\n\nPLAY [Group hosts based on configuration] **************************************\n\nTASK [Group hosts based on Kolla action] ***************************************\nTuesday 24 February 2026  01:05:58 +0000 (0:00:00.208)       0:00:00.208 ****** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [Group hosts based on enabled services] ***********************************\nTuesday 24 February 2026  01:05:58 +0000 (0:00:00.311)       0:00:00.520 ****** \n\u001b[0;32mok: [testbed-node-0] => (item=enable_keystone_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item=enable_keystone_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item=enable_keystone_True)\u001b[0m\n\nPLAY [Wait for the Keystone service] *******************************************\n\nTASK [Waiting for Keystone public port to be UP] *******************************\nTuesday 24 February 2026  01:05:59 +0000 (0:00:00.748)       0:00:01.268 ****** \n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;32mtestbed-node-0\u001b[0m             : \u001b[0;32mok=3   \u001b[0m changed=0    unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\u001b[0;32mtestbed-node-1\u001b[0m             : \u001b[0;32mok=3   \u001b[0m changed=0    unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\u001b[0;32mtestbed-node-2\u001b[0m             : \u001b[0;32mok=3   \u001b[0m changed=0    unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nTuesday 24 February 2026  01:06:00 +0000 (0:00:00.662)       0:00:01.931 ****** \n=============================================================================== \nGroup hosts based on enabled services ----------------------------------- 0.75s\nWaiting for Keystone public port to be UP ------------------------------- 0.66s\nGroup hosts based on Kolla action --------------------------------------- 0.31s\n\n\nPLAY [Group hosts based on configuration] **************************************\n\nTASK [Group hosts based on Kolla action] ***************************************\nTuesday 24 February 2026  01:02:32 +0000 (0:00:00.288)       0:00:00.288 ****** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [Group hosts based on enabled services] ***********************************\nTuesday 24 February 2026  01:02:32 +0000 (0:00:00.301)       0:00:00.590 ****** \n\u001b[0;32mok: [testbed-node-0] => (item=enable_designate_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item=enable_designate_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item=enable_designate_True)\u001b[0m\n\nPLAY [Apply role designate] ****************************************************\n\nTASK [designate : include_tasks] ***********************************************\nTuesday 24 February 2026  01:02:33 +0000 (0:00:00.448)       0:00:01.038 ****** \n\u001b[0;36mincluded: /ansible/roles/designate/tasks/deploy.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [service-ks-register : designate | Creating services] *********************\nTuesday 24 February 2026  01:02:33 +0000 (0:00:00.510)       0:00:01.549 ****** \n\u001b[0;33mchanged: [testbed-node-0] => (item=designate (dns))\u001b[0m\n\nTASK [service-ks-register : designate | Creating endpoints] ********************\nTuesday 24 February 2026  01:02:38 +0000 (0:00:04.482)       0:00:06.032 ****** \n\u001b[0;33mchanged: [testbed-node-0] => (item=designate -> https://api-int.testbed.osism.xyz:9001 -> internal)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=designate -> https://api.testbed.osism.xyz:9001 -> public)\u001b[0m\n\nTASK [service-ks-register : designate | Creating projects] *********************\nTuesday 24 February 2026  01:02:46 +0000 (0:00:08.061)       0:00:14.093 ****** \n\u001b[0;33mchanged: [testbed-node-0] => (item=service)\u001b[0m\n\nTASK [service-ks-register : designate | Creating users] ************************\nTuesday 24 February 2026  01:02:50 +0000 (0:00:04.045)       0:00:18.139 ****** \n\u001b[0;33mchanged: [testbed-node-0] => (item=designate -> service)\u001b[0m\n\u001b[1;35m[WARNING]: Module did not set no_log for update_password\u001b[0m\n\nTASK [service-ks-register : designate | Creating roles] ************************\nTuesday 24 February 2026  01:02:54 +0000 (0:00:04.576)       0:00:22.715 ****** \n\u001b[0;32mok: [testbed-node-0] => (item=admin)\u001b[0m\n\nTASK [service-ks-register : designate | Granting user roles] *******************\nTuesday 24 February 2026  01:02:59 +0000 (0:00:04.160)       0:00:26.876 ****** \n\u001b[0;33mchanged: [testbed-node-0] => (item=designate -> service -> admin)\u001b[0m\n\nTASK [designate : Ensuring config directories exist] ***************************\nTuesday 24 February 2026  01:03:03 +0000 (0:00:04.804)       0:00:31.680 ****** \n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-api:2024.2', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-api:2024.2', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-api:2024.2', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-backend-bind9:2024.2', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-backend-bind9:2024.2', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-backend-bind9:2024.2', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-central:2024.2', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-central:2024.2', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-central:2024.2', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-mdns:2024.2', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-mdns:2024.2', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-mdns:2024.2', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-producer:2024.2', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-producer:2024.2', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-producer:2024.2', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-worker:2024.2', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-worker:2024.2', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-worker:2024.2', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}})\u001b[0m\n\nTASK [designate : Check if policies shall be overwritten] **********************\nTuesday 24 February 2026  01:03:07 +0000 (0:00:03.279)       0:00:34.959 ****** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [designate : Set designate policy file] ***********************************\nTuesday 24 February 2026  01:03:07 +0000 (0:00:00.151)       0:00:35.111 ****** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [designate : include_tasks] ***********************************************\nTuesday 24 February 2026  01:03:07 +0000 (0:00:00.287)       0:00:35.398 ****** \n\u001b[0;36mincluded: /ansible/roles/designate/tasks/copy-certs.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [service-cert-copy : designate | Copying over extra CA certificates] ******\nTuesday 24 February 2026  01:03:08 +0000 (0:00:00.860)       0:00:36.258 ****** \n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-api:2024.2', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-api:2024.2', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-api:2024.2', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-backend-bind9:2024.2', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-backend-bind9:2024.2', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-backend-bind9:2024.2', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-central:2024.2', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-central:2024.2', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-central:2024.2', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-mdns:2024.2', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-mdns:2024.2', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-mdns:2024.2', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-producer:2024.2', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-producer:2024.2', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-producer:2024.2', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-worker:2024.2', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-worker:2024.2', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-worker:2024.2', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}})\u001b[0m\n\nTASK [service-cert-copy : designate | Copying over backend internal TLS certificate] ***\nTuesday 24 February 2026  01:03:14 +0000 (0:00:06.356)       0:00:42.615 ****** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-api:2024.2', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-backend-bind9:2024.2', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-central:2024.2', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-mdns:2024.2', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-producer:2024.2', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-worker:2024.2', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-api:2024.2', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-backend-bind9:2024.2', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-central:2024.2', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-mdns:2024.2', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-producer:2024.2', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-worker:2024.2', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-api:2024.2', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-backend-bind9:2024.2', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-central:2024.2', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-mdns:2024.2', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-producer:2024.2', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-worker:2024.2', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [service-cert-copy : designate | Copying over backend internal TLS key] ***\nTuesday 24 February 2026  01:03:16 +0000 (0:00:01.398)       0:00:44.014 ****** \n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-api:2024.2', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-backend-bind9:2024.2', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-central:2024.2', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-mdns:2024.2', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-producer:2024.2', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-api:2024.2', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-worker:2024.2', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-backend-bind9:2024.2', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-central:2024.2', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-mdns:2024.2', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-producer:2024.2', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-worker:2024.2', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-api:2024.2', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-backend-bind9:2024.2', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-central:2024.2', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-mdns:2024.2', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-producer:2024.2', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-worker:2024.2', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [designate : Copying over config.json files for services] *****************\nTuesday 24 February 2026  01:03:18 +0000 (0:00:02.139)       0:00:46.154 ****** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-api:2024.2', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-api:2024.2', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-api:2024.2', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-backend-bind9:2024.2', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-backend-bind9:2024.2', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-backend-bind9:2024.2', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-central:2024.2', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-central:2024.2', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-central:2024.2', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-mdns:2024.2', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-mdns:2024.2', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-mdns:2024.2', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-producer:2024.2', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-producer:2024.2', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-producer:2024.2', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-worker:2024.2', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-worker:2024.2', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-worker:2024.2', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}})\u001b[0m\n\nTASK [designate : Copying over designate.conf] *********************************\nTuesday 24 February 2026  01:03:25 +0000 (0:00:07.464)       0:00:53.619 ****** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-api:2024.2', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-api:2024.2', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-api:2024.2', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-backend-bind9:2024.2', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-backend-bind9:2024.2', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-backend-bind9:2024.2', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-central:2024.2', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-central:2024.2', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-central:2024.2', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-mdns:2024.2', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-mdns:2024.2', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-mdns:2024.2', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-producer:2024.2', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-producer:2024.2', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-producer:2024.2', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-worker:2024.2', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-worker:2024.2', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-worker:2024.2', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}})\u001b[0m\n\nTASK [designate : Copying over pools.yaml] *************************************\nTuesday 24 February 2026  01:03:50 +0000 (0:00:24.326)       0:01:17.945 ****** \n\u001b[0;33mchanged: [testbed-node-0] => (item=/ansible/roles/designate/templates/pools.yaml.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=/ansible/roles/designate/templates/pools.yaml.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=/ansible/roles/designate/templates/pools.yaml.j2)\u001b[0m\n\nTASK [designate : Copying over named.conf] *************************************\nTuesday 24 February 2026  01:03:58 +0000 (0:00:08.679)       0:01:26.625 ****** \n\u001b[0;33mchanged: [testbed-node-0] => (item=/ansible/roles/designate/templates/named.conf.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=/ansible/roles/designate/templates/named.conf.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=/ansible/roles/designate/templates/named.conf.j2)\u001b[0m\n\nTASK [designate : Copying over rndc.conf] **************************************\nTuesday 24 February 2026  01:04:03 +0000 (0:00:04.651)       0:01:31.276 ****** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-api:2024.2', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-api:2024.2', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-api:2024.2', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}}) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-backend-bind9:2024.2', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-central:2024.2', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-mdns:2024.2', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-producer:2024.2', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-backend-bind9:2024.2', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-central:2024.2', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-mdns:2024.2', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-producer:2024.2', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-backend-bind9:2024.2', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-central:2024.2', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-mdns:2024.2', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-producer:2024.2', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-worker:2024.2', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-worker:2024.2', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-worker:2024.2', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}})\u001b[0m\n\nTASK [designate : Copying over rndc.key] ***************************************\nTuesday 24 February 2026  01:04:07 +0000 (0:00:03.940)       0:01:35.216 ****** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-api:2024.2', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-api:2024.2', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-api:2024.2', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}}) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-backend-bind9:2024.2', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-central:2024.2', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-backend-bind9:2024.2', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-mdns:2024.2', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-central:2024.2', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-mdns:2024.2', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-producer:2024.2', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-producer:2024.2', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-backend-bind9:2024.2', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-central:2024.2', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-mdns:2024.2', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-producer:2024.2', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-worker:2024.2', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-worker:2024.2', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-worker:2024.2', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}})\u001b[0m\n\nTASK [designate : include_tasks] ***********************************************\nTuesday 24 February 2026  01:04:12 +0000 (0:00:05.065)       0:01:40.281 ****** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [designate : Copying over existing policy file] ***************************\nTuesday 24 February 2026  01:04:13 +0000 (0:00:01.104)       0:01:41.385 ****** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-api:2024.2', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-backend-bind9:2024.2', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-central:2024.2', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-mdns:2024.2', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-producer:2024.2', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-worker:2024.2', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-api:2024.2', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-backend-bind9:2024.2', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-central:2024.2', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-mdns:2024.2', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-producer:2024.2', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-worker:2024.2', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-api:2024.2', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-backend-bind9:2024.2', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-central:2024.2', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-mdns:2024.2', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-producer:2024.2', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-worker:2024.2', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [designate : Check designate containers] **********************************\nTuesday 24 February 2026  01:04:14 +0000 (0:00:01.014)       0:01:42.400 ****** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-api:2024.2', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-backend-bind9:2024.2', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-api:2024.2', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-api:2024.2', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-central:2024.2', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-backend-bind9:2024.2', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-backend-bind9:2024.2', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-mdns:2024.2', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-central:2024.2', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-central:2024.2', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-producer:2024.2', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-mdns:2024.2', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-mdns:2024.2', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-producer:2024.2', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-worker:2024.2', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-producer:2024.2', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-worker:2024.2', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-worker:2024.2', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}})\u001b[0m\n\nTASK [designate : include_tasks] ***********************************************\nTuesday 24 February 2026  01:04:21 +0000 (0:00:07.070)       0:01:49.470 ****** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [designate : Creating Designate databases] ********************************\nTuesday 24 February 2026  01:04:22 +0000 (0:00:00.488)       0:01:49.958 ****** \n\u001b[0;33mchanged: [testbed-node-0] => (item=designate)\u001b[0m\n\nTASK [designate : Creating Designate databases user and setting permissions] ***\nTuesday 24 February 2026  01:04:24 +0000 (0:00:02.750)       0:01:52.709 ****** \n\u001b[0;33mchanged: [testbed-node-0] => (item=None)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0 -> {{ groups['designate-central'][0] }}]\u001b[0m\n\nTASK [designate : Running Designate bootstrap container] ***********************\nTuesday 24 February 2026  01:04:27 +0000 (0:00:02.892)       0:01:55.602 ****** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [designate : Flush handlers] **********************************************\nTuesday 24 February 2026  01:04:47 +0000 (0:00:19.447)       0:02:15.050 ****** \n\nTASK [designate : Flush handlers] **********************************************\nTuesday 24 February 2026  01:04:47 +0000 (0:00:00.070)       0:02:15.120 ****** \n\nTASK [designate : Flush handlers] **********************************************\nTuesday 24 February 2026  01:04:47 +0000 (0:00:00.067)       0:02:15.187 ****** \n\nRUNNING HANDLER [designate : Restart designate-backend-bind9 container] ********\nTuesday 24 February 2026  01:04:47 +0000 (0:00:00.065)       0:02:15.253 ****** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [designate : Restart designate-api container] ******************\nTuesday 24 February 2026  01:05:01 +0000 (0:00:13.945)       0:02:29.199 ****** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [designate : Restart designate-central container] **************\nTuesday 24 February 2026  01:05:15 +0000 (0:00:14.513)       0:02:43.712 ****** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\nRUNNING HANDLER [designate : Restart designate-producer container] *************\nTuesday 24 February 2026  01:05:31 +0000 (0:00:15.292)       0:02:59.004 ****** \n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nRUNNING HANDLER [designate : Restart designate-mdns container] *****************\nTuesday 24 February 2026  01:05:40 +0000 (0:00:09.370)       0:03:08.375 ****** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [designate : Restart designate-worker container] ***************\nTuesday 24 February 2026  01:05:52 +0000 (0:00:11.935)       0:03:20.310 ****** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\nTASK [designate : Non-destructive DNS pools update] ****************************\nTuesday 24 February 2026  01:06:04 +0000 (0:00:11.520)       0:03:31.831 ****** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=29  \u001b[0m \u001b[0;33mchanged=24  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=7   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=19  \u001b[0m \u001b[0;33mchanged=15  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=6   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=19  \u001b[0m \u001b[0;33mchanged=15  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=6   \u001b[0m rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nTuesday 24 February 2026  01:06:12 +0000 (0:00:08.881)       0:03:40.713 ****** \n=============================================================================== \ndesignate : Copying over designate.conf -------------------------------- 24.33s\ndesignate : Running Designate bootstrap container ---------------------- 19.45s\ndesignate : Restart designate-central container ------------------------ 15.29s\ndesignate : Restart designate-api container ---------------------------- 14.51s\ndesignate : Restart designate-backend-bind9 container ------------------ 13.95s\ndesignate : Restart designate-mdns container --------------------------- 11.94s\ndesignate : Restart designate-worker container ------------------------- 11.52s\ndesignate : Restart designate-producer container ------------------------ 9.37s\ndesignate : Non-destructive DNS pools update ---------------------------- 8.88s\ndesignate : Copying over pools.yaml ------------------------------------- 8.68s\nservice-ks-register : designate | Creating endpoints -------------------- 8.06s\ndesignate : Copying over config.json files for services ----------------- 7.46s\ndesignate : Check designate containers ---------------------------------- 7.07s\nservice-cert-copy : designate | Copying over extra CA certificates ------ 6.36s\ndesignate : Copying over rndc.key --------------------------------------- 5.07s\nservice-ks-register : designate | Granting user roles ------------------- 4.80s\ndesignate : Copying over named.conf ------------------------------------- 4.65s\nservice-ks-register : designate | Creating users ------------------------ 4.58s\nservice-ks-register : designate | Creating services --------------------- 4.48s\nservice-ks-register : designate | Creating roles ------------------------ 4.16s\n\u001b[32m2026-02-24 01:06:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b44e23a9-4820-4202-81f0-47da60f63246 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:06:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a8ebcda5-cdc0-4611-9f70-ca910ead0481 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:06:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a29489db-95ad-418c-bc67-f87a0737c252 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:06:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44aa611b-f898-4352-b174-ada638d8f758 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:06:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 01:06:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b44e23a9-4820-4202-81f0-47da60f63246 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:06:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a8ebcda5-cdc0-4611-9f70-ca910ead0481 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:06:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a29489db-95ad-418c-bc67-f87a0737c252 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:06:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44aa611b-f898-4352-b174-ada638d8f758 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:06:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 01:06:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b44e23a9-4820-4202-81f0-47da60f63246 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:06:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a8ebcda5-cdc0-4611-9f70-ca910ead0481 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:06:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a29489db-95ad-418c-bc67-f87a0737c252 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:06:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44aa611b-f898-4352-b174-ada638d8f758 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:06:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 01:06:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b44e23a9-4820-4202-81f0-47da60f63246 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:06:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a8ebcda5-cdc0-4611-9f70-ca910ead0481 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:06:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a29489db-95ad-418c-bc67-f87a0737c252 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:06:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44aa611b-f898-4352-b174-ada638d8f758 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:06:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 01:06:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b44e23a9-4820-4202-81f0-47da60f63246 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:06:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a8ebcda5-cdc0-4611-9f70-ca910ead0481 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:06:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a29489db-95ad-418c-bc67-f87a0737c252 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:06:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44aa611b-f898-4352-b174-ada638d8f758 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:06:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 01:06:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b44e23a9-4820-4202-81f0-47da60f63246 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:06:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a8ebcda5-cdc0-4611-9f70-ca910ead0481 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:06:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a29489db-95ad-418c-bc67-f87a0737c252 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:06:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44aa611b-f898-4352-b174-ada638d8f758 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:06:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 01:06:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b44e23a9-4820-4202-81f0-47da60f63246 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:06:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a8ebcda5-cdc0-4611-9f70-ca910ead0481 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:06:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a29489db-95ad-418c-bc67-f87a0737c252 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:06:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44aa611b-f898-4352-b174-ada638d8f758 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:06:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 01:06:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b44e23a9-4820-4202-81f0-47da60f63246 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:06:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a8ebcda5-cdc0-4611-9f70-ca910ead0481 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:06:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a29489db-95ad-418c-bc67-f87a0737c252 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:06:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44aa611b-f898-4352-b174-ada638d8f758 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:06:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 01:06:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b44e23a9-4820-4202-81f0-47da60f63246 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:06:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a8ebcda5-cdc0-4611-9f70-ca910ead0481 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:06:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a29489db-95ad-418c-bc67-f87a0737c252 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:06:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44aa611b-f898-4352-b174-ada638d8f758 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:06:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 01:06:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b44e23a9-4820-4202-81f0-47da60f63246 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:06:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a8ebcda5-cdc0-4611-9f70-ca910ead0481 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:06:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a29489db-95ad-418c-bc67-f87a0737c252 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:06:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44aa611b-f898-4352-b174-ada638d8f758 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:06:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 01:06:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b44e23a9-4820-4202-81f0-47da60f63246 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:06:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a8ebcda5-cdc0-4611-9f70-ca910ead0481 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:06:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a29489db-95ad-418c-bc67-f87a0737c252 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:06:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44aa611b-f898-4352-b174-ada638d8f758 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:06:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 01:06:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b44e23a9-4820-4202-81f0-47da60f63246 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:06:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a8ebcda5-cdc0-4611-9f70-ca910ead0481 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:06:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a29489db-95ad-418c-bc67-f87a0737c252 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:06:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44aa611b-f898-4352-b174-ada638d8f758 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:06:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 01:06:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b44e23a9-4820-4202-81f0-47da60f63246 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:06:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a8ebcda5-cdc0-4611-9f70-ca910ead0481 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:06:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a29489db-95ad-418c-bc67-f87a0737c252 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:06:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44aa611b-f898-4352-b174-ada638d8f758 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:06:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 01:06:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b44e23a9-4820-4202-81f0-47da60f63246 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:06:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a8ebcda5-cdc0-4611-9f70-ca910ead0481 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:06:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a29489db-95ad-418c-bc67-f87a0737c252 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:06:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44aa611b-f898-4352-b174-ada638d8f758 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:06:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 01:06:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b44e23a9-4820-4202-81f0-47da60f63246 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:06:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a8ebcda5-cdc0-4611-9f70-ca910ead0481 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:06:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a29489db-95ad-418c-bc67-f87a0737c252 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:06:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44aa611b-f898-4352-b174-ada638d8f758 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:06:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 01:07:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b44e23a9-4820-4202-81f0-47da60f63246 is in state SUCCESS\u001b[0m\n\u001b[32m2026-02-24 01:07:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a8ebcda5-cdc0-4611-9f70-ca910ead0481 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:07:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a29489db-95ad-418c-bc67-f87a0737c252 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:07:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44aa611b-f898-4352-b174-ada638d8f758 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:07:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 01:07:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fba6bf07-65a1-432e-9a3d-72209557a21d is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:07:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a8ebcda5-cdc0-4611-9f70-ca910ead0481 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:07:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a29489db-95ad-418c-bc67-f87a0737c252 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:07:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44aa611b-f898-4352-b174-ada638d8f758 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:07:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 01:07:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fba6bf07-65a1-432e-9a3d-72209557a21d is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:07:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a8ebcda5-cdc0-4611-9f70-ca910ead0481 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:07:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a29489db-95ad-418c-bc67-f87a0737c252 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:07:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44aa611b-f898-4352-b174-ada638d8f758 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:07:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 01:07:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fba6bf07-65a1-432e-9a3d-72209557a21d is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:07:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a8ebcda5-cdc0-4611-9f70-ca910ead0481 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:07:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a29489db-95ad-418c-bc67-f87a0737c252 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:07:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44aa611b-f898-4352-b174-ada638d8f758 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:07:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 01:07:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fba6bf07-65a1-432e-9a3d-72209557a21d is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:07:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a8ebcda5-cdc0-4611-9f70-ca910ead0481 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:07:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a29489db-95ad-418c-bc67-f87a0737c252 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:07:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44aa611b-f898-4352-b174-ada638d8f758 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:07:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 01:07:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fba6bf07-65a1-432e-9a3d-72209557a21d is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:07:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a8ebcda5-cdc0-4611-9f70-ca910ead0481 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:07:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a29489db-95ad-418c-bc67-f87a0737c252 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:07:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44aa611b-f898-4352-b174-ada638d8f758 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:07:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 01:07:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fba6bf07-65a1-432e-9a3d-72209557a21d is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:07:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a8ebcda5-cdc0-4611-9f70-ca910ead0481 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:07:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a29489db-95ad-418c-bc67-f87a0737c252 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:07:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44aa611b-f898-4352-b174-ada638d8f758 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:07:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 01:07:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fba6bf07-65a1-432e-9a3d-72209557a21d is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:07:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a8ebcda5-cdc0-4611-9f70-ca910ead0481 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:07:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a29489db-95ad-418c-bc67-f87a0737c252 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:07:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44aa611b-f898-4352-b174-ada638d8f758 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:07:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 01:07:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fba6bf07-65a1-432e-9a3d-72209557a21d is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:07:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a8ebcda5-cdc0-4611-9f70-ca910ead0481 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:07:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a29489db-95ad-418c-bc67-f87a0737c252 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:07:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44aa611b-f898-4352-b174-ada638d8f758 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:07:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 01:07:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fba6bf07-65a1-432e-9a3d-72209557a21d is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:07:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a8ebcda5-cdc0-4611-9f70-ca910ead0481 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:07:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a29489db-95ad-418c-bc67-f87a0737c252 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:07:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44aa611b-f898-4352-b174-ada638d8f758 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:07:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 01:07:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fba6bf07-65a1-432e-9a3d-72209557a21d is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:07:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a8ebcda5-cdc0-4611-9f70-ca910ead0481 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:07:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a29489db-95ad-418c-bc67-f87a0737c252 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:07:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44aa611b-f898-4352-b174-ada638d8f758 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:07:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 01:07:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fba6bf07-65a1-432e-9a3d-72209557a21d is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:07:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a8ebcda5-cdc0-4611-9f70-ca910ead0481 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:07:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a29489db-95ad-418c-bc67-f87a0737c252 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:07:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44aa611b-f898-4352-b174-ada638d8f758 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:07:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 01:07:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fba6bf07-65a1-432e-9a3d-72209557a21d is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:07:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask eddb1a5f-448f-4517-b064-c30da006d691 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:07:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a8ebcda5-cdc0-4611-9f70-ca910ead0481 is in state SUCCESS\u001b[0m\n\n\nPLAY [Group hosts based on configuration] **************************************\n\nTASK [Group hosts based on Kolla action] ***************************************\nTuesday 24 February 2026  01:06:19 +0000 (0:00:00.431)       0:00:00.431 ****** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [Group hosts based on enabled services] ***********************************\nTuesday 24 February 2026  01:06:20 +0000 (0:00:01.058)       0:00:01.489 ****** \n\u001b[0;32mok: [testbed-manager] => (item=enable_ceph_rgw_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-0] => (item=enable_ceph_rgw_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item=enable_ceph_rgw_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item=enable_ceph_rgw_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item=enable_ceph_rgw_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item=enable_ceph_rgw_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item=enable_ceph_rgw_True)\u001b[0m\n\nPLAY [Apply role ceph-rgw] *****************************************************\n\nTASK [ceph-rgw : include_tasks] ************************************************\nTuesday 24 February 2026  01:06:22 +0000 (0:00:01.650)       0:00:03.139 ****** \n\u001b[0;36mincluded: /ansible/roles/ceph-rgw/tasks/deploy.yml for testbed-manager, testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m\n\nTASK [service-ks-register : ceph-rgw | Creating services] **********************\nTuesday 24 February 2026  01:06:26 +0000 (0:00:03.865)       0:00:07.005 ****** \n\u001b[0;33mchanged: [testbed-manager] => (item=swift (object-store))\u001b[0m\n\nTASK [service-ks-register : ceph-rgw | Creating endpoints] *********************\nTuesday 24 February 2026  01:06:30 +0000 (0:00:03.828)       0:00:10.833 ****** \n\u001b[0;33mchanged: [testbed-manager] => (item=swift -> https://api-int.testbed.osism.xyz:6780/swift/v1/AUTH_%(project_id)s -> internal)\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item=swift -> https://api.testbed.osism.xyz:6780/swift/v1/AUTH_%(project_id)s -> public)\u001b[0m\n\nTASK [service-ks-register : ceph-rgw | Creating projects] **********************\nTuesday 24 February 2026  01:06:37 +0000 (0:00:07.352)       0:00:18.185 ****** \n\u001b[0;32mok: [testbed-manager] => (item=service)\u001b[0m\n\nTASK [service-ks-register : ceph-rgw | Creating users] *************************\nTuesday 24 February 2026  01:06:40 +0000 (0:00:03.458)       0:00:21.644 ****** \n\u001b[0;33mchanged: [testbed-manager] => (item=ceph_rgw -> service)\u001b[0m\n\u001b[1;35m[WARNING]: Module did not set no_log for update_password\u001b[0m\n\nTASK [service-ks-register : ceph-rgw | Creating roles] *************************\nTuesday 24 February 2026  01:06:46 +0000 (0:00:05.445)       0:00:27.090 ****** \n\u001b[0;32mok: [testbed-manager] => (item=admin)\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item=ResellerAdmin)\u001b[0m\n\nTASK [service-ks-register : ceph-rgw | Granting user roles] ********************\nTuesday 24 February 2026  01:06:53 +0000 (0:00:07.412)       0:00:34.502 ****** \n\u001b[0;33mchanged: [testbed-manager] => (item=ceph_rgw -> service -> admin)\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;33mtestbed-manager\u001b[0m            : \u001b[0;32mok=9   \u001b[0m \u001b[0;33mchanged=5   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\u001b[0;32mtestbed-node-0\u001b[0m             : \u001b[0;32mok=3   \u001b[0m changed=0    unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\u001b[0;32mtestbed-node-1\u001b[0m             : \u001b[0;32mok=3   \u001b[0m changed=0    unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\u001b[0;32mtestbed-node-2\u001b[0m             : \u001b[0;32mok=3   \u001b[0m changed=0    unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\u001b[0;32mtestbed-node-3\u001b[0m             : \u001b[0;32mok=3   \u001b[0m changed=0    unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\u001b[0;32mtestbed-node-4\u001b[0m             : \u001b[0;32mok=3   \u001b[0m changed=0    unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\u001b[0;32mtestbed-node-5\u001b[0m             : \u001b[0;32mok=3   \u001b[0m changed=0    unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nTuesday 24 February 2026  01:06:58 +0000 (0:00:05.225)       0:00:39.727 ****** \n=============================================================================== \nservice-ks-register : ceph-rgw | Creating roles ------------------------- 7.41s\nservice-ks-register : ceph-rgw | Creating endpoints --------------------- 7.35s\nservice-ks-register : ceph-rgw | Creating users ------------------------- 5.45s\nservice-ks-register : ceph-rgw | Granting user roles -------------------- 5.23s\nceph-rgw : include_tasks ------------------------------------------------ 3.86s\nservice-ks-register : ceph-rgw | Creating services ---------------------- 3.83s\nservice-ks-register : ceph-rgw | Creating projects ---------------------- 3.46s\nGroup hosts based on enabled services ----------------------------------- 1.65s\nGroup hosts based on Kolla action --------------------------------------- 1.06s\n\n\nPLAY [Group hosts based on configuration] **************************************\n\nTASK [Group hosts based on Kolla action] ***************************************\nTuesday 24 February 2026  01:05:22 +0000 (0:00:01.108)       0:00:01.108 ****** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [Group hosts based on enabled services] ***********************************\nTuesday 24 February 2026  01:05:22 +0000 (0:00:00.446)       0:00:01.555 ****** \n\u001b[0;32mok: [testbed-node-0] => (item=enable_magnum_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item=enable_magnum_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item=enable_magnum_True)\u001b[0m\n\nPLAY [Apply role magnum] *******************************************************\n\nTASK [magnum : include_tasks] **************************************************\nTuesday 24 February 2026  01:05:23 +0000 (0:00:00.633)       0:00:02.189 ****** \n\u001b[0;36mincluded: /ansible/roles/magnum/tasks/deploy.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [service-ks-register : magnum | Creating services] ************************\nTuesday 24 February 2026  01:05:23 +0000 (0:00:00.641)       0:00:02.830 ****** \n\u001b[0;33mchanged: [testbed-node-0] => (item=magnum (container-infra))\u001b[0m\n\nTASK [service-ks-register : magnum | Creating endpoints] ***********************\nTuesday 24 February 2026  01:05:28 +0000 (0:00:04.321)       0:00:07.151 ****** \n\u001b[0;33mchanged: [testbed-node-0] => (item=magnum -> https://api-int.testbed.osism.xyz:9511/v1 -> internal)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=magnum -> https://api.testbed.osism.xyz:9511/v1 -> public)\u001b[0m\n\nTASK [service-ks-register : magnum | Creating projects] ************************\nTuesday 24 February 2026  01:05:36 +0000 (0:00:08.209)       0:00:15.361 ****** \n\u001b[0;32mok: [testbed-node-0] => (item=service)\u001b[0m\n\nTASK [service-ks-register : magnum | Creating users] ***************************\nTuesday 24 February 2026  01:05:40 +0000 (0:00:04.070)       0:00:19.432 ****** \n\u001b[0;33mchanged: [testbed-node-0] => (item=magnum -> service)\u001b[0m\n\u001b[1;35m[WARNING]: Module did not set no_log for update_password\u001b[0m\n\nTASK [service-ks-register : magnum | Creating roles] ***************************\nTuesday 24 February 2026  01:05:45 +0000 (0:00:04.896)       0:00:24.328 ****** \n\u001b[0;32mok: [testbed-node-0] => (item=admin)\u001b[0m\n\nTASK [service-ks-register : magnum | Granting user roles] **********************\nTuesday 24 February 2026  01:05:49 +0000 (0:00:04.366)       0:00:28.695 ****** \n\u001b[0;33mchanged: [testbed-node-0] => (item=magnum -> service -> admin)\u001b[0m\n\nTASK [magnum : Creating Magnum trustee domain] *********************************\nTuesday 24 February 2026  01:05:54 +0000 (0:00:04.721)       0:00:33.416 ****** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [magnum : Creating Magnum trustee user] ***********************************\nTuesday 24 February 2026  01:05:58 +0000 (0:00:03.987)       0:00:37.404 ****** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [magnum : Creating Magnum trustee user role] ******************************\nTuesday 24 February 2026  01:06:02 +0000 (0:00:04.211)       0:00:41.615 ****** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [magnum : Ensuring config directories exist] ******************************\nTuesday 24 February 2026  01:06:07 +0000 (0:00:04.403)       0:00:46.019 ****** \n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-api:2024.2', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-api:2024.2', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-api:2024.2', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-conductor:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.11,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-conductor:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.10,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-conductor:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.12,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}})\u001b[0m\n\nTASK [magnum : Check if policies shall be overwritten] *************************\nTuesday 24 February 2026  01:06:08 +0000 (0:00:01.379)       0:00:47.398 ****** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [magnum : Set magnum policy file] *****************************************\nTuesday 24 February 2026  01:06:08 +0000 (0:00:00.144)       0:00:47.543 ****** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [magnum : Check if kubeconfig file is supplied] ***************************\nTuesday 24 February 2026  01:06:09 +0000 (0:00:00.744)       0:00:48.288 ****** \n\u001b[0;32mok: [testbed-node-0 -> localhost]\u001b[0m\n\nTASK [magnum : Copying over kubeconfig file] ***********************************\nTuesday 24 February 2026  01:06:10 +0000 (0:00:01.697)       0:00:49.985 ****** \n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-api:2024.2', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-api:2024.2', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-api:2024.2', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-conductor:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.12,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-conductor:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.11,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-conductor:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.10,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}})\u001b[0m\n\nTASK [magnum : Set magnum kubeconfig file's path] ******************************\nTuesday 24 February 2026  01:06:13 +0000 (0:00:02.739)       0:00:52.724 ****** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [magnum : include_tasks] **************************************************\nTuesday 24 February 2026  01:06:14 +0000 (0:00:00.356)       0:00:53.081 ****** \n\u001b[0;36mincluded: /ansible/roles/magnum/tasks/copy-certs.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [service-cert-copy : magnum | Copying over extra CA certificates] *********\nTuesday 24 February 2026  01:06:15 +0000 (0:00:01.477)       0:00:54.559 ****** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-api:2024.2', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-api:2024.2', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-api:2024.2', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-conductor:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.10,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-conductor:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.11,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-conductor:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.12,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}})\u001b[0m\n\nTASK [service-cert-copy : magnum | Copying over backend internal TLS certificate] ***\nTuesday 24 February 2026  01:06:18 +0000 (0:00:03.228)       0:00:57.787 ****** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-api:2024.2', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-conductor:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.10,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-api:2024.2', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-conductor:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.11,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-api:2024.2', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-conductor:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.12,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [service-cert-copy : magnum | Copying over backend internal TLS key] ******\nTuesday 24 February 2026  01:06:19 +0000 (0:00:01.016)       0:00:58.804 ****** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-api:2024.2', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-api:2024.2', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-conductor:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.10,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-conductor:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.11,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-api:2024.2', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-conductor:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.12,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [magnum : Copying over config.json files for services] ********************\nTuesday 24 February 2026  01:06:22 +0000 (0:00:02.593)       0:01:01.397 ****** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-api:2024.2', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-api:2024.2', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-api:2024.2', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-conductor:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.10,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-conductor:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.12,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-conductor:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.11,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}})\u001b[0m\n\nTASK [magnum : Copying over magnum.conf] ***************************************\nTuesday 24 February 2026  01:06:26 +0000 (0:00:03.834)       0:01:05.232 ****** \n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-api:2024.2', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-api:2024.2', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-api:2024.2', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-conductor:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.12,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-conductor:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.10,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-conductor:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.11,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}})\u001b[0m\n\nTASK [magnum : Copying over existing policy file] ******************************\nTuesday 24 February 2026  01:06:36 +0000 (0:00:10.617)       0:01:15.850 ****** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-api:2024.2', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-conductor:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.10,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-api:2024.2', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-conductor:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.11,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-api:2024.2', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-conductor:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.12,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [magnum : Check magnum containers] ****************************************\nTuesday 24 February 2026  01:06:37 +0000 (0:00:00.662)       0:01:16.512 ****** \n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-api:2024.2', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-api:2024.2', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-api:2024.2', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-conductor:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.11,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-conductor:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.10,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-conductor:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.12,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}})\u001b[0m\n\nTASK [magnum : include_tasks] **************************************************\nTuesday 24 February 2026  01:06:39 +0000 (0:00:02.405)       0:01:18.917 ****** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [magnum : Creating Magnum database] ***************************************\nTuesday 24 February 2026  01:06:40 +0000 (0:00:00.530)       0:01:19.447 ****** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [magnum : Creating Magnum database user and setting permissions] **********\nTuesday 24 February 2026  01:06:43 +0000 (0:00:02.683)       0:01:22.130 ****** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [magnum : Running Magnum bootstrap container] *****************************\nTuesday 24 February 2026  01:06:46 +0000 (0:00:03.531)       0:01:25.662 ****** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [magnum : Flush handlers] *************************************************\nTuesday 24 February 2026  01:07:05 +0000 (0:00:18.495)       0:01:44.157 ****** \n\nTASK [magnum : Flush handlers] *************************************************\nTuesday 24 February 2026  01:07:05 +0000 (0:00:00.140)       0:01:44.298 ****** \n\nTASK [magnum : Flush handlers] *************************************************\nTuesday 24 February 2026  01:07:05 +0000 (0:00:00.068)       0:01:44.367 ****** \n\nRUNNING HANDLER [magnum : Restart magnum-api container] ************************\nTuesday 24 February 2026  01:07:05 +0000 (0:00:00.106)       0:01:44.473 ****** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [magnum : Restart magnum-conductor container] ******************\nTuesday 24 February 2026  01:07:21 +0000 (0:00:15.914)       0:02:00.388 ****** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=26  \u001b[0m \u001b[0;33mchanged=18  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=6   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=13  \u001b[0m \u001b[0;33mchanged=8   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=5   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=13  \u001b[0m \u001b[0;33mchanged=8   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=5   \u001b[0m rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nTuesday 24 February 2026  01:07:35 +0000 (0:00:13.930)       0:02:14.319 ****** \n=============================================================================== \nmagnum : Running Magnum bootstrap container ---------------------------- 18.50s\nmagnum : Restart magnum-api container ---------------------------------- 15.91s\nmagnum : Restart magnum-conductor container ---------------------------- 13.93s\nmagnum : Copying over magnum.conf -------------------------------------- 10.62s\nservice-ks-register : magnum | Creating endpoints ----------------------- 8.21s\nservice-ks-register : magnum | Creating users --------------------------- 4.90s\nservice-ks-register : magnum | Granting user roles ---------------------- 4.72s\nmagnum : Creating Magnum trustee user role ------------------------------ 4.40s\nservice-ks-register : magnum | Creating roles --------------------------- 4.37s\nservice-ks-register : magnum | Creating services ------------------------ 4.32s\nmagnum : Creating Magnum trustee user ----------------------------------- 4.21s\nservice-ks-register : magnum | Creating projects ------------------------ 4.07s\nmagnum : Creating Magnum trustee domain --------------------------------- 3.99s\nmagnum : Copying over config.json files for services -------------------- 3.83s\nmagnum : Creating Magnum database user and setting permissions ---------- 3.53s\nservice-cert-copy : magnum | Copying over extra CA certificates --------- 3.23s\nmagnum : Copying over kubeconfig file ----------------------------------- 2.74s\nmagnum : Creating Magnum database --------------------------------------- 2.68s\nservice-cert-copy : magnum | Copying over backend internal TLS key ------ 2.59s\nmagnum : Check magnum containers ---------------------------------------- 2.41s\n\u001b[32m2026-02-24 01:07:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a29489db-95ad-418c-bc67-f87a0737c252 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:07:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44aa611b-f898-4352-b174-ada638d8f758 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:07:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 01:07:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fba6bf07-65a1-432e-9a3d-72209557a21d is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:07:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask eddb1a5f-448f-4517-b064-c30da006d691 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:07:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a29489db-95ad-418c-bc67-f87a0737c252 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:07:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44aa611b-f898-4352-b174-ada638d8f758 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:07:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 01:07:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fba6bf07-65a1-432e-9a3d-72209557a21d is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:07:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask eddb1a5f-448f-4517-b064-c30da006d691 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:07:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a29489db-95ad-418c-bc67-f87a0737c252 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:07:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44aa611b-f898-4352-b174-ada638d8f758 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:07:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 01:07:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fba6bf07-65a1-432e-9a3d-72209557a21d is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:07:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask eddb1a5f-448f-4517-b064-c30da006d691 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:07:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a29489db-95ad-418c-bc67-f87a0737c252 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:07:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44aa611b-f898-4352-b174-ada638d8f758 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:07:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 01:07:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fba6bf07-65a1-432e-9a3d-72209557a21d is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:07:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask eddb1a5f-448f-4517-b064-c30da006d691 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:07:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a29489db-95ad-418c-bc67-f87a0737c252 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:07:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44aa611b-f898-4352-b174-ada638d8f758 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:07:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 01:07:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fba6bf07-65a1-432e-9a3d-72209557a21d is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:07:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask eddb1a5f-448f-4517-b064-c30da006d691 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:07:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a29489db-95ad-418c-bc67-f87a0737c252 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:07:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4baf3a0c-387b-4215-b490-5edd972623e9 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:07:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44aa611b-f898-4352-b174-ada638d8f758 is in state SUCCESS\u001b[0m\n\u001b[32m2026-02-24 01:07:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\n\nPLAY [Group hosts based on configuration] **************************************\n\nTASK [Group hosts based on Kolla action] ***************************************\nTuesday 24 February 2026  01:02:32 +0000 (0:00:00.205)       0:00:00.205 ****** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [Group hosts based on enabled services] ***********************************\nTuesday 24 February 2026  01:02:32 +0000 (0:00:00.706)       0:00:00.912 ****** \n\u001b[0;32mok: [testbed-node-0] => (item=enable_neutron_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item=enable_neutron_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item=enable_neutron_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item=enable_neutron_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item=enable_neutron_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item=enable_neutron_True)\u001b[0m\n\nPLAY [Apply role neutron] ******************************************************\n\nTASK [neutron : include_tasks] *************************************************\nTuesday 24 February 2026  01:02:33 +0000 (0:00:00.636)       0:00:01.549 ****** \n\u001b[0;36mincluded: /ansible/roles/neutron/tasks/deploy.yml for testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m\n\nTASK [neutron : Get container facts] *******************************************\nTuesday 24 February 2026  01:02:34 +0000 (0:00:01.217)       0:00:02.766 ****** \n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [neutron : Get container volume facts] ************************************\nTuesday 24 February 2026  01:02:35 +0000 (0:00:01.244)       0:00:04.010 ****** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [neutron : Check for ML2/OVN presence] ************************************\nTuesday 24 February 2026  01:02:37 +0000 (0:00:01.199)       0:00:05.210 ****** \n\u001b[0;32mok: [testbed-node-0] => {\u001b[0m\n\u001b[0;32m    \"changed\": false,\u001b[0m\n\u001b[0;32m    \"msg\": \"All assertions passed\"\u001b[0m\n\u001b[0;32m}\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => {\u001b[0m\n\u001b[0;32m    \"changed\": false,\u001b[0m\n\u001b[0;32m    \"msg\": \"All assertions passed\"\u001b[0m\n\u001b[0;32m}\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => {\u001b[0m\n\u001b[0;32m    \"changed\": false,\u001b[0m\n\u001b[0;32m    \"msg\": \"All assertions passed\"\u001b[0m\n\u001b[0;32m}\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => {\u001b[0m\n\u001b[0;32m    \"changed\": false,\u001b[0m\n\u001b[0;32m    \"msg\": \"All assertions passed\"\u001b[0m\n\u001b[0;32m}\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => {\u001b[0m\n\u001b[0;32m    \"changed\": false,\u001b[0m\n\u001b[0;32m    \"msg\": \"All assertions passed\"\u001b[0m\n\u001b[0;32m}\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => {\u001b[0m\n\u001b[0;32m    \"changed\": false,\u001b[0m\n\u001b[0;32m    \"msg\": \"All assertions passed\"\u001b[0m\n\u001b[0;32m}\u001b[0m\n\nTASK [neutron : Check for ML2/OVS presence] ************************************\nTuesday 24 February 2026  01:02:37 +0000 (0:00:00.827)       0:00:06.037 ****** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [service-ks-register : neutron | Creating services] ***********************\nTuesday 24 February 2026  01:02:38 +0000 (0:00:00.694)       0:00:06.731 ****** \n\u001b[0;33mchanged: [testbed-node-0] => (item=neutron (network))\u001b[0m\n\nTASK [service-ks-register : neutron | Creating endpoints] **********************\nTuesday 24 February 2026  01:02:42 +0000 (0:00:03.804)       0:00:10.536 ****** \n\u001b[0;33mchanged: [testbed-node-0] => (item=neutron -> https://api-int.testbed.osism.xyz:9696 -> internal)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=neutron -> https://api.testbed.osism.xyz:9696 -> public)\u001b[0m\n\nTASK [service-ks-register : neutron | Creating projects] ***********************\nTuesday 24 February 2026  01:02:50 +0000 (0:00:07.910)       0:00:18.446 ****** \n\u001b[0;32mok: [testbed-node-0] => (item=service)\u001b[0m\n\nTASK [service-ks-register : neutron | Creating users] **************************\nTuesday 24 February 2026  01:02:54 +0000 (0:00:03.900)       0:00:22.346 ****** \n\u001b[0;33mchanged: [testbed-node-0] => (item=neutron -> service)\u001b[0m\n\u001b[1;35m[WARNING]: Module did not set no_log for update_password\u001b[0m\n\nTASK [service-ks-register : neutron | Creating roles] **************************\nTuesday 24 February 2026  01:02:58 +0000 (0:00:04.487)       0:00:26.834 ****** \n\u001b[0;32mok: [testbed-node-0] => (item=admin)\u001b[0m\n\nTASK [service-ks-register : neutron | Granting user roles] *********************\nTuesday 24 February 2026  01:03:02 +0000 (0:00:04.108)       0:00:30.943 ****** \n\u001b[0;33mchanged: [testbed-node-0] => (item=neutron -> service -> admin)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=neutron -> service -> service)\u001b[0m\n\nTASK [neutron : include_tasks] *************************************************\nTuesday 24 February 2026  01:03:11 +0000 (0:00:08.840)       0:00:39.783 ****** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Load and persist kernel modules] *****************************************\nTuesday 24 February 2026  01:03:12 +0000 (0:00:00.924)       0:00:40.707 ****** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [neutron : Check IPv6 support] ********************************************\nTuesday 24 February 2026  01:03:15 +0000 (0:00:02.463)       0:00:43.171 ****** \n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [Setting sysctl values] ***************************************************\nTuesday 24 February 2026  01:03:16 +0000 (0:00:01.462)       0:00:44.633 ****** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [neutron : Ensuring config directories exist] *****************************\nTuesday 24 February 2026  01:03:20 +0000 (0:00:03.540)       0:00:48.174 ****** \n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2024.2', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2024.2', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2024.2', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}})\u001b[0m\n\nTASK [neutron : Check if extra ml2 plugins exists] *****************************\nTuesday 24 February 2026  01:03:23 +0000 (0:00:03.577)       0:00:51.752 ****** \n\u001b[1;35m[WARNING]: Skipped\u001b[0m\n\u001b[1;35m'/opt/configuration/environments/kolla/files/overlays/neutron/plugins/' path\u001b[0m\n\u001b[1;35mdue to this access issue:\u001b[0m\n\u001b[1;35m'/opt/configuration/environments/kolla/files/overlays/neutron/plugins/' is not\u001b[0m\n\u001b[1;35ma directory\u001b[0m\n\u001b[0;32mok: [testbed-node-0 -> localhost]\u001b[0m\n\nTASK [neutron : include_tasks] *************************************************\nTuesday 24 February 2026  01:03:24 +0000 (0:00:01.044)       0:00:52.796 ****** \n\u001b[0;36mincluded: /ansible/roles/neutron/tasks/copy-certs.yml for testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m\n\nTASK [service-cert-copy : neutron | Copying over extra CA certificates] ********\nTuesday 24 February 2026  01:03:26 +0000 (0:00:01.417)       0:00:54.214 ****** \n\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2024.2', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2024.2', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2024.2', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}})\u001b[0m\n\nTASK [service-cert-copy : neutron | Copying over backend internal TLS certificate] ***\nTuesday 24 February 2026  01:03:30 +0000 (0:00:04.087)       0:00:58.302 ****** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2024.2', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2024.2', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2024.2', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [service-cert-copy : neutron | Copying over backend internal TLS key] *****\nTuesday 24 February 2026  01:03:34 +0000 (0:00:04.521)       0:01:02.823 ****** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2024.2', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2024.2', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2024.2', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [neutron : Creating TLS backend PEM File] *********************************\nTuesday 24 February 2026  01:03:39 +0000 (0:00:04.500)       0:01:07.324 ****** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [neutron : Check if policies shall be overwritten] ************************\nTuesday 24 February 2026  01:03:42 +0000 (0:00:03.343)       0:01:10.668 ****** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [neutron : Set neutron policy file] ***************************************\nTuesday 24 February 2026  01:03:42 +0000 (0:00:00.137)       0:01:10.805 ****** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [neutron : Copying over existing policy file] *****************************\nTuesday 24 February 2026  01:03:43 +0000 (0:00:00.802)       0:01:11.608 ****** \n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2024.2', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2024.2', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2024.2', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [neutron : Copying over config.json files for services] *******************\nTuesday 24 February 2026  01:03:46 +0000 (0:00:03.366)       0:01:14.975 ****** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2024.2', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2024.2', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2024.2', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}})\u001b[0m\n\nTASK [neutron : Copying over neutron.conf] *************************************\nTuesday 24 February 2026  01:03:51 +0000 (0:00:04.364)       0:01:19.340 ****** \n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2024.2', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2024.2', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2024.2', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}})\u001b[0m\n\nTASK [neutron : Copying over neutron_vpnaas.conf] ******************************\nTuesday 24 February 2026  01:03:59 +0000 (0:00:08.285)       0:01:27.626 ****** \n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2024.2', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2024.2', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2024.2', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [neutron : Copying over ssh key] ******************************************\nTuesday 24 February 2026  01:04:03 +0000 (0:00:04.009)       0:01:31.636 ****** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [neutron : Copying over ml2_conf.ini] *************************************\nTuesday 24 February 2026  01:04:07 +0000 (0:00:03.835)       0:01:35.471 ****** \n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2024.2', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2024.2', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2024.2', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}})\u001b[0m\n\nTASK [neutron : Copying over linuxbridge_agent.ini] ****************************\nTuesday 24 February 2026  01:04:14 +0000 (0:00:06.979)       0:01:42.451 ****** \n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [neutron : Copying over openvswitch_agent.ini] ****************************\nTuesday 24 February 2026  01:04:18 +0000 (0:00:04.459)       0:01:46.910 ****** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [neutron : Copying over sriov_agent.ini] **********************************\nTuesday 24 February 2026  01:04:22 +0000 (0:00:03.744)       0:01:50.654 ****** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [neutron : Copying over mlnx_agent.ini] ***********************************\nTuesday 24 February 2026  01:04:25 +0000 (0:00:02.611)       0:01:53.266 ****** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [neutron : Copying over eswitchd.conf] ************************************\nTuesday 24 February 2026  01:04:27 +0000 (0:00:02.467)       0:01:55.734 ****** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [neutron : Copying over dhcp_agent.ini] ***********************************\nTuesday 24 February 2026  01:04:30 +0000 (0:00:03.238)       0:01:58.973 ****** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [neutron : Copying over dnsmasq.conf] *************************************\nTuesday 24 February 2026  01:04:33 +0000 (0:00:02.532)       0:02:01.506 ****** \n\u001b[0;36mskipping: [testbed-node-1] => (item=/ansible/roles/neutron/templates/dnsmasq.conf.j2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=/ansible/roles/neutron/templates/dnsmasq.conf.j2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=/ansible/roles/neutron/templates/dnsmasq.conf.j2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=/ansible/roles/neutron/templates/dnsmasq.conf.j2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=/ansible/roles/neutron/templates/dnsmasq.conf.j2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=/ansible/roles/neutron/templates/dnsmasq.conf.j2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [neutron : Copying over l3_agent.ini] *************************************\nTuesday 24 February 2026  01:04:35 +0000 (0:00:02.547)       0:02:04.053 ****** \n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2024.2', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2024.2', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2024.2', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [neutron : Copying over fwaas_driver.ini] *********************************\nTuesday 24 February 2026  01:04:38 +0000 (0:00:02.698)       0:02:06.752 ****** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2024.2', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2024.2', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2024.2', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [neutron : Copying over metadata_agent.ini] *******************************\nTuesday 24 February 2026  01:04:41 +0000 (0:00:03.254)       0:02:10.006 ****** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [neutron : Copying over neutron_ovn_metadata_agent.ini] *******************\nTuesday 24 February 2026  01:04:44 +0000 (0:00:02.323)       0:02:12.330 ****** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nTASK [neutron : Copying over metering_agent.ini] *******************************\nTuesday 24 February 2026  01:04:48 +0000 (0:00:04.486)       0:02:16.816 ****** \n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [neutron : Copying over ironic_neutron_agent.ini] *************************\nTuesday 24 February 2026  01:04:52 +0000 (0:00:03.925)       0:02:20.742 ****** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [neutron : Copying over bgp_dragent.ini] **********************************\nTuesday 24 February 2026  01:04:55 +0000 (0:00:03.198)       0:02:23.940 ****** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [neutron : Copying over ovn_agent.ini] ************************************\nTuesday 24 February 2026  01:04:58 +0000 (0:00:02.808)       0:02:26.749 ****** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [neutron : Copying over nsx.ini] ******************************************\nTuesday 24 February 2026  01:05:02 +0000 (0:00:03.648)       0:02:30.398 ****** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [neutron : Copy neutron-l3-agent-wrapper script] **************************\nTuesday 24 February 2026  01:05:07 +0000 (0:00:05.364)       0:02:35.762 ****** \n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [neutron : Copying over extra ml2 plugins] ********************************\nTuesday 24 February 2026  01:05:11 +0000 (0:00:03.693)       0:02:39.456 ****** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [neutron : Copying over neutron-tls-proxy.cfg] ****************************\nTuesday 24 February 2026  01:05:14 +0000 (0:00:03.127)       0:02:42.584 ****** \n\u001b[0;36mskipping: [testbed-node-1] => (item=/ansible/roles/neutron/templates/neutron-tls-proxy.cfg.j2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=/ansible/roles/neutron/templates/neutron-tls-proxy.cfg.j2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=/ansible/roles/neutron/templates/neutron-tls-proxy.cfg.j2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=/ansible/roles/neutron/templates/neutron-tls-proxy.cfg.j2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=/ansible/roles/neutron/templates/neutron-tls-proxy.cfg.j2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=/ansible/roles/neutron/templates/neutron-tls-proxy.cfg.j2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [neutron : Copying over neutron_taas.conf] ********************************\nTuesday 24 February 2026  01:05:19 +0000 (0:00:04.731)       0:02:47.315 ****** \n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2024.2', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2024.2', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2024.2', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [neutron : Check neutron containers] **************************************\nTuesday 24 February 2026  01:05:23 +0000 (0:00:04.114)       0:02:51.429 ****** \n\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2024.2', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2024.2', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2024.2', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}})\u001b[0m\n\nTASK [neutron : include_tasks] *************************************************\nTuesday 24 February 2026  01:05:26 +0000 (0:00:03.510)       0:02:54.939 ****** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [neutron : Creating Neutron database] *************************************\nTuesday 24 February 2026  01:05:27 +0000 (0:00:00.608)       0:02:55.548 ****** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [neutron : Creating Neutron database user and setting permissions] ********\nTuesday 24 February 2026  01:05:29 +0000 (0:00:02.512)       0:02:58.061 ****** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [neutron : Running Neutron bootstrap container] ***************************\nTuesday 24 February 2026  01:05:32 +0000 (0:00:02.915)       0:03:00.976 ****** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [neutron : Flush Handlers] ************************************************\nTuesday 24 February 2026  01:06:20 +0000 (0:00:47.724)       0:03:48.700 ****** \n\nTASK [neutron : Flush Handlers] ************************************************\nTuesday 24 February 2026  01:06:20 +0000 (0:00:00.069)       0:03:48.770 ****** \n\nTASK [neutron : Flush Handlers] ************************************************\nTuesday 24 February 2026  01:06:21 +0000 (0:00:00.397)       0:03:49.167 ****** \n\nTASK [neutron : Flush Handlers] ************************************************\nTuesday 24 February 2026  01:06:21 +0000 (0:00:00.074)       0:03:49.241 ****** \n\nTASK [neutron : Flush Handlers] ************************************************\nTuesday 24 February 2026  01:06:21 +0000 (0:00:00.120)       0:03:49.361 ****** \n\nTASK [neutron : Flush Handlers] ************************************************\nTuesday 24 February 2026  01:06:21 +0000 (0:00:00.135)       0:03:49.496 ****** \n\nRUNNING HANDLER [neutron : Restart neutron-server container] *******************\nTuesday 24 February 2026  01:06:21 +0000 (0:00:00.146)       0:03:49.643 ****** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\nRUNNING HANDLER [neutron : Restart neutron-ovn-metadata-agent container] *******\nTuesday 24 February 2026  01:06:56 +0000 (0:00:34.425)       0:04:24.068 ****** \n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=26  \u001b[0m \u001b[0;33mchanged=15  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=32  \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=16  \u001b[0m \u001b[0;33mchanged=8   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=31  \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=16  \u001b[0m \u001b[0;33mchanged=8   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=31  \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-3\u001b[0m             : \u001b[0;32mok=15  \u001b[0m \u001b[0;33mchanged=7   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=32  \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-4\u001b[0m             : \u001b[0;32mok=15  \u001b[0m \u001b[0;33mchanged=7   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=32  \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-5\u001b[0m             : \u001b[0;32mok=15  \u001b[0m \u001b[0;33mchanged=7   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=32  \u001b[0m rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nTuesday 24 February 2026  01:07:50 +0000 (0:00:54.110)       0:05:18.179 ****** \n=============================================================================== \nneutron : Restart neutron-ovn-metadata-agent container ----------------- 54.11s\nneutron : Running Neutron bootstrap container -------------------------- 47.72s\nneutron : Restart neutron-server container ----------------------------- 34.43s\nservice-ks-register : neutron | Granting user roles --------------------- 8.84s\nneutron : Copying over neutron.conf ------------------------------------- 8.29s\nservice-ks-register : neutron | Creating endpoints ---------------------- 7.91s\nneutron : Copying over ml2_conf.ini ------------------------------------- 6.98s\nneutron : Copying over nsx.ini ------------------------------------------ 5.36s\nneutron : Copying over neutron-tls-proxy.cfg ---------------------------- 4.73s\nservice-cert-copy : neutron | Copying over backend internal TLS certificate --- 4.52s\nservice-cert-copy : neutron | Copying over backend internal TLS key ----- 4.50s\nservice-ks-register : neutron | Creating users -------------------------- 4.49s\nneutron : Copying over neutron_ovn_metadata_agent.ini ------------------- 4.49s\nneutron : Copying over linuxbridge_agent.ini ---------------------------- 4.46s\nneutron : Copying over config.json files for services ------------------- 4.36s\nneutron : Copying over neutron_taas.conf -------------------------------- 4.11s\nservice-ks-register : neutron | Creating roles -------------------------- 4.11s\nservice-cert-copy : neutron | Copying over extra CA certificates -------- 4.09s\nneutron : Copying over neutron_vpnaas.conf ------------------------------ 4.01s\nneutron : Copying over metering_agent.ini ------------------------------- 3.93s\n\u001b[32m2026-02-24 01:07:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fba6bf07-65a1-432e-9a3d-72209557a21d is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:07:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask eddb1a5f-448f-4517-b064-c30da006d691 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:07:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a29489db-95ad-418c-bc67-f87a0737c252 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:07:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4baf3a0c-387b-4215-b490-5edd972623e9 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:07:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 01:07:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fba6bf07-65a1-432e-9a3d-72209557a21d is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:07:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask eddb1a5f-448f-4517-b064-c30da006d691 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:07:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a29489db-95ad-418c-bc67-f87a0737c252 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:07:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4baf3a0c-387b-4215-b490-5edd972623e9 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:07:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 01:08:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fba6bf07-65a1-432e-9a3d-72209557a21d is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:08:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask eddb1a5f-448f-4517-b064-c30da006d691 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:08:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a29489db-95ad-418c-bc67-f87a0737c252 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:08:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4baf3a0c-387b-4215-b490-5edd972623e9 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:08:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 01:08:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fba6bf07-65a1-432e-9a3d-72209557a21d is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:08:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask eddb1a5f-448f-4517-b064-c30da006d691 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:08:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a29489db-95ad-418c-bc67-f87a0737c252 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:08:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4baf3a0c-387b-4215-b490-5edd972623e9 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:08:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 01:08:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fba6bf07-65a1-432e-9a3d-72209557a21d is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:08:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask eddb1a5f-448f-4517-b064-c30da006d691 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:08:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a29489db-95ad-418c-bc67-f87a0737c252 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:08:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4baf3a0c-387b-4215-b490-5edd972623e9 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:08:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 01:08:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fba6bf07-65a1-432e-9a3d-72209557a21d is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:08:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask eddb1a5f-448f-4517-b064-c30da006d691 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:08:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a29489db-95ad-418c-bc67-f87a0737c252 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:08:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4baf3a0c-387b-4215-b490-5edd972623e9 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:08:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 01:08:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fba6bf07-65a1-432e-9a3d-72209557a21d is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:08:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask eddb1a5f-448f-4517-b064-c30da006d691 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:08:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a29489db-95ad-418c-bc67-f87a0737c252 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:08:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4baf3a0c-387b-4215-b490-5edd972623e9 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:08:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 01:08:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fba6bf07-65a1-432e-9a3d-72209557a21d is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:08:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask eddb1a5f-448f-4517-b064-c30da006d691 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:08:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a29489db-95ad-418c-bc67-f87a0737c252 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:08:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4baf3a0c-387b-4215-b490-5edd972623e9 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:08:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 01:08:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fba6bf07-65a1-432e-9a3d-72209557a21d is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:08:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask eddb1a5f-448f-4517-b064-c30da006d691 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:08:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a29489db-95ad-418c-bc67-f87a0737c252 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:08:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4baf3a0c-387b-4215-b490-5edd972623e9 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:08:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 01:08:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fba6bf07-65a1-432e-9a3d-72209557a21d is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:08:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask eddb1a5f-448f-4517-b064-c30da006d691 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:08:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a29489db-95ad-418c-bc67-f87a0737c252 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:08:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4baf3a0c-387b-4215-b490-5edd972623e9 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:08:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 01:08:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fba6bf07-65a1-432e-9a3d-72209557a21d is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:08:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask eddb1a5f-448f-4517-b064-c30da006d691 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:08:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a29489db-95ad-418c-bc67-f87a0737c252 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:08:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4baf3a0c-387b-4215-b490-5edd972623e9 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:08:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 01:08:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fba6bf07-65a1-432e-9a3d-72209557a21d is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:08:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask eddb1a5f-448f-4517-b064-c30da006d691 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:08:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a29489db-95ad-418c-bc67-f87a0737c252 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:08:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4baf3a0c-387b-4215-b490-5edd972623e9 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:08:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 01:08:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fba6bf07-65a1-432e-9a3d-72209557a21d is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:08:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask eddb1a5f-448f-4517-b064-c30da006d691 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:08:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a29489db-95ad-418c-bc67-f87a0737c252 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:08:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4baf3a0c-387b-4215-b490-5edd972623e9 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:08:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 01:08:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fba6bf07-65a1-432e-9a3d-72209557a21d is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:08:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask eddb1a5f-448f-4517-b064-c30da006d691 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:08:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a29489db-95ad-418c-bc67-f87a0737c252 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:08:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4baf3a0c-387b-4215-b490-5edd972623e9 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:08:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 01:08:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fba6bf07-65a1-432e-9a3d-72209557a21d is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:08:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask eddb1a5f-448f-4517-b064-c30da006d691 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:08:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a29489db-95ad-418c-bc67-f87a0737c252 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:08:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4baf3a0c-387b-4215-b490-5edd972623e9 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:08:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 01:08:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fba6bf07-65a1-432e-9a3d-72209557a21d is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:08:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask eddb1a5f-448f-4517-b064-c30da006d691 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:08:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a29489db-95ad-418c-bc67-f87a0737c252 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:08:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4baf3a0c-387b-4215-b490-5edd972623e9 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:08:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 01:08:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fba6bf07-65a1-432e-9a3d-72209557a21d is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:08:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask eddb1a5f-448f-4517-b064-c30da006d691 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:08:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a29489db-95ad-418c-bc67-f87a0737c252 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:08:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4baf3a0c-387b-4215-b490-5edd972623e9 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:08:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 01:08:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fba6bf07-65a1-432e-9a3d-72209557a21d is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:08:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask eddb1a5f-448f-4517-b064-c30da006d691 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:08:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a29489db-95ad-418c-bc67-f87a0737c252 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:08:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4baf3a0c-387b-4215-b490-5edd972623e9 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:08:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 01:08:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fba6bf07-65a1-432e-9a3d-72209557a21d is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:08:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask eddb1a5f-448f-4517-b064-c30da006d691 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:08:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a29489db-95ad-418c-bc67-f87a0737c252 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:08:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4baf3a0c-387b-4215-b490-5edd972623e9 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:08:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 01:08:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fba6bf07-65a1-432e-9a3d-72209557a21d is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:08:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask eddb1a5f-448f-4517-b064-c30da006d691 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:08:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a29489db-95ad-418c-bc67-f87a0737c252 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:08:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4baf3a0c-387b-4215-b490-5edd972623e9 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:08:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 01:08:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fba6bf07-65a1-432e-9a3d-72209557a21d is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:08:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask eddb1a5f-448f-4517-b064-c30da006d691 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:08:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a29489db-95ad-418c-bc67-f87a0737c252 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:08:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4baf3a0c-387b-4215-b490-5edd972623e9 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:08:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 01:08:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fba6bf07-65a1-432e-9a3d-72209557a21d is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:08:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask eddb1a5f-448f-4517-b064-c30da006d691 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:08:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a29489db-95ad-418c-bc67-f87a0737c252 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:08:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4baf3a0c-387b-4215-b490-5edd972623e9 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:08:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 01:09:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fba6bf07-65a1-432e-9a3d-72209557a21d is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:09:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask eddb1a5f-448f-4517-b064-c30da006d691 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:09:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a29489db-95ad-418c-bc67-f87a0737c252 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:09:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4baf3a0c-387b-4215-b490-5edd972623e9 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:09:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 01:09:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fba6bf07-65a1-432e-9a3d-72209557a21d is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:09:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask eddb1a5f-448f-4517-b064-c30da006d691 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:09:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a29489db-95ad-418c-bc67-f87a0737c252 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:09:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4baf3a0c-387b-4215-b490-5edd972623e9 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:09:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 01:09:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fba6bf07-65a1-432e-9a3d-72209557a21d is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:09:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask eddb1a5f-448f-4517-b064-c30da006d691 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:09:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a29489db-95ad-418c-bc67-f87a0737c252 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:09:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4baf3a0c-387b-4215-b490-5edd972623e9 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:09:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 01:09:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fba6bf07-65a1-432e-9a3d-72209557a21d is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:09:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask eddb1a5f-448f-4517-b064-c30da006d691 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:09:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a29489db-95ad-418c-bc67-f87a0737c252 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:09:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4baf3a0c-387b-4215-b490-5edd972623e9 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:09:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 01:09:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fba6bf07-65a1-432e-9a3d-72209557a21d is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:09:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask eddb1a5f-448f-4517-b064-c30da006d691 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:09:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a29489db-95ad-418c-bc67-f87a0737c252 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:09:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4baf3a0c-387b-4215-b490-5edd972623e9 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:09:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 01:09:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fba6bf07-65a1-432e-9a3d-72209557a21d is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:09:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask eddb1a5f-448f-4517-b064-c30da006d691 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:09:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a29489db-95ad-418c-bc67-f87a0737c252 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:09:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4baf3a0c-387b-4215-b490-5edd972623e9 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:09:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 01:09:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fba6bf07-65a1-432e-9a3d-72209557a21d is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:09:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask eddb1a5f-448f-4517-b064-c30da006d691 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:09:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a29489db-95ad-418c-bc67-f87a0737c252 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:09:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4baf3a0c-387b-4215-b490-5edd972623e9 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:09:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 01:09:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fba6bf07-65a1-432e-9a3d-72209557a21d is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:09:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask eddb1a5f-448f-4517-b064-c30da006d691 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:09:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a29489db-95ad-418c-bc67-f87a0737c252 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:09:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4baf3a0c-387b-4215-b490-5edd972623e9 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:09:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 01:09:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fba6bf07-65a1-432e-9a3d-72209557a21d is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:09:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask eddb1a5f-448f-4517-b064-c30da006d691 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:09:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a29489db-95ad-418c-bc67-f87a0737c252 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:09:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4baf3a0c-387b-4215-b490-5edd972623e9 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:09:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 01:09:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fba6bf07-65a1-432e-9a3d-72209557a21d is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:09:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask eddb1a5f-448f-4517-b064-c30da006d691 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:09:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a29489db-95ad-418c-bc67-f87a0737c252 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:09:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4baf3a0c-387b-4215-b490-5edd972623e9 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:09:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 01:09:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fba6bf07-65a1-432e-9a3d-72209557a21d is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:09:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask eddb1a5f-448f-4517-b064-c30da006d691 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:09:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a29489db-95ad-418c-bc67-f87a0737c252 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:09:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4baf3a0c-387b-4215-b490-5edd972623e9 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:09:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 01:09:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fba6bf07-65a1-432e-9a3d-72209557a21d is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:09:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask eddb1a5f-448f-4517-b064-c30da006d691 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:09:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a29489db-95ad-418c-bc67-f87a0737c252 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:09:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4baf3a0c-387b-4215-b490-5edd972623e9 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:09:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 01:09:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fba6bf07-65a1-432e-9a3d-72209557a21d is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:09:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask eddb1a5f-448f-4517-b064-c30da006d691 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:09:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a29489db-95ad-418c-bc67-f87a0737c252 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:09:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4baf3a0c-387b-4215-b490-5edd972623e9 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:09:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 01:09:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fba6bf07-65a1-432e-9a3d-72209557a21d is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:09:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask eddb1a5f-448f-4517-b064-c30da006d691 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:09:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a29489db-95ad-418c-bc67-f87a0737c252 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:09:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4baf3a0c-387b-4215-b490-5edd972623e9 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:09:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 01:09:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fba6bf07-65a1-432e-9a3d-72209557a21d is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:09:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask eddb1a5f-448f-4517-b064-c30da006d691 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:09:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a29489db-95ad-418c-bc67-f87a0737c252 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:09:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4baf3a0c-387b-4215-b490-5edd972623e9 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:09:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 01:09:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fba6bf07-65a1-432e-9a3d-72209557a21d is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:09:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask eddb1a5f-448f-4517-b064-c30da006d691 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:09:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask de6aa5ad-1dfd-4a58-bf00-4755d79ef102 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:09:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a29489db-95ad-418c-bc67-f87a0737c252 is in state SUCCESS\u001b[0m\n\n\nPLAY [Group hosts based on configuration] **************************************\n\nTASK [Group hosts based on Kolla action] ***************************************\nTuesday 24 February 2026  01:06:05 +0000 (0:00:00.293)       0:00:00.293 ****** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [Group hosts based on enabled services] ***********************************\nTuesday 24 February 2026  01:06:06 +0000 (0:00:01.006)       0:00:01.300 ****** \n\u001b[0;32mok: [testbed-manager] => (item=enable_prometheus_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-0] => (item=enable_prometheus_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item=enable_prometheus_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item=enable_prometheus_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item=enable_prometheus_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item=enable_prometheus_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item=enable_prometheus_True)\u001b[0m\n\nPLAY [Apply role prometheus] ***************************************************\n\nTASK [prometheus : include_tasks] **********************************************\nTuesday 24 February 2026  01:06:06 +0000 (0:00:00.776)       0:00:02.077 ****** \n\u001b[0;36mincluded: /ansible/roles/prometheus/tasks/deploy.yml for testbed-manager, testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m\n\nTASK [prometheus : Ensuring config directories exist] **************************\nTuesday 24 February 2026  01:06:08 +0000 (0:00:01.785)       0:00:03.863 ****** \n\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'prometheus-server', 'value': {'container_name': 'prometheus_server', 'group': 'prometheus', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-v2-server:2024.2', 'volumes': ['/etc/kolla/prometheus-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'prometheus_v2:/var/lib/prometheus', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'prometheus_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9091', 'active_passive': True}, 'prometheus_server_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9091', 'listen_port': '9091', 'active_passive': True}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2024.2', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2024.2', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2024.2', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2024.2', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'prometheus-mysqld-exporter', 'value': {'container_name': 'prometheus_mysqld_exporter', 'group': 'prometheus-mysqld-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-mysqld-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-mysqld-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'prometheus-mysqld-exporter', 'value': {'container_name': 'prometheus_mysqld_exporter', 'group': 'prometheus-mysqld-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-mysqld-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-mysqld-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2024.2', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'prometheus-mysqld-exporter', 'value': {'container_name': 'prometheus_mysqld_exporter', 'group': 'prometheus-mysqld-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-mysqld-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-mysqld-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2024.2', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'prometheus-memcached-exporter', 'value': {'container_name': 'prometheus_memcached_exporter', 'group': 'prometheus-memcached-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-memcached-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-memcached-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2024.2', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'prometheus-memcached-exporter', 'value': {'container_name': 'prometheus_memcached_exporter', 'group': 'prometheus-memcached-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-memcached-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-memcached-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2024.2', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'prometheus-alertmanager', 'value': {'container_name': 'prometheus_alertmanager', 'group': 'prometheus-alertmanager', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-alertmanager:2024.2', 'volumes': ['/etc/kolla/prometheus-alertmanager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'prometheus:/var/lib/prometheus'], 'dimensions': {}, 'haproxy': {'prometheus_alertmanager': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True}, 'prometheus_alertmanager_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9093', 'listen_port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2024.2', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'prometheus-memcached-exporter', 'value': {'container_name': 'prometheus_memcached_exporter', 'group': 'prometheus-memcached-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-memcached-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-memcached-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2024.2', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2024.2', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2024.2', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'prometheus-libvirt-exporter', 'value': {'container_name': 'prometheus_libvirt_exporter', 'group': 'prometheus-libvirt-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-libvirt-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-libvirt-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run/libvirt:/run/libvirt:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'prometheus-blackbox-exporter', 'value': {'container_name': 'prometheus_blackbox_exporter', 'group': 'prometheus-blackbox-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-blackbox-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-blackbox-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2024.2', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2024.2', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'prometheus-elasticsearch-exporter', 'value': {'container_name': 'prometheus_elasticsearch_exporter', 'group': 'prometheus-elasticsearch-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-elasticsearch-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-elasticsearch-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'prometheus-elasticsearch-exporter', 'value': {'container_name': 'prometheus_elasticsearch_exporter', 'group': 'prometheus-elasticsearch-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-elasticsearch-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-elasticsearch-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'prometheus-libvirt-exporter', 'value': {'container_name': 'prometheus_libvirt_exporter', 'group': 'prometheus-libvirt-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-libvirt-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-libvirt-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run/libvirt:/run/libvirt:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'prometheus-libvirt-exporter', 'value': {'container_name': 'prometheus_libvirt_exporter', 'group': 'prometheus-libvirt-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-libvirt-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-libvirt-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run/libvirt:/run/libvirt:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'prometheus-elasticsearch-exporter', 'value': {'container_name': 'prometheus_elasticsearch_exporter', 'group': 'prometheus-elasticsearch-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-elasticsearch-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-elasticsearch-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\nTASK [prometheus : include_tasks] **********************************************\nTuesday 24 February 2026  01:06:11 +0000 (0:00:03.323)       0:00:07.187 ****** \n\u001b[0;36mincluded: /ansible/roles/prometheus/tasks/copy-certs.yml for testbed-manager, testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m\n\nTASK [service-cert-copy : prometheus | Copying over extra CA certificates] *****\nTuesday 24 February 2026  01:06:13 +0000 (0:00:01.666)       0:00:08.854 ****** \n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2024.2', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'prometheus-server', 'value': {'container_name': 'prometheus_server', 'group': 'prometheus', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-v2-server:2024.2', 'volumes': ['/etc/kolla/prometheus-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'prometheus_v2:/var/lib/prometheus', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'prometheus_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9091', 'active_passive': True}, 'prometheus_server_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9091', 'listen_port': '9091', 'active_passive': True}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2024.2', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2024.2', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2024.2', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2024.2', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2024.2', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'prometheus-mysqld-exporter', 'value': {'container_name': 'prometheus_mysqld_exporter', 'group': 'prometheus-mysqld-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-mysqld-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-mysqld-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2024.2', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2024.2', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'prometheus-mysqld-exporter', 'value': {'container_name': 'prometheus_mysqld_exporter', 'group': 'prometheus-mysqld-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-mysqld-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-mysqld-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'prometheus-mysqld-exporter', 'value': {'container_name': 'prometheus_mysqld_exporter', 'group': 'prometheus-mysqld-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-mysqld-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-mysqld-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2024.2', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2024.2', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'prometheus-memcached-exporter', 'value': {'container_name': 'prometheus_memcached_exporter', 'group': 'prometheus-memcached-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-memcached-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-memcached-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'prometheus-libvirt-exporter', 'value': {'container_name': 'prometheus_libvirt_exporter', 'group': 'prometheus-libvirt-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-libvirt-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-libvirt-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run/libvirt:/run/libvirt:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'prometheus-memcached-exporter', 'value': {'container_name': 'prometheus_memcached_exporter', 'group': 'prometheus-memcached-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-memcached-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-memcached-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2024.2', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'prometheus-memcached-exporter', 'value': {'container_name': 'prometheus_memcached_exporter', 'group': 'prometheus-memcached-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-memcached-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-memcached-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'prometheus-libvirt-exporter', 'value': {'container_name': 'prometheus_libvirt_exporter', 'group': 'prometheus-libvirt-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-libvirt-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-libvirt-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run/libvirt:/run/libvirt:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'prometheus-alertmanager', 'value': {'container_name': 'prometheus_alertmanager', 'group': 'prometheus-alertmanager', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-alertmanager:2024.2', 'volumes': ['/etc/kolla/prometheus-alertmanager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'prometheus:/var/lib/prometheus'], 'dimensions': {}, 'haproxy': {'prometheus_alertmanager': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True}, 'prometheus_alertmanager_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9093', 'listen_port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2024.2', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'prometheus-libvirt-exporter', 'value': {'container_name': 'prometheus_libvirt_exporter', 'group': 'prometheus-libvirt-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-libvirt-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-libvirt-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run/libvirt:/run/libvirt:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2024.2', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2024.2', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'prometheus-blackbox-exporter', 'value': {'container_name': 'prometheus_blackbox_exporter', 'group': 'prometheus-blackbox-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-blackbox-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-blackbox-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'prometheus-elasticsearch-exporter', 'value': {'container_name': 'prometheus_elasticsearch_exporter', 'group': 'prometheus-elasticsearch-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-elasticsearch-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-elasticsearch-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'prometheus-elasticsearch-exporter', 'value': {'container_name': 'prometheus_elasticsearch_exporter', 'group': 'prometheus-elasticsearch-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-elasticsearch-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-elasticsearch-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'prometheus-elasticsearch-exporter', 'value': {'container_name': 'prometheus_elasticsearch_exporter', 'group': 'prometheus-elasticsearch-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-elasticsearch-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-elasticsearch-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\nTASK [service-cert-copy : prometheus | Copying over backend internal TLS certificate] ***\nTuesday 24 February 2026  01:06:20 +0000 (0:00:07.061)       0:00:15.915 ****** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2024.2', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-mysqld-exporter', 'value': {'container_name': 'prometheus_mysqld_exporter', 'group': 'prometheus-mysqld-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-mysqld-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-mysqld-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-memcached-exporter', 'value': {'container_name': 'prometheus_memcached_exporter', 'group': 'prometheus-memcached-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-memcached-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-memcached-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2024.2', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-elasticsearch-exporter', 'value': {'container_name': 'prometheus_elasticsearch_exporter', 'group': 'prometheus-elasticsearch-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-elasticsearch-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-elasticsearch-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-manager] => (item={'key': 'prometheus-server', 'value': {'container_name': 'prometheus_server', 'group': 'prometheus', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-v2-server:2024.2', 'volumes': ['/etc/kolla/prometheus-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'prometheus_v2:/var/lib/prometheus', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'prometheus_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9091', 'active_passive': True}, 'prometheus_server_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9091', 'listen_port': '9091', 'active_passive': True}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2024.2', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2024.2', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager] => (item={'key': 'prometheus-alertmanager', 'value': {'container_name': 'prometheus_alertmanager', 'group': 'prometheus-alertmanager', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-alertmanager:2024.2', 'volumes': ['/etc/kolla/prometheus-alertmanager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'prometheus:/var/lib/prometheus'], 'dimensions': {}, 'haproxy': {'prometheus_alertmanager': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True}, 'prometheus_alertmanager_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9093', 'listen_port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager] => (item={'key': 'prometheus-blackbox-exporter', 'value': {'container_name': 'prometheus_blackbox_exporter', 'group': 'prometheus-blackbox-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-blackbox-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-blackbox-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2024.2', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-mysqld-exporter', 'value': {'container_name': 'prometheus_mysqld_exporter', 'group': 'prometheus-mysqld-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-mysqld-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-mysqld-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-memcached-exporter', 'value': {'container_name': 'prometheus_memcached_exporter', 'group': 'prometheus-memcached-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-memcached-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-memcached-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2024.2', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-elasticsearch-exporter', 'value': {'container_name': 'prometheus_elasticsearch_exporter', 'group': 'prometheus-elasticsearch-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-elasticsearch-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-elasticsearch-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2024.2', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-mysqld-exporter', 'value': {'container_name': 'prometheus_mysqld_exporter', 'group': 'prometheus-mysqld-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-mysqld-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-mysqld-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-memcached-exporter', 'value': {'container_name': 'prometheus_memcached_exporter', 'group': 'prometheus-memcached-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-memcached-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-memcached-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2024.2', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-elasticsearch-exporter', 'value': {'container_name': 'prometheus_elasticsearch_exporter', 'group': 'prometheus-elasticsearch-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-elasticsearch-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-elasticsearch-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2024.2', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2024.2', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'prometheus-libvirt-exporter', 'value': {'container_name': 'prometheus_libvirt_exporter', 'group': 'prometheus-libvirt-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-libvirt-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-libvirt-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run/libvirt:/run/libvirt:ro'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2024.2', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2024.2', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'prometheus-libvirt-exporter', 'value': {'container_name': 'prometheus_libvirt_exporter', 'group': 'prometheus-libvirt-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-libvirt-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-libvirt-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run/libvirt:/run/libvirt:ro'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2024.2', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2024.2', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'prometheus-libvirt-exporter', 'value': {'container_name': 'prometheus_libvirt_exporter', 'group': 'prometheus-libvirt-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-libvirt-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-libvirt-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run/libvirt:/run/libvirt:ro'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [service-cert-copy : prometheus | Copying over backend internal TLS key] ***\nTuesday 24 February 2026  01:06:24 +0000 (0:00:03.620)       0:00:19.536 ****** \n\u001b[0;36mskipping: [testbed-manager] => (item={'key': 'prometheus-server', 'value': {'container_name': 'prometheus_server', 'group': 'prometheus', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-v2-server:2024.2', 'volumes': ['/etc/kolla/prometheus-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'prometheus_v2:/var/lib/prometheus', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'prometheus_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9091', 'active_passive': True}, 'prometheus_server_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9091', 'listen_port': '9091', 'active_passive': True}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2024.2', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2024.2', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager] => (item={'key': 'prometheus-alertmanager', 'value': {'container_name': 'prometheus_alertmanager', 'group': 'prometheus-alertmanager', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-alertmanager:2024.2', 'volumes': ['/etc/kolla/prometheus-alertmanager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'prometheus:/var/lib/prometheus'], 'dimensions': {}, 'haproxy': {'prometheus_alertmanager': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True}, 'prometheus_alertmanager_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9093', 'listen_port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager] => (item={'key': 'prometheus-blackbox-exporter', 'value': {'container_name': 'prometheus_blackbox_exporter', 'group': 'prometheus-blackbox-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-blackbox-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-blackbox-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2024.2', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-mysqld-exporter', 'value': {'container_name': 'prometheus_mysqld_exporter', 'group': 'prometheus-mysqld-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-mysqld-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-mysqld-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-memcached-exporter', 'value': {'container_name': 'prometheus_memcached_exporter', 'group': 'prometheus-memcached-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-memcached-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-memcached-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2024.2', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-elasticsearch-exporter', 'value': {'container_name': 'prometheus_elasticsearch_exporter', 'group': 'prometheus-elasticsearch-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-elasticsearch-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-elasticsearch-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2024.2', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-mysqld-exporter', 'value': {'container_name': 'prometheus_mysqld_exporter', 'group': 'prometheus-mysqld-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-mysqld-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-mysqld-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-memcached-exporter', 'value': {'container_name': 'prometheus_memcached_exporter', 'group': 'prometheus-memcached-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-memcached-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-memcached-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2024.2', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-elasticsearch-exporter', 'value': {'container_name': 'prometheus_elasticsearch_exporter', 'group': 'prometheus-elasticsearch-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-elasticsearch-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-elasticsearch-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2024.2', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-mysqld-exporter', 'value': {'container_name': 'prometheus_mysqld_exporter', 'group': 'prometheus-mysqld-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-mysqld-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-mysqld-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-memcached-exporter', 'value': {'container_name': 'prometheus_memcached_exporter', 'group': 'prometheus-memcached-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-memcached-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-memcached-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2024.2', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-elasticsearch-exporter', 'value': {'container_name': 'prometheus_elasticsearch_exporter', 'group': 'prometheus-elasticsearch-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-elasticsearch-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-elasticsearch-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2024.2', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2024.2', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'prometheus-libvirt-exporter', 'value': {'container_name': 'prometheus_libvirt_exporter', 'group': 'prometheus-libvirt-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-libvirt-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-libvirt-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run/libvirt:/run/libvirt:ro'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2024.2', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2024.2', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'prometheus-libvirt-exporter', 'value': {'container_name': 'prometheus_libvirt_exporter', 'group': 'prometheus-libvirt-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-libvirt-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-libvirt-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run/libvirt:/run/libvirt:ro'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2024.2', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2024.2', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'prometheus-libvirt-exporter', 'value': {'container_name': 'prometheus_libvirt_exporter', 'group': 'prometheus-libvirt-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-libvirt-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-libvirt-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run/libvirt:/run/libvirt:ro'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [prometheus : Copying over config.json files] *****************************\nTuesday 24 February 2026  01:06:28 +0000 (0:00:04.616)       0:00:24.153 ****** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2024.2', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2024.2', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2024.2', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'prometheus-server', 'value': {'container_name': 'prometheus_server', 'group': 'prometheus', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-v2-server:2024.2', 'volumes': ['/etc/kolla/prometheus-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'prometheus_v2:/var/lib/prometheus', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'prometheus_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9091', 'active_passive': True}, 'prometheus_server_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9091', 'listen_port': '9091', 'active_passive': True}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'prometheus-mysqld-exporter', 'value': {'container_name': 'prometheus_mysqld_exporter', 'group': 'prometheus-mysqld-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-mysqld-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-mysqld-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2024.2', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2024.2', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'prometheus-mysqld-exporter', 'value': {'container_name': 'prometheus_mysqld_exporter', 'group': 'prometheus-mysqld-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-mysqld-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-mysqld-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2024.2', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'prometheus-memcached-exporter', 'value': {'container_name': 'prometheus_memcached_exporter', 'group': 'prometheus-memcached-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-memcached-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-memcached-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2024.2', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2024.2', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2024.2', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'prometheus-memcached-exporter', 'value': {'container_name': 'prometheus_memcached_exporter', 'group': 'prometheus-memcached-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-memcached-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-memcached-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'prometheus-mysqld-exporter', 'value': {'container_name': 'prometheus_mysqld_exporter', 'group': 'prometheus-mysqld-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-mysqld-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-mysqld-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'prometheus-libvirt-exporter', 'value': {'container_name': 'prometheus_libvirt_exporter', 'group': 'prometheus-libvirt-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-libvirt-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-libvirt-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run/libvirt:/run/libvirt:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2024.2', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2024.2', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2024.2', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2024.2', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'prometheus-alertmanager', 'value': {'container_name': 'prometheus_alertmanager', 'group': 'prometheus-alertmanager', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-alertmanager:2024.2', 'volumes': ['/etc/kolla/prometheus-alertmanager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'prometheus:/var/lib/prometheus'], 'dimensions': {}, 'haproxy': {'prometheus_alertmanager': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True}, 'prometheus_alertmanager_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9093', 'listen_port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'prometheus-memcached-exporter', 'value': {'container_name': 'prometheus_memcached_exporter', 'group': 'prometheus-memcached-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-memcached-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-memcached-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'prometheus-libvirt-exporter', 'value': {'container_name': 'prometheus_libvirt_exporter', 'group': 'prometheus-libvirt-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-libvirt-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-libvirt-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run/libvirt:/run/libvirt:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'prometheus-libvirt-exporter', 'value': {'container_name': 'prometheus_libvirt_exporter', 'group': 'prometheus-libvirt-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-libvirt-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-libvirt-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run/libvirt:/run/libvirt:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'prometheus-elasticsearch-exporter', 'value': {'container_name': 'prometheus_elasticsearch_exporter', 'group': 'prometheus-elasticsearch-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-elasticsearch-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-elasticsearch-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'prometheus-elasticsearch-exporter', 'value': {'container_name': 'prometheus_elasticsearch_exporter', 'group': 'prometheus-elasticsearch-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-elasticsearch-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-elasticsearch-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'prometheus-blackbox-exporter', 'value': {'container_name': 'prometheus_blackbox_exporter', 'group': 'prometheus-blackbox-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-blackbox-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-blackbox-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2024.2', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'prometheus-elasticsearch-exporter', 'value': {'container_name': 'prometheus_elasticsearch_exporter', 'group': 'prometheus-elasticsearch-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-elasticsearch-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-elasticsearch-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\nTASK [prometheus : Find custom prometheus alert rules files] *******************\nTuesday 24 February 2026  01:06:38 +0000 (0:00:09.707)       0:00:33.860 ****** \n\u001b[0;32mok: [testbed-manager -> localhost]\u001b[0m\n\nTASK [prometheus : Copying over custom prometheus alert rules files] ***********\nTuesday 24 February 2026  01:06:40 +0000 (0:00:01.708)       0:00:35.568 ****** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'path': '/operations/prometheus/fluentd-aggregator.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 996, 'inode': 1317250, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.3637931, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'path': '/operations/prometheus/fluentd-aggregator.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 996, 'inode': 1317250, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.3637931, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'path': '/operations/prometheus/fluentd-aggregator.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 996, 'inode': 1317250, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.3637931, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'path': '/operations/prometheus/prometheus.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 12944, 'inode': 1317263, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.368886, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'path': '/operations/prometheus/prometheus.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 12944, 'inode': 1317263, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.368886, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'path': '/operations/prometheus/ceph.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 56929, 'inode': 1317246, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.362469, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'path': '/operations/prometheus/prometheus.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 12944, 'inode': 1317263, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.368886, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item={'path': '/operations/prometheus/fluentd-aggregator.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 996, 'inode': 1317250, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.3637931, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'path': '/operations/prometheus/fluentd-aggregator.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 996, 'inode': 1317250, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.3637931, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'path': '/operations/prometheus/openstack.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 12293, 'inode': 1317258, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.3675466, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'path': '/operations/prometheus/ceph.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 56929, 'inode': 1317246, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.362469, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'path': '/operations/prometheus/fluentd-aggregator.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 996, 'inode': 1317250, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.3637931, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'path': '/operations/prometheus/fluentd-aggregator.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 996, 'inode': 1317250, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.3637931, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'path': '/operations/prometheus/ceph.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 56929, 'inode': 1317246, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.362469, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'path': '/operations/prometheus/cadvisor.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3900, 'inode': 1317244, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.3611212, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'path': '/operations/prometheus/haproxy.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 7933, 'inode': 1317251, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.3647292, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'path': '/operations/prometheus/openstack.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 12293, 'inode': 1317258, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.3675466, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'path': '/operations/prometheus/prometheus.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 12944, 'inode': 1317263, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.368886, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'path': '/operations/prometheus/prometheus.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 12944, 'inode': 1317263, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.368886, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'path': '/operations/prometheus/openstack.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 12293, 'inode': 1317258, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.3675466, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'path': '/operations/prometheus/prometheus.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 12944, 'inode': 1317263, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.368886, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'path': '/operations/prometheus/cadvisor.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3900, 'inode': 1317244, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.3611212, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'path': '/operations/prometheus/ceph.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 56929, 'inode': 1317246, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.362469, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'path': '/operations/prometheus/node.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 14018, 'inode': 1317257, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.366479, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'path': '/operations/prometheus/ceph.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 56929, 'inode': 1317246, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.362469, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'path': '/operations/prometheus/cadvisor.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3900, 'inode': 1317244, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.3611212, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'path': '/operations/prometheus/haproxy.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 7933, 'inode': 1317251, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.3647292, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item={'path': '/operations/prometheus/prometheus.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 12944, 'inode': 1317263, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.368886, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'path': '/operations/prometheus/ceph.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 56929, 'inode': 1317246, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.362469, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'path': '/operations/prometheus/hardware.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 5593, 'inode': 1317253, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.36546, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'path': '/operations/prometheus/openstack.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 12293, 'inode': 1317258, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.3675466, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'path': '/operations/prometheus/openstack.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 12293, 'inode': 1317258, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.3675466, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'path': '/operations/prometheus/haproxy.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 7933, 'inode': 1317251, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.3647292, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'path': '/operations/prometheus/cadvisor.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3900, 'inode': 1317244, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.3611212, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'path': '/operations/prometheus/node.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 14018, 'inode': 1317257, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.366479, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'path': '/operations/prometheus/elasticsearch.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 5987, 'inode': 1317249, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.3633637, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'path': '/operations/prometheus/cadvisor.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3900, 'inode': 1317244, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.3611212, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'path': '/operations/prometheus/prometheus.rec.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3, 'inode': 1317262, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.368623, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'path': '/operations/prometheus/openstack.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 12293, 'inode': 1317258, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.3675466, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'path': '/operations/prometheus/haproxy.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 7933, 'inode': 1317251, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.3647292, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'path': '/operations/prometheus/cadvisor.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3900, 'inode': 1317244, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.3611212, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item={'path': '/operations/prometheus/ceph.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 56929, 'inode': 1317246, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.362469, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'path': '/operations/prometheus/alertmanager.rec.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3, 'inode': 1317241, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.3604016, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'path': '/operations/prometheus/haproxy.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 7933, 'inode': 1317251, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.3647292, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'path': '/operations/prometheus/redfish.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 334, 'inode': 1317272, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.371548, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'path': '/operations/prometheus/haproxy.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 7933, 'inode': 1317251, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.3647292, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'path': '/operations/prometheus/node.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 14018, 'inode': 1317257, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.366479, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'path': '/operations/prometheus/node.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 14018, 'inode': 1317257, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.366479, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'path': '/operations/prometheus/hardware.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 5593, 'inode': 1317253, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.36546, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'path': '/operations/prometheus/node.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 14018, 'inode': 1317257, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.366479, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'path': '/operations/prometheus/node.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 14018, 'inode': 1317257, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.366479, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'path': '/operations/prometheus/hardware.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 5593, 'inode': 1317253, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.36546, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'path': '/operations/prometheus/prometheus-extra.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 7408, 'inode': 1317260, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.3682377, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'path': '/operations/prometheus/elasticsearch.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 5987, 'inode': 1317249, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.3633637, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'path': '/operations/prometheus/hardware.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 5593, 'inode': 1317253, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.36546, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'path': '/operations/prometheus/hardware.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 5593, 'inode': 1317253, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.36546, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'path': '/operations/prometheus/hardware.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 5593, 'inode': 1317253, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.36546, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'path': '/operations/prometheus/ceph.rec.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3, 'inode': 1317245, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.361316, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'path': '/operations/prometheus/elasticsearch.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 5987, 'inode': 1317249, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.3633637, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'path': '/operations/prometheus/elasticsearch.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 5987, 'inode': 1317249, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.3633637, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'path': '/operations/prometheus/elasticsearch.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 5987, 'inode': 1317249, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.3633637, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'path': '/operations/prometheus/elasticsearch.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 5987, 'inode': 1317249, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.3633637, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'path': '/operations/prometheus/prometheus.rec.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3, 'inode': 1317262, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.368623, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item={'path': '/operations/prometheus/openstack.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 12293, 'inode': 1317258, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.3675466, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'path': '/operations/prometheus/prometheus.rec.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3, 'inode': 1317262, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.368623, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'path': '/operations/prometheus/prometheus.rec.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3, 'inode': 1317262, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.368623, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'path': '/operations/prometheus/alertmanager.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 5065, 'inode': 1317242, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.3607938, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'path': '/operations/prometheus/prometheus.rec.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3, 'inode': 1317262, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.368623, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'path': '/operations/prometheus/alertmanager.rec.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3, 'inode': 1317241, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.3604016, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'path': '/operations/prometheus/prometheus.rec.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3, 'inode': 1317262, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.368623, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'path': '/operations/prometheus/alertmanager.rec.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3, 'inode': 1317241, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.3604016, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'path': '/operations/prometheus/alertmanager.rec.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3, 'inode': 1317241, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.3604016, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'path': '/operations/prometheus/alertmanager.rec.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3, 'inode': 1317241, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.3604016, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'path': '/operations/prometheus/redfish.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 334, 'inode': 1317272, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.371548, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'path': '/operations/prometheus/node.rec.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 2309, 'inode': 1317256, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.3662472, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'path': '/operations/prometheus/alertmanager.rec.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3, 'inode': 1317241, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.3604016, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'path': '/operations/prometheus/redfish.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 334, 'inode': 1317272, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.371548, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'path': '/operations/prometheus/mysql.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3792, 'inode': 1317255, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.3659542, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'path': '/operations/prometheus/redfish.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 334, 'inode': 1317272, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.371548, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item={'path': '/operations/prometheus/cadvisor.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3900, 'inode': 1317244, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.3611212, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'path': '/operations/prometheus/redfish.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 334, 'inode': 1317272, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.371548, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'path': '/operations/prometheus/rabbitmq.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3539, 'inode': 1317271, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.3709881, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'path': '/operations/prometheus/prometheus-extra.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 7408, 'inode': 1317260, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.3682377, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'path': '/operations/prometheus/prometheus-extra.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 7408, 'inode': 1317260, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.3682377, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'path': '/operations/prometheus/redfish.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 334, 'inode': 1317272, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.371548, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'path': '/operations/prometheus/prometheus-extra.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 7408, 'inode': 1317260, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.3682377, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'path': '/operations/prometheus/prometheus-extra.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 7408, 'inode': 1317260, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.3682377, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'path': '/operations/prometheus/ceph.rec.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3, 'inode': 1317245, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.361316, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'path': '/operations/prometheus/ceph.rec.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3, 'inode': 1317245, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.361316, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'path': '/operations/prometheus/ceph.rec.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3, 'inode': 1317245, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.361316, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'path': '/operations/prometheus/alertmanager.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 5065, 'inode': 1317242, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.3607938, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'path': '/operations/prometheus/prometheus-extra.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 7408, 'inode': 1317260, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.3682377, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'path': '/operations/prometheus/ceph.rec.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3, 'inode': 1317245, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.361316, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'path': '/operations/prometheus/alertmanager.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 5065, 'inode': 1317242, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.3607938, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item={'path': '/operations/prometheus/haproxy.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 7933, 'inode': 1317251, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.3647292, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'path': '/operations/prometheus/alertmanager.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 5065, 'inode': 1317242, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.3607938, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'path': '/operations/prometheus/node.rec.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 2309, 'inode': 1317256, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.3662472, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'path': '/operations/prometheus/node.rec.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 2309, 'inode': 1317256, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.3662472, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'path': '/operations/prometheus/ceph.rec.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3, 'inode': 1317245, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.361316, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'path': '/operations/prometheus/node.rec.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 2309, 'inode': 1317256, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.3662472, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'path': '/operations/prometheus/alertmanager.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 5065, 'inode': 1317242, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.3607938, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'path': '/operations/prometheus/mysql.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3792, 'inode': 1317255, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.3659542, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'path': '/operations/prometheus/mysql.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3792, 'inode': 1317255, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.3659542, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'path': '/operations/prometheus/alertmanager.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 5065, 'inode': 1317242, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.3607938, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'path': '/operations/prometheus/mysql.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3792, 'inode': 1317255, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.3659542, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'path': '/operations/prometheus/node.rec.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 2309, 'inode': 1317256, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.3662472, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item={'path': '/operations/prometheus/node.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 14018, 'inode': 1317257, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.366479, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'path': '/operations/prometheus/rabbitmq.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3539, 'inode': 1317271, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.3709881, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'path': '/operations/prometheus/node.rec.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 2309, 'inode': 1317256, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.3662472, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'path': '/operations/prometheus/mysql.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3792, 'inode': 1317255, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.3659542, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'path': '/operations/prometheus/rabbitmq.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3539, 'inode': 1317271, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.3709881, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'path': '/operations/prometheus/rabbitmq.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3539, 'inode': 1317271, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.3709881, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'path': '/operations/prometheus/rabbitmq.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3539, 'inode': 1317271, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.3709881, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'path': '/operations/prometheus/mysql.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3792, 'inode': 1317255, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.3659542, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'path': '/operations/prometheus/rabbitmq.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3539, 'inode': 1317271, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.3709881, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item={'path': '/operations/prometheus/hardware.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 5593, 'inode': 1317253, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.36546, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item={'path': '/operations/prometheus/elasticsearch.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 5987, 'inode': 1317249, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.3633637, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item={'path': '/operations/prometheus/prometheus.rec.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3, 'inode': 1317262, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.368623, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item={'path': '/operations/prometheus/alertmanager.rec.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3, 'inode': 1317241, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.3604016, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item={'path': '/operations/prometheus/redfish.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 334, 'inode': 1317272, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.371548, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item={'path': '/operations/prometheus/prometheus-extra.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 7408, 'inode': 1317260, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.3682377, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item={'path': '/operations/prometheus/ceph.rec.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3, 'inode': 1317245, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.361316, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item={'path': '/operations/prometheus/alertmanager.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 5065, 'inode': 1317242, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.3607938, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item={'path': '/operations/prometheus/node.rec.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 2309, 'inode': 1317256, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.3662472, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item={'path': '/operations/prometheus/mysql.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3792, 'inode': 1317255, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.3659542, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item={'path': '/operations/prometheus/rabbitmq.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3539, 'inode': 1317271, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.3709881, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})\u001b[0m\n\nTASK [prometheus : Find prometheus common config overrides] ********************\nTuesday 24 February 2026  01:07:16 +0000 (0:00:35.749)       0:01:11.318 ****** \n\u001b[0;32mok: [testbed-manager -> localhost]\u001b[0m\n\nTASK [prometheus : Find prometheus host config overrides] **********************\nTuesday 24 February 2026  01:07:16 +0000 (0:00:00.865)       0:01:12.184 ****** \n\u001b[1;35m[WARNING]: Skipped\u001b[0m\n\u001b[1;35m'/opt/configuration/environments/kolla/files/overlays/prometheus/testbed-\u001b[0m\n\u001b[1;35mmanager/prometheus.yml.d' path due to this access issue:\u001b[0m\n\u001b[1;35m'/opt/configuration/environments/kolla/files/overlays/prometheus/testbed-\u001b[0m\n\u001b[1;35mmanager/prometheus.yml.d' is not a directory\u001b[0m\n\u001b[0;32mok: [testbed-manager -> localhost]\u001b[0m\n\u001b[1;35m[WARNING]: Skipped\u001b[0m\n\u001b[1;35m'/opt/configuration/environments/kolla/files/overlays/prometheus/testbed-\u001b[0m\n\u001b[1;35mnode-0/prometheus.yml.d' path due to this access issue:\u001b[0m\n\u001b[1;35m'/opt/configuration/environments/kolla/files/overlays/prometheus/testbed-\u001b[0m\n\u001b[1;35mnode-0/prometheus.yml.d' is not a directory\u001b[0m\n\u001b[0;32mok: [testbed-node-0 -> localhost]\u001b[0m\n\u001b[1;35m[WARNING]: Skipped\u001b[0m\n\u001b[1;35m'/opt/configuration/environments/kolla/files/overlays/prometheus/testbed-\u001b[0m\n\u001b[1;35mnode-1/prometheus.yml.d' path due to this access issue:\u001b[0m\n\u001b[1;35m'/opt/configuration/environments/kolla/files/overlays/prometheus/testbed-\u001b[0m\n\u001b[1;35mnode-1/prometheus.yml.d' is not a directory\u001b[0m\n\u001b[0;32mok: [testbed-node-1 -> localhost]\u001b[0m\n\u001b[1;35m[WARNING]: Skipped\u001b[0m\n\u001b[1;35m'/opt/configuration/environments/kolla/files/overlays/prometheus/testbed-\u001b[0m\n\u001b[1;35mnode-2/prometheus.yml.d' path due to this access issue:\u001b[0m\n\u001b[1;35m'/opt/configuration/environments/kolla/files/overlays/prometheus/testbed-\u001b[0m\n\u001b[1;35mnode-2/prometheus.yml.d' is not a directory\u001b[0m\n\u001b[0;32mok: [testbed-node-2 -> localhost]\u001b[0m\n\u001b[1;35m[WARNING]: Skipped\u001b[0m\n\u001b[1;35m'/opt/configuration/environments/kolla/files/overlays/prometheus/testbed-\u001b[0m\n\u001b[1;35mnode-3/prometheus.yml.d' path due to this access issue:\u001b[0m\n\u001b[1;35m'/opt/configuration/environments/kolla/files/overlays/prometheus/testbed-\u001b[0m\n\u001b[1;35mnode-3/prometheus.yml.d' is not a directory\u001b[0m\n\u001b[0;32mok: [testbed-node-3 -> localhost]\u001b[0m\n\u001b[1;35m[WARNING]: Skipped\u001b[0m\n\u001b[1;35m'/opt/configuration/environments/kolla/files/overlays/prometheus/testbed-\u001b[0m\n\u001b[1;35mnode-5/prometheus.yml.d' path due to this access issue:\u001b[0m\n\u001b[1;35m'/opt/configuration/environments/kolla/files/overlays/prometheus/testbed-\u001b[0m\n\u001b[1;35mnode-5/prometheus.yml.d' is not a directory\u001b[0m\n\u001b[0;32mok: [testbed-node-5 -> localhost]\u001b[0m\n\u001b[1;35m[WARNING]: Skipped\u001b[0m\n\u001b[1;35m'/opt/configuration/environments/kolla/files/overlays/prometheus/testbed-\u001b[0m\n\u001b[1;35mnode-4/prometheus.yml.d' path due to this access issue:\u001b[0m\n\u001b[1;35m'/opt/configuration/environments/kolla/files/overlays/prometheus/testbed-\u001b[0m\n\u001b[1;35mnode-4/prometheus.yml.d' is not a directory\u001b[0m\n\u001b[0;32mok: [testbed-node-4 -> localhost]\u001b[0m\n\nTASK [prometheus : Copying over prometheus config file] ************************\nTuesday 24 February 2026  01:07:18 +0000 (0:00:01.910)       0:01:14.094 ****** \n\u001b[0;36mskipping: [testbed-node-0] => (item=/ansible/roles/prometheus/templates/prometheus.yml.j2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=/ansible/roles/prometheus/templates/prometheus.yml.j2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=/ansible/roles/prometheus/templates/prometheus.yml.j2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=/ansible/roles/prometheus/templates/prometheus.yml.j2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=/ansible/roles/prometheus/templates/prometheus.yml.j2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=/ansible/roles/prometheus/templates/prometheus.yml.j2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item=/ansible/roles/prometheus/templates/prometheus.yml.j2)\u001b[0m\n\nTASK [prometheus : Copying over prometheus web config file] ********************\nTuesday 24 February 2026  01:07:37 +0000 (0:00:18.746)       0:01:32.840 ****** \n\u001b[0;36mskipping: [testbed-node-1] => (item=/ansible/roles/prometheus/templates/prometheus-web.yml.j2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=/ansible/roles/prometheus/templates/prometheus-web.yml.j2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=/ansible/roles/prometheus/templates/prometheus-web.yml.j2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=/ansible/roles/prometheus/templates/prometheus-web.yml.j2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=/ansible/roles/prometheus/templates/prometheus-web.yml.j2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=/ansible/roles/prometheus/templates/prometheus-web.yml.j2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item=/ansible/roles/prometheus/templates/prometheus-web.yml.j2)\u001b[0m\n\nTASK [prometheus : Copying over prometheus alertmanager config file] ***********\nTuesday 24 February 2026  01:07:42 +0000 (0:00:04.630)       0:01:37.471 ****** \n\u001b[0;36mskipping: [testbed-node-0] => (item=/opt/configuration/environments/kolla/files/overlays/prometheus/prometheus-alertmanager.yml) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=/opt/configuration/environments/kolla/files/overlays/prometheus/prometheus-alertmanager.yml) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=/opt/configuration/environments/kolla/files/overlays/prometheus/prometheus-alertmanager.yml) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=/opt/configuration/environments/kolla/files/overlays/prometheus/prometheus-alertmanager.yml) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=/opt/configuration/environments/kolla/files/overlays/prometheus/prometheus-alertmanager.yml) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item=/opt/configuration/environments/kolla/files/overlays/prometheus/prometheus-alertmanager.yml)\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=/opt/configuration/environments/kolla/files/overlays/prometheus/prometheus-alertmanager.yml) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\nTASK [prometheus : Find custom Alertmanager alert notification templates] ******\nTuesday 24 February 2026  01:07:45 +0000 (0:00:03.081)       0:01:40.553 ****** \n\u001b[0;32mok: [testbed-manager -> localhost]\u001b[0m\n\nTASK [prometheus : Copying over custom Alertmanager alert notification templates] ***\nTuesday 24 February 2026  01:07:46 +0000 (0:00:01.367)       0:01:41.920 ****** \n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [prometheus : Copying over my.cnf for mysqld_exporter] ********************\nTuesday 24 February 2026  01:07:47 +0000 (0:00:01.191)       0:01:43.111 ****** \n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [prometheus : Copying cloud config file for openstack exporter] ***********\nTuesday 24 February 2026  01:07:51 +0000 (0:00:03.254)       0:01:46.366 ****** \n\u001b[0;36mskipping: [testbed-node-1] => (item=/ansible/roles/prometheus/templates/clouds.yml.j2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=/ansible/roles/prometheus/templates/clouds.yml.j2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-manager] => (item=/ansible/roles/prometheus/templates/clouds.yml.j2) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=/ansible/roles/prometheus/templates/clouds.yml.j2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=/ansible/roles/prometheus/templates/clouds.yml.j2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=/ansible/roles/prometheus/templates/clouds.yml.j2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=/ansible/roles/prometheus/templates/clouds.yml.j2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [prometheus : Copying config file for blackbox exporter] ******************\nTuesday 24 February 2026  01:07:53 +0000 (0:00:02.360)       0:01:48.726 ****** \n\u001b[0;36mskipping: [testbed-node-0] => (item=/ansible/roles/prometheus/templates/prometheus-blackbox-exporter.yml.j2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=/ansible/roles/prometheus/templates/prometheus-blackbox-exporter.yml.j2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=/ansible/roles/prometheus/templates/prometheus-blackbox-exporter.yml.j2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item=/ansible/roles/prometheus/templates/prometheus-blackbox-exporter.yml.j2)\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=/ansible/roles/prometheus/templates/prometheus-blackbox-exporter.yml.j2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=/ansible/roles/prometheus/templates/prometheus-blackbox-exporter.yml.j2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=/ansible/roles/prometheus/templates/prometheus-blackbox-exporter.yml.j2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [prometheus : Find extra prometheus server config files] ******************\nTuesday 24 February 2026  01:07:55 +0000 (0:00:02.105)       0:01:50.832 ****** \n\u001b[1;35m[WARNING]: Skipped\u001b[0m\n\u001b[1;35m'/opt/configuration/environments/kolla/files/overlays/prometheus/extras/' path\u001b[0m\n\u001b[1;35mdue to this access issue:\u001b[0m\n\u001b[1;35m'/opt/configuration/environments/kolla/files/overlays/prometheus/extras/' is\u001b[0m\n\u001b[1;35mnot a directory\u001b[0m\n\u001b[0;32mok: [testbed-manager -> localhost]\u001b[0m\n\nTASK [prometheus : Create subdirectories for extra config files] ***************\nTuesday 24 February 2026  01:07:57 +0000 (0:00:01.599)       0:01:52.432 ****** \n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [prometheus : Template extra prometheus server config files] **************\nTuesday 24 February 2026  01:07:59 +0000 (0:00:02.145)       0:01:54.577 ****** \n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [prometheus : Check prometheus containers] ********************************\nTuesday 24 February 2026  01:08:00 +0000 (0:00:00.925)       0:01:55.502 ****** \n\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'prometheus-server', 'value': {'container_name': 'prometheus_server', 'group': 'prometheus', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-v2-server:2024.2', 'volumes': ['/etc/kolla/prometheus-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'prometheus_v2:/var/lib/prometheus', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'prometheus_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9091', 'active_passive': True}, 'prometheus_server_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9091', 'listen_port': '9091', 'active_passive': True}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2024.2', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2024.2', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2024.2', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2024.2', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2024.2', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2024.2', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2024.2', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2024.2', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'prometheus-mysqld-exporter', 'value': {'container_name': 'prometheus_mysqld_exporter', 'group': 'prometheus-mysqld-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-mysqld-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-mysqld-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'prometheus-mysqld-exporter', 'value': {'container_name': 'prometheus_mysqld_exporter', 'group': 'prometheus-mysqld-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-mysqld-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-mysqld-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'prometheus-mysqld-exporter', 'value': {'container_name': 'prometheus_mysqld_exporter', 'group': 'prometheus-mysqld-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-mysqld-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-mysqld-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2024.2', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2024.2', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2024.2', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'prometheus-libvirt-exporter', 'value': {'container_name': 'prometheus_libvirt_exporter', 'group': 'prometheus-libvirt-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-libvirt-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-libvirt-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run/libvirt:/run/libvirt:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'prometheus-alertmanager', 'value': {'container_name': 'prometheus_alertmanager', 'group': 'prometheus-alertmanager', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-alertmanager:2024.2', 'volumes': ['/etc/kolla/prometheus-alertmanager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'prometheus:/var/lib/prometheus'], 'dimensions': {}, 'haproxy': {'prometheus_alertmanager': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True}, 'prometheus_alertmanager_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9093', 'listen_port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'prometheus-memcached-exporter', 'value': {'container_name': 'prometheus_memcached_exporter', 'group': 'prometheus-memcached-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-memcached-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-memcached-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'prometheus-memcached-exporter', 'value': {'container_name': 'prometheus_memcached_exporter', 'group': 'prometheus-memcached-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-memcached-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-memcached-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'prometheus-libvirt-exporter', 'value': {'container_name': 'prometheus_libvirt_exporter', 'group': 'prometheus-libvirt-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-libvirt-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-libvirt-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run/libvirt:/run/libvirt:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'prometheus-memcached-exporter', 'value': {'container_name': 'prometheus_memcached_exporter', 'group': 'prometheus-memcached-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-memcached-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-memcached-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'prometheus-libvirt-exporter', 'value': {'container_name': 'prometheus_libvirt_exporter', 'group': 'prometheus-libvirt-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-libvirt-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-libvirt-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run/libvirt:/run/libvirt:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'prometheus-blackbox-exporter', 'value': {'container_name': 'prometheus_blackbox_exporter', 'group': 'prometheus-blackbox-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-blackbox-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-blackbox-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2024.2', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2024.2', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2024.2', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'prometheus-elasticsearch-exporter', 'value': {'container_name': 'prometheus_elasticsearch_exporter', 'group': 'prometheus-elasticsearch-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-elasticsearch-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-elasticsearch-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'prometheus-elasticsearch-exporter', 'value': {'container_name': 'prometheus_elasticsearch_exporter', 'group': 'prometheus-elasticsearch-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-elasticsearch-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-elasticsearch-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'prometheus-elasticsearch-exporter', 'value': {'container_name': 'prometheus_elasticsearch_exporter', 'group': 'prometheus-elasticsearch-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-elasticsearch-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-elasticsearch-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m\n\nTASK [prometheus : Creating prometheus database user and setting permissions] ***\nTuesday 24 February 2026  01:08:04 +0000 (0:00:04.587)       0:02:00.090 ****** \n\u001b[0;36mskipping: [testbed-manager] => (item=testbed-node-0) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\nTASK [prometheus : Flush handlers] *********************************************\nTuesday 24 February 2026  01:08:06 +0000 (0:00:01.662)       0:02:01.753 ****** \n\nTASK [prometheus : Flush handlers] *********************************************\nTuesday 24 February 2026  01:08:06 +0000 (0:00:00.087)       0:02:01.840 ****** \n\nTASK [prometheus : Flush handlers] *********************************************\nTuesday 24 February 2026  01:08:06 +0000 (0:00:00.069)       0:02:01.910 ****** \n\nTASK [prometheus : Flush handlers] *********************************************\nTuesday 24 February 2026  01:08:06 +0000 (0:00:00.071)       0:02:01.982 ****** \n\nTASK [prometheus : Flush handlers] *********************************************\nTuesday 24 February 2026  01:08:07 +0000 (0:00:00.405)       0:02:02.387 ****** \n\nTASK [prometheus : Flush handlers] *********************************************\nTuesday 24 February 2026  01:08:07 +0000 (0:00:00.079)       0:02:02.467 ****** \n\nTASK [prometheus : Flush handlers] *********************************************\nTuesday 24 February 2026  01:08:07 +0000 (0:00:00.072)       0:02:02.539 ****** \n\nRUNNING HANDLER [prometheus : Restart prometheus-server container] *************\nTuesday 24 February 2026  01:08:07 +0000 (0:00:00.111)       0:02:02.651 ****** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nRUNNING HANDLER [prometheus : Restart prometheus-node-exporter container] ******\nTuesday 24 February 2026  01:08:23 +0000 (0:00:15.973)       0:02:18.625 ****** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\nRUNNING HANDLER [prometheus : Restart prometheus-mysqld-exporter container] ****\nTuesday 24 February 2026  01:08:38 +0000 (0:00:15.586)       0:02:34.211 ****** \n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [prometheus : Restart prometheus-memcached-exporter container] ***\nTuesday 24 February 2026  01:08:50 +0000 (0:00:11.283)       0:02:45.494 ****** \n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [prometheus : Restart prometheus-cadvisor container] ***********\nTuesday 24 February 2026  01:08:57 +0000 (0:00:07.008)       0:02:52.503 ****** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\nRUNNING HANDLER [prometheus : Restart prometheus-alertmanager container] *******\nTuesday 24 February 2026  01:09:13 +0000 (0:00:16.517)       0:03:09.021 ****** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nRUNNING HANDLER [prometheus : Restart prometheus-elasticsearch-exporter container] ***\nTuesday 24 February 2026  01:09:20 +0000 (0:00:06.976)       0:03:15.997 ****** \n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nRUNNING HANDLER [prometheus : Restart prometheus-blackbox-exporter container] ***\nTuesday 24 February 2026  01:09:30 +0000 (0:00:09.898)       0:03:25.895 ****** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nRUNNING HANDLER [prometheus : Restart prometheus-libvirt-exporter container] ***\nTuesday 24 February 2026  01:09:35 +0000 (0:00:05.162)       0:03:31.058 ****** \n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;33mtestbed-manager\u001b[0m            : \u001b[0;32mok=23  \u001b[0m \u001b[0;33mchanged=14  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=8   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=15  \u001b[0m \u001b[0;33mchanged=10  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=11  \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=15  \u001b[0m \u001b[0;33mchanged=10  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=11  \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=15  \u001b[0m \u001b[0;33mchanged=10  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=11  \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-3\u001b[0m             : \u001b[0;32mok=12  \u001b[0m \u001b[0;33mchanged=7   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=12  \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-4\u001b[0m             : \u001b[0;32mok=12  \u001b[0m \u001b[0;33mchanged=7   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=12  \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-5\u001b[0m             : \u001b[0;32mok=12  \u001b[0m \u001b[0;33mchanged=7   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=12  \u001b[0m rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nTuesday 24 February 2026  01:09:46 +0000 (0:00:10.244)       0:03:41.303 ****** \n=============================================================================== \nprometheus : Copying over custom prometheus alert rules files ---------- 35.75s\nprometheus : Copying over prometheus config file ----------------------- 18.75s\nprometheus : Restart prometheus-cadvisor container --------------------- 16.52s\nprometheus : Restart prometheus-server container ----------------------- 15.97s\nprometheus : Restart prometheus-node-exporter container ---------------- 15.59s\nprometheus : Restart prometheus-mysqld-exporter container -------------- 11.28s\nprometheus : Restart prometheus-libvirt-exporter container ------------- 10.24s\nprometheus : Restart prometheus-elasticsearch-exporter container -------- 9.90s\nprometheus : Copying over config.json files ----------------------------- 9.71s\nservice-cert-copy : prometheus | Copying over extra CA certificates ----- 7.06s\nprometheus : Restart prometheus-memcached-exporter container ------------ 7.01s\nprometheus : Restart prometheus-alertmanager container ------------------ 6.98s\nprometheus : Restart prometheus-blackbox-exporter container ------------- 5.16s\nprometheus : Copying over prometheus web config file -------------------- 4.63s\nservice-cert-copy : prometheus | Copying over backend internal TLS key --- 4.62s\nprometheus : Check prometheus containers -------------------------------- 4.59s\nservice-cert-copy : prometheus | Copying over backend internal TLS certificate --- 3.62s\nprometheus : Ensuring config directories exist -------------------------- 3.32s\nprometheus : Copying over my.cnf for mysqld_exporter -------------------- 3.25s\nprometheus : Copying over prometheus alertmanager config file ----------- 3.08s\n\u001b[32m2026-02-24 01:09:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4baf3a0c-387b-4215-b490-5edd972623e9 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:09:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 01:09:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fba6bf07-65a1-432e-9a3d-72209557a21d is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:09:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask eddb1a5f-448f-4517-b064-c30da006d691 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:09:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask de6aa5ad-1dfd-4a58-bf00-4755d79ef102 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:09:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4baf3a0c-387b-4215-b490-5edd972623e9 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:09:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 01:09:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fba6bf07-65a1-432e-9a3d-72209557a21d is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:09:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask eddb1a5f-448f-4517-b064-c30da006d691 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:09:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask de6aa5ad-1dfd-4a58-bf00-4755d79ef102 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:09:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4baf3a0c-387b-4215-b490-5edd972623e9 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:09:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 01:09:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fba6bf07-65a1-432e-9a3d-72209557a21d is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:09:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask eddb1a5f-448f-4517-b064-c30da006d691 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:09:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask de6aa5ad-1dfd-4a58-bf00-4755d79ef102 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:09:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4baf3a0c-387b-4215-b490-5edd972623e9 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:09:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 01:10:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fba6bf07-65a1-432e-9a3d-72209557a21d is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:10:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask eddb1a5f-448f-4517-b064-c30da006d691 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:10:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask de6aa5ad-1dfd-4a58-bf00-4755d79ef102 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:10:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4baf3a0c-387b-4215-b490-5edd972623e9 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:10:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 01:10:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fba6bf07-65a1-432e-9a3d-72209557a21d is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:10:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask eddb1a5f-448f-4517-b064-c30da006d691 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:10:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask de6aa5ad-1dfd-4a58-bf00-4755d79ef102 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:10:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4baf3a0c-387b-4215-b490-5edd972623e9 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:10:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 01:10:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fba6bf07-65a1-432e-9a3d-72209557a21d is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:10:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask eddb1a5f-448f-4517-b064-c30da006d691 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:10:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask de6aa5ad-1dfd-4a58-bf00-4755d79ef102 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:10:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4baf3a0c-387b-4215-b490-5edd972623e9 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:10:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 01:10:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fba6bf07-65a1-432e-9a3d-72209557a21d is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:10:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask eddb1a5f-448f-4517-b064-c30da006d691 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:10:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask de6aa5ad-1dfd-4a58-bf00-4755d79ef102 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:10:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4baf3a0c-387b-4215-b490-5edd972623e9 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:10:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 01:10:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fba6bf07-65a1-432e-9a3d-72209557a21d is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:10:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask eddb1a5f-448f-4517-b064-c30da006d691 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:10:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask de6aa5ad-1dfd-4a58-bf00-4755d79ef102 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:10:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4baf3a0c-387b-4215-b490-5edd972623e9 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:10:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 01:10:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fba6bf07-65a1-432e-9a3d-72209557a21d is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:10:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask eddb1a5f-448f-4517-b064-c30da006d691 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:10:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask de6aa5ad-1dfd-4a58-bf00-4755d79ef102 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:10:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4baf3a0c-387b-4215-b490-5edd972623e9 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:10:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 01:10:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fba6bf07-65a1-432e-9a3d-72209557a21d is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:10:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask eddb1a5f-448f-4517-b064-c30da006d691 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:10:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask de6aa5ad-1dfd-4a58-bf00-4755d79ef102 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:10:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4baf3a0c-387b-4215-b490-5edd972623e9 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:10:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 01:10:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fba6bf07-65a1-432e-9a3d-72209557a21d is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:10:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask eddb1a5f-448f-4517-b064-c30da006d691 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:10:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask de6aa5ad-1dfd-4a58-bf00-4755d79ef102 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:10:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4baf3a0c-387b-4215-b490-5edd972623e9 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:10:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 01:10:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fba6bf07-65a1-432e-9a3d-72209557a21d is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:10:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask eddb1a5f-448f-4517-b064-c30da006d691 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:10:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask de6aa5ad-1dfd-4a58-bf00-4755d79ef102 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:10:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4baf3a0c-387b-4215-b490-5edd972623e9 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:10:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 01:10:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fba6bf07-65a1-432e-9a3d-72209557a21d is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:10:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask eddb1a5f-448f-4517-b064-c30da006d691 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:10:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask de6aa5ad-1dfd-4a58-bf00-4755d79ef102 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:10:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4baf3a0c-387b-4215-b490-5edd972623e9 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:10:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 01:10:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fba6bf07-65a1-432e-9a3d-72209557a21d is in state SUCCESS\u001b[0m\n\n\nPLAY [Group hosts based on configuration] **************************************\n\nTASK [Group hosts based on Kolla action] ***************************************\nTuesday 24 February 2026  01:07:05 +0000 (0:00:00.529)       0:00:00.529 ****** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [Group hosts based on enabled services] ***********************************\nTuesday 24 February 2026  01:07:06 +0000 (0:00:00.606)       0:00:01.135 ****** \n\u001b[0;32mok: [testbed-node-0] => (item=enable_glance_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item=enable_glance_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item=enable_glance_True)\u001b[0m\n\nPLAY [Apply role glance] *******************************************************\n\nTASK [glance : include_tasks] **************************************************\nTuesday 24 February 2026  01:07:07 +0000 (0:00:01.432)       0:00:02.568 ****** \n\u001b[0;36mincluded: /ansible/roles/glance/tasks/deploy.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [service-ks-register : glance | Creating services] ************************\nTuesday 24 February 2026  01:07:09 +0000 (0:00:01.471)       0:00:04.040 ****** \n\u001b[0;33mchanged: [testbed-node-0] => (item=glance (image))\u001b[0m\n\nTASK [service-ks-register : glance | Creating endpoints] ***********************\nTuesday 24 February 2026  01:07:13 +0000 (0:00:04.321)       0:00:08.362 ****** \n\u001b[0;33mchanged: [testbed-node-0] => (item=glance -> https://api-int.testbed.osism.xyz:9292 -> internal)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=glance -> https://api.testbed.osism.xyz:9292 -> public)\u001b[0m\n\nTASK [service-ks-register : glance | Creating projects] ************************\nTuesday 24 February 2026  01:07:20 +0000 (0:00:06.822)       0:00:15.184 ****** \n\u001b[0;32mok: [testbed-node-0] => (item=service)\u001b[0m\n\nTASK [service-ks-register : glance | Creating users] ***************************\nTuesday 24 February 2026  01:07:23 +0000 (0:00:03.335)       0:00:18.520 ****** \n\u001b[0;33mchanged: [testbed-node-0] => (item=glance -> service)\u001b[0m\n\u001b[1;35m[WARNING]: Module did not set no_log for update_password\u001b[0m\n\nTASK [service-ks-register : glance | Creating roles] ***************************\nTuesday 24 February 2026  01:07:27 +0000 (0:00:04.402)       0:00:22.922 ****** \n\u001b[0;32mok: [testbed-node-0] => (item=admin)\u001b[0m\n\nTASK [service-ks-register : glance | Granting user roles] **********************\nTuesday 24 February 2026  01:07:31 +0000 (0:00:03.688)       0:00:26.611 ****** \n\u001b[0;33mchanged: [testbed-node-0] => (item=glance -> service -> admin)\u001b[0m\n\nTASK [glance : Ensuring config directories exist] ******************************\nTuesday 24 February 2026  01:07:36 +0000 (0:00:04.453)       0:00:31.064 ****** \n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'glance-api', 'value': {'container_name': 'glance_api', 'group': 'glance-api', 'host_in_groups': True, 'enabled': True, 'image': 'registry.osism.tech/kolla/glance-api:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.11,192.168.16.9'}, 'privileged': True, 'volumes': ['/etc/kolla/glance-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'glance:/var/lib/glance/', '', 'kolla_logs:/var/log/kolla/', '', '', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9292'], 'timeout': '30'}, 'haproxy': {'glance_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}, 'glance_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'glance-api', 'value': {'container_name': 'glance_api', 'group': 'glance-api', 'host_in_groups': True, 'enabled': True, 'image': 'registry.osism.tech/kolla/glance-api:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.10,192.168.16.9'}, 'privileged': True, 'volumes': ['/etc/kolla/glance-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'glance:/var/lib/glance/', '', 'kolla_logs:/var/log/kolla/', '', '', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9292'], 'timeout': '30'}, 'haproxy': {'glance_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}, 'glance_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'glance-api', 'value': {'container_name': 'glance_api', 'group': 'glance-api', 'host_in_groups': True, 'enabled': True, 'image': 'registry.osism.tech/kolla/glance-api:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.12,192.168.16.9'}, 'privileged': True, 'volumes': ['/etc/kolla/glance-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'glance:/var/lib/glance/', '', 'kolla_logs:/var/log/kolla/', '', '', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9292'], 'timeout': '30'}, 'haproxy': {'glance_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}, 'glance_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}}})\u001b[0m\n\nTASK [glance : include_tasks] **************************************************\nTuesday 24 February 2026  01:07:42 +0000 (0:00:05.994)       0:00:37.059 ****** \n\u001b[0;36mincluded: /ansible/roles/glance/tasks/external_ceph.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [glance : Ensuring glance service ceph config subdir exists] **************\nTuesday 24 February 2026  01:07:43 +0000 (0:00:01.396)       0:00:38.456 ****** \n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\nTASK [glance : Copy over multiple ceph configs for Glance] *********************\nTuesday 24 February 2026  01:07:49 +0000 (0:00:05.557)       0:00:44.014 ****** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'name': 'rbd', 'type': 'rbd', 'cluster': 'ceph', 'enabled': True})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'name': 'rbd', 'type': 'rbd', 'cluster': 'ceph', 'enabled': True})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'name': 'rbd', 'type': 'rbd', 'cluster': 'ceph', 'enabled': True})\u001b[0m\n\nTASK [glance : Copy over ceph Glance keyrings] *********************************\nTuesday 24 February 2026  01:07:51 +0000 (0:00:02.304)       0:00:46.319 ****** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'name': 'rbd', 'type': 'rbd', 'cluster': 'ceph', 'enabled': True})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'name': 'rbd', 'type': 'rbd', 'cluster': 'ceph', 'enabled': True})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'name': 'rbd', 'type': 'rbd', 'cluster': 'ceph', 'enabled': True})\u001b[0m\n\nTASK [glance : Ensuring config directory has correct owner and permission] *****\nTuesday 24 February 2026  01:07:53 +0000 (0:00:01.724)       0:00:48.043 ****** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [glance : Check if policies shall be overwritten] *************************\nTuesday 24 February 2026  01:07:54 +0000 (0:00:01.343)       0:00:49.387 ****** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [glance : Set glance policy file] *****************************************\nTuesday 24 February 2026  01:07:54 +0000 (0:00:00.366)       0:00:49.754 ****** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [glance : include_tasks] **************************************************\nTuesday 24 February 2026  01:07:55 +0000 (0:00:00.616)       0:00:50.371 ****** \n\u001b[0;36mincluded: /ansible/roles/glance/tasks/copy-certs.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [service-cert-copy : glance | Copying over extra CA certificates] *********\nTuesday 24 February 2026  01:07:55 +0000 (0:00:00.614)       0:00:50.985 ****** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'glance-api', 'value': {'container_name': 'glance_api', 'group': 'glance-api', 'host_in_groups': True, 'enabled': True, 'image': 'registry.osism.tech/kolla/glance-api:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.10,192.168.16.9'}, 'privileged': True, 'volumes': ['/etc/kolla/glance-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'glance:/var/lib/glance/', '', 'kolla_logs:/var/log/kolla/', '', '', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9292'], 'timeout': '30'}, 'haproxy': {'glance_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}, 'glance_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'glance-api', 'value': {'container_name': 'glance_api', 'group': 'glance-api', 'host_in_groups': True, 'enabled': True, 'image': 'registry.osism.tech/kolla/glance-api:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.11,192.168.16.9'}, 'privileged': True, 'volumes': ['/etc/kolla/glance-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'glance:/var/lib/glance/', '', 'kolla_logs:/var/log/kolla/', '', '', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9292'], 'timeout': '30'}, 'haproxy': {'glance_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}, 'glance_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'glance-api', 'value': {'container_name': 'glance_api', 'group': 'glance-api', 'host_in_groups': True, 'enabled': True, 'image': 'registry.osism.tech/kolla/glance-api:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.12,192.168.16.9'}, 'privileged': True, 'volumes': ['/etc/kolla/glance-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'glance:/var/lib/glance/', '', 'kolla_logs:/var/log/kolla/', '', '', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9292'], 'timeout': '30'}, 'haproxy': {'glance_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}, 'glance_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}}})\u001b[0m\n\nTASK [service-cert-copy : glance | Copying over backend internal TLS certificate] ***\nTuesday 24 February 2026  01:08:01 +0000 (0:00:05.846)       0:00:56.831 ****** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'glance-api', 'value': {'container_name': 'glance_api', 'group': 'glance-api', 'host_in_groups': True, 'enabled': True, 'image': 'registry.osism.tech/kolla/glance-api:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.10,192.168.16.9'}, 'privileged': True, 'volumes': ['/etc/kolla/glance-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'glance:/var/lib/glance/', '', 'kolla_logs:/var/log/kolla/', '', '', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9292'], 'timeout': '30'}, 'haproxy': {'glance_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}, 'glance_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'glance-api', 'value': {'container_name': 'glance_api', 'group': 'glance-api', 'host_in_groups': True, 'enabled': True, 'image': 'registry.osism.tech/kolla/glance-api:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.11,192.168.16.9'}, 'privileged': True, 'volumes': ['/etc/kolla/glance-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'glance:/var/lib/glance/', '', 'kolla_logs:/var/log/kolla/', '', '', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9292'], 'timeout': '30'}, 'haproxy': {'glance_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}, 'glance_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'glance-api', 'value': {'container_name': 'glance_api', 'group': 'glance-api', 'host_in_groups': True, 'enabled': True, 'image': 'registry.osism.tech/kolla/glance-api:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.12,192.168.16.9'}, 'privileged': True, 'volumes': ['/etc/kolla/glance-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'glance:/var/lib/glance/', '', 'kolla_logs:/var/log/kolla/', '', '', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9292'], 'timeout': '30'}, 'haproxy': {'glance_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}, 'glance_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [service-cert-copy : glance | Copying over backend internal TLS key] ******\nTuesday 24 February 2026  01:08:04 +0000 (0:00:03.120)       0:00:59.952 ****** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'glance-api', 'value': {'container_name': 'glance_api', 'group': 'glance-api', 'host_in_groups': True, 'enabled': True, 'image': 'registry.osism.tech/kolla/glance-api:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.10,192.168.16.9'}, 'privileged': True, 'volumes': ['/etc/kolla/glance-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'glance:/var/lib/glance/', '', 'kolla_logs:/var/log/kolla/', '', '', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9292'], 'timeout': '30'}, 'haproxy': {'glance_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}, 'glance_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'glance-api', 'value': {'container_name': 'glance_api', 'group': 'glance-api', 'host_in_groups': True, 'enabled': True, 'image': 'registry.osism.tech/kolla/glance-api:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.11,192.168.16.9'}, 'privileged': True, 'volumes': ['/etc/kolla/glance-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'glance:/var/lib/glance/', '', 'kolla_logs:/var/log/kolla/', '', '', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9292'], 'timeout': '30'}, 'haproxy': {'glance_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}, 'glance_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'glance-api', 'value': {'container_name': 'glance_api', 'group': 'glance-api', 'host_in_groups': True, 'enabled': True, 'image': 'registry.osism.tech/kolla/glance-api:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.12,192.168.16.9'}, 'privileged': True, 'volumes': ['/etc/kolla/glance-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'glance:/var/lib/glance/', '', 'kolla_logs:/var/log/kolla/', '', '', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9292'], 'timeout': '30'}, 'haproxy': {'glance_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}, 'glance_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [glance : Creating TLS backend PEM File] **********************************\nTuesday 24 February 2026  01:08:08 +0000 (0:00:03.603)       0:01:03.556 ****** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [glance : Copying over config.json files for services] ********************\nTuesday 24 February 2026  01:08:13 +0000 (0:00:05.412)       0:01:08.968 ****** \n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'glance-api', 'value': {'container_name': 'glance_api', 'group': 'glance-api', 'host_in_groups': True, 'enabled': True, 'image': 'registry.osism.tech/kolla/glance-api:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.11,192.168.16.9'}, 'privileged': True, 'volumes': ['/etc/kolla/glance-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'glance:/var/lib/glance/', '', 'kolla_logs:/var/log/kolla/', '', '', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9292'], 'timeout': '30'}, 'haproxy': {'glance_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}, 'glance_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'glance-api', 'value': {'container_name': 'glance_api', 'group': 'glance-api', 'host_in_groups': True, 'enabled': True, 'image': 'registry.osism.tech/kolla/glance-api:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.10,192.168.16.9'}, 'privileged': True, 'volumes': ['/etc/kolla/glance-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'glance:/var/lib/glance/', '', 'kolla_logs:/var/log/kolla/', '', '', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9292'], 'timeout': '30'}, 'haproxy': {'glance_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}, 'glance_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'glance-api', 'value': {'container_name': 'glance_api', 'group': 'glance-api', 'host_in_groups': True, 'enabled': True, 'image': 'registry.osism.tech/kolla/glance-api:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.12,192.168.16.9'}, 'privileged': True, 'volumes': ['/etc/kolla/glance-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'glance:/var/lib/glance/', '', 'kolla_logs:/var/log/kolla/', '', '', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9292'], 'timeout': '30'}, 'haproxy': {'glance_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}, 'glance_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}}})\u001b[0m\n\nTASK [glance : Copying over glance-api.conf] ***********************************\nTuesday 24 February 2026  01:08:19 +0000 (0:00:05.947)       0:01:14.916 ****** \n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [glance : Copying over glance-cache.conf for glance_api] ******************\nTuesday 24 February 2026  01:08:31 +0000 (0:00:11.398)       0:01:26.314 ****** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [glance : Copying over glance-swift.conf for glance_api] ******************\nTuesday 24 February 2026  01:08:35 +0000 (0:00:04.409)       0:01:30.724 ****** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [glance : Copying over glance-image-import.conf] **************************\nTuesday 24 February 2026  01:08:40 +0000 (0:00:04.911)       0:01:35.635 ****** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [glance : Copying over property-protections-rules.conf] *******************\nTuesday 24 February 2026  01:08:45 +0000 (0:00:04.503)       0:01:40.139 ****** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [glance : Copying over existing policy file] ******************************\nTuesday 24 February 2026  01:08:49 +0000 (0:00:04.070)       0:01:44.209 ****** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [glance : Copying over glance-haproxy-tls.cfg] ****************************\nTuesday 24 February 2026  01:08:49 +0000 (0:00:00.425)       0:01:44.634 ****** \n\u001b[0;36mskipping: [testbed-node-1] => (item=/ansible/roles/glance/templates/glance-tls-proxy.cfg.j2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=/ansible/roles/glance/templates/glance-tls-proxy.cfg.j2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=/ansible/roles/glance/templates/glance-tls-proxy.cfg.j2) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [glance : Generating 'hostnqn' file for glance_api] ***********************\nTuesday 24 February 2026  01:08:55 +0000 (0:00:05.794)       0:01:50.429 ****** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\nTASK [glance : Check glance containers] ****************************************\nTuesday 24 February 2026  01:09:02 +0000 (0:00:07.563)       0:01:57.992 ****** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'glance-api', 'value': {'container_name': 'glance_api', 'group': 'glance-api', 'host_in_groups': True, 'enabled': True, 'image': 'registry.osism.tech/kolla/glance-api:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.10,192.168.16.9'}, 'privileged': True, 'volumes': ['/etc/kolla/glance-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'glance:/var/lib/glance/', '', 'kolla_logs:/var/log/kolla/', '', '', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9292'], 'timeout': '30'}, 'haproxy': {'glance_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}, 'glance_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'glance-api', 'value': {'container_name': 'glance_api', 'group': 'glance-api', 'host_in_groups': True, 'enabled': True, 'image': 'registry.osism.tech/kolla/glance-api:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.11,192.168.16.9'}, 'privileged': True, 'volumes': ['/etc/kolla/glance-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'glance:/var/lib/glance/', '', 'kolla_logs:/var/log/kolla/', '', '', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9292'], 'timeout': '30'}, 'haproxy': {'glance_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}, 'glance_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'glance-api', 'value': {'container_name': 'glance_api', 'group': 'glance-api', 'host_in_groups': True, 'enabled': True, 'image': 'registry.osism.tech/kolla/glance-api:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.12,192.168.16.9'}, 'privileged': True, 'volumes': ['/etc/kolla/glance-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'glance:/var/lib/glance/', '', 'kolla_logs:/var/log/kolla/', '', '', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9292'], 'timeout': '30'}, 'haproxy': {'glance_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}, 'glance_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}}})\u001b[0m\n\nTASK [glance : include_tasks] **************************************************\nTuesday 24 February 2026  01:09:09 +0000 (0:00:06.758)       0:02:04.751 ****** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [glance : Creating Glance database] ***************************************\nTuesday 24 February 2026  01:09:10 +0000 (0:00:00.327)       0:02:05.079 ****** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [glance : Creating Glance database user and setting permissions] **********\nTuesday 24 February 2026  01:09:12 +0000 (0:00:02.019)       0:02:07.099 ****** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [glance : Enable log_bin_trust_function_creators function] ****************\nTuesday 24 February 2026  01:09:14 +0000 (0:00:02.366)       0:02:09.465 ****** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [glance : Running Glance bootstrap container] *****************************\nTuesday 24 February 2026  01:09:16 +0000 (0:00:02.433)       0:02:11.899 ****** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [glance : Disable log_bin_trust_function_creators function] ***************\nTuesday 24 February 2026  01:09:51 +0000 (0:00:34.287)       0:02:46.187 ****** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [glance : Flush handlers] *************************************************\nTuesday 24 February 2026  01:09:53 +0000 (0:00:02.220)       0:02:48.407 ****** \n\nTASK [glance : Flush handlers] *************************************************\nTuesday 24 February 2026  01:09:54 +0000 (0:00:00.689)       0:02:49.096 ****** \n\nTASK [glance : Flush handlers] *************************************************\nTuesday 24 February 2026  01:09:54 +0000 (0:00:00.074)       0:02:49.171 ****** \n\nRUNNING HANDLER [glance : Restart glance-api container] ************************\nTuesday 24 February 2026  01:09:54 +0000 (0:00:00.086)       0:02:49.257 ****** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=27  \u001b[0m \u001b[0;33mchanged=19  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=12  \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=16  \u001b[0m \u001b[0;33mchanged=10  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=11  \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=16  \u001b[0m \u001b[0;33mchanged=10  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=11  \u001b[0m rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nTuesday 24 February 2026  01:10:28 +0000 (0:00:33.819)       0:03:23.077 ****** \n=============================================================================== \nglance : Running Glance bootstrap container ---------------------------- 34.29s\nglance : Restart glance-api container ---------------------------------- 33.82s\nglance : Copying over glance-api.conf ---------------------------------- 11.40s\nglance : Generating 'hostnqn' file for glance_api ----------------------- 7.56s\nservice-ks-register : glance | Creating endpoints ----------------------- 6.82s\nglance : Check glance containers ---------------------------------------- 6.76s\nglance : Ensuring config directories exist ------------------------------ 5.99s\nglance : Copying over config.json files for services -------------------- 5.95s\nservice-cert-copy : glance | Copying over extra CA certificates --------- 5.85s\nglance : Copying over glance-haproxy-tls.cfg ---------------------------- 5.79s\nglance : Ensuring glance service ceph config subdir exists -------------- 5.56s\nglance : Creating TLS backend PEM File ---------------------------------- 5.41s\nglance : Copying over glance-swift.conf for glance_api ------------------ 4.91s\nglance : Copying over glance-image-import.conf -------------------------- 4.50s\nservice-ks-register : glance | Granting user roles ---------------------- 4.45s\nglance : Copying over glance-cache.conf for glance_api ------------------ 4.41s\nservice-ks-register : glance | Creating users --------------------------- 4.40s\nservice-ks-register : glance | Creating services ------------------------ 4.32s\nglance : Copying over property-protections-rules.conf ------------------- 4.07s\nservice-ks-register : glance | Creating roles --------------------------- 3.69s\n\u001b[32m2026-02-24 01:10:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask eddb1a5f-448f-4517-b064-c30da006d691 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:10:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask de6aa5ad-1dfd-4a58-bf00-4755d79ef102 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:10:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 918da279-e500-4c4b-bdd7-1bd26f3de1b1 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:10:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4baf3a0c-387b-4215-b490-5edd972623e9 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:10:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 01:10:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask eddb1a5f-448f-4517-b064-c30da006d691 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:10:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask de6aa5ad-1dfd-4a58-bf00-4755d79ef102 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:10:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 918da279-e500-4c4b-bdd7-1bd26f3de1b1 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:10:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4baf3a0c-387b-4215-b490-5edd972623e9 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:10:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 01:10:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask eddb1a5f-448f-4517-b064-c30da006d691 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:10:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask de6aa5ad-1dfd-4a58-bf00-4755d79ef102 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:10:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 918da279-e500-4c4b-bdd7-1bd26f3de1b1 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:10:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4baf3a0c-387b-4215-b490-5edd972623e9 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:10:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 01:10:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask eddb1a5f-448f-4517-b064-c30da006d691 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:10:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask de6aa5ad-1dfd-4a58-bf00-4755d79ef102 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:10:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 918da279-e500-4c4b-bdd7-1bd26f3de1b1 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:10:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4baf3a0c-387b-4215-b490-5edd972623e9 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:10:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 01:10:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask eddb1a5f-448f-4517-b064-c30da006d691 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:10:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask de6aa5ad-1dfd-4a58-bf00-4755d79ef102 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:10:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 918da279-e500-4c4b-bdd7-1bd26f3de1b1 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:10:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4baf3a0c-387b-4215-b490-5edd972623e9 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:10:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 01:10:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask eddb1a5f-448f-4517-b064-c30da006d691 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:10:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask de6aa5ad-1dfd-4a58-bf00-4755d79ef102 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:10:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 918da279-e500-4c4b-bdd7-1bd26f3de1b1 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:10:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4baf3a0c-387b-4215-b490-5edd972623e9 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:10:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 01:10:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask eddb1a5f-448f-4517-b064-c30da006d691 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:10:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask de6aa5ad-1dfd-4a58-bf00-4755d79ef102 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:10:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 918da279-e500-4c4b-bdd7-1bd26f3de1b1 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:10:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4baf3a0c-387b-4215-b490-5edd972623e9 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:10:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 01:10:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask eddb1a5f-448f-4517-b064-c30da006d691 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:10:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask de6aa5ad-1dfd-4a58-bf00-4755d79ef102 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:10:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 918da279-e500-4c4b-bdd7-1bd26f3de1b1 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:10:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4baf3a0c-387b-4215-b490-5edd972623e9 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:10:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 01:10:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask eddb1a5f-448f-4517-b064-c30da006d691 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:10:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask de6aa5ad-1dfd-4a58-bf00-4755d79ef102 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:10:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 918da279-e500-4c4b-bdd7-1bd26f3de1b1 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:10:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4baf3a0c-387b-4215-b490-5edd972623e9 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:10:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 01:10:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask eddb1a5f-448f-4517-b064-c30da006d691 is in state SUCCESS\u001b[0m\n\n\nPLAY [Group hosts based on configuration] **************************************\n\nTASK [Group hosts based on Kolla action] ***************************************\nTuesday 24 February 2026  01:07:44 +0000 (0:00:00.350)       0:00:00.350 ****** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [Group hosts based on enabled services] ***********************************\nTuesday 24 February 2026  01:07:44 +0000 (0:00:00.452)       0:00:00.802 ****** \n\u001b[0;32mok: [testbed-node-0] => (item=enable_cinder_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item=enable_cinder_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item=enable_cinder_True)\u001b[0m\n\nPLAY [Apply role cinder] *******************************************************\n\nTASK [cinder : include_tasks] **************************************************\nTuesday 24 February 2026  01:07:46 +0000 (0:00:01.080)       0:00:01.882 ****** \n\u001b[0;36mincluded: /ansible/roles/cinder/tasks/deploy.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [service-ks-register : cinder | Creating services] ************************\nTuesday 24 February 2026  01:07:47 +0000 (0:00:01.080)       0:00:02.964 ****** \n\u001b[0;33mchanged: [testbed-node-0] => (item=cinderv3 (volumev3))\u001b[0m\n\nTASK [service-ks-register : cinder | Creating endpoints] ***********************\nTuesday 24 February 2026  01:07:51 +0000 (0:00:03.950)       0:00:06.914 ****** \n\u001b[0;33mchanged: [testbed-node-0] => (item=cinderv3 -> https://api-int.testbed.osism.xyz:8776/v3/%(tenant_id)s -> internal)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=cinderv3 -> https://api.testbed.osism.xyz:8776/v3/%(tenant_id)s -> public)\u001b[0m\n\nTASK [service-ks-register : cinder | Creating projects] ************************\nTuesday 24 February 2026  01:07:58 +0000 (0:00:07.411)       0:00:14.325 ****** \n\u001b[0;32mok: [testbed-node-0] => (item=service)\u001b[0m\n\nTASK [service-ks-register : cinder | Creating users] ***************************\nTuesday 24 February 2026  01:08:02 +0000 (0:00:03.961)       0:00:18.287 ****** \n\u001b[0;33mchanged: [testbed-node-0] => (item=cinder -> service)\u001b[0m\n\u001b[1;35m[WARNING]: Module did not set no_log for update_password\u001b[0m\n\nTASK [service-ks-register : cinder | Creating roles] ***************************\nTuesday 24 February 2026  01:08:06 +0000 (0:00:04.491)       0:00:22.778 ****** \n\u001b[0;32mok: [testbed-node-0] => (item=admin)\u001b[0m\n\nTASK [service-ks-register : cinder | Granting user roles] **********************\nTuesday 24 February 2026  01:08:11 +0000 (0:00:04.385)       0:00:27.164 ****** \n\u001b[0;33mchanged: [testbed-node-0] => (item=cinder -> service -> admin)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=cinder -> service -> service)\u001b[0m\n\nTASK [cinder : Ensuring config directories exist] ******************************\nTuesday 24 February 2026  01:08:19 +0000 (0:00:08.019)       0:00:35.183 ****** \n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-api:2024.2', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8776'], 'timeout': '30'}, 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-api:2024.2', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8776'], 'timeout': '30'}, 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-api:2024.2', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8776'], 'timeout': '30'}, 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-scheduler:2024.2', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-scheduler:2024.2', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-scheduler:2024.2', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-volume:2024.2', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-volume:2024.2', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-volume:2024.2', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-backup:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-backup:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-backup:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}})\u001b[0m\n\nTASK [cinder : include_tasks] **************************************************\nTuesday 24 February 2026  01:08:22 +0000 (0:00:02.803)       0:00:37.986 ****** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [cinder : include_tasks] **************************************************\nTuesday 24 February 2026  01:08:22 +0000 (0:00:00.580)       0:00:38.567 ****** \n\u001b[0;36mincluded: /ansible/roles/cinder/tasks/external_ceph.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [cinder : Ensuring cinder service ceph config subdirs exists] *************\nTuesday 24 February 2026  01:08:23 +0000 (0:00:00.923)       0:00:39.491 ****** \n\u001b[0;33mchanged: [testbed-node-0] => (item=cinder-volume)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=cinder-volume)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=cinder-backup)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=cinder-volume)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=cinder-backup)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=cinder-backup)\u001b[0m\n\nTASK [cinder : Copying over multiple ceph.conf for cinder services] ************\nTuesday 24 February 2026  01:08:28 +0000 (0:00:04.425)       0:00:43.916 ****** \n\u001b[0;36mskipping: [testbed-node-1] => (item=[{'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-api:2024.2', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8776'], 'timeout': '30'}, 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}}}, {'name': 'rbd-1', 'cluster': 'ceph', 'enabled': True}]) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=[{'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-scheduler:2024.2', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}}, {'name': 'rbd-1', 'cluster': 'ceph', 'enabled': True}]) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=[{'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-api:2024.2', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8776'], 'timeout': '30'}, 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}}}, {'name': 'rbd-1', 'cluster': 'ceph', 'enabled': True}]) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=[{'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-scheduler:2024.2', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}}, {'name': 'rbd-1', 'cluster': 'ceph', 'enabled': True}]) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=[{'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-api:2024.2', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8776'], 'timeout': '30'}, 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}}}, {'name': 'rbd-1', 'cluster': 'ceph', 'enabled': True}]) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=[{'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-scheduler:2024.2', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}}, {'name': 'rbd-1', 'cluster': 'ceph', 'enabled': True}]) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=[{'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-volume:2024.2', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}}, {'name': 'rbd-1', 'cluster': 'ceph', 'enabled': True}])\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=[{'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-volume:2024.2', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}}, {'name': 'rbd-1', 'cluster': 'ceph', 'enabled': True}])\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=[{'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-volume:2024.2', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}}, {'name': 'rbd-1', 'cluster': 'ceph', 'enabled': True}])\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=[{'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-backup:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}}, {'name': 'rbd-1', 'cluster': 'ceph', 'enabled': True}])\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=[{'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-backup:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}}, {'name': 'rbd-1', 'cluster': 'ceph', 'enabled': True}])\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=[{'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-backup:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}}, {'name': 'rbd-1', 'cluster': 'ceph', 'enabled': True}])\u001b[0m\n\nTASK [cinder : Copy over Ceph keyring files for cinder-volume] *****************\nTuesday 24 February 2026  01:08:33 +0000 (0:00:05.051)       0:00:48.967 ****** \n\u001b[0;33mchanged: [testbed-node-1] => (item={'name': 'rbd-1', 'cluster': 'ceph', 'enabled': True})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'name': 'rbd-1', 'cluster': 'ceph', 'enabled': True})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'name': 'rbd-1', 'cluster': 'ceph', 'enabled': True})\u001b[0m\n\nTASK [cinder : Copy over Ceph keyring files for cinder-backup] *****************\nTuesday 24 February 2026  01:08:35 +0000 (0:00:02.638)       0:00:51.605 ****** \n\u001b[0;33mchanged: [testbed-node-0] => (item=ceph.client.cinder.keyring)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=ceph.client.cinder.keyring)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=ceph.client.cinder.keyring)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=ceph.client.cinder-backup.keyring)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=ceph.client.cinder-backup.keyring)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=ceph.client.cinder-backup.keyring)\u001b[0m\n\nTASK [cinder : Ensuring config directory has correct owner and permission] *****\nTuesday 24 February 2026  01:08:39 +0000 (0:00:03.572)       0:00:55.178 ****** \n\u001b[0;32mok: [testbed-node-0] => (item=cinder-volume)\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item=cinder-volume)\u001b[0m\n\u001b[0;32mok: [testbed-node-0] => (item=cinder-backup)\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item=cinder-volume)\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item=cinder-backup)\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item=cinder-backup)\u001b[0m\n\nTASK [cinder : Check if policies shall be overwritten] *************************\nTuesday 24 February 2026  01:08:40 +0000 (0:00:01.564)       0:00:56.743 ****** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [cinder : Set cinder policy file] *****************************************\nTuesday 24 February 2026  01:08:41 +0000 (0:00:00.429)       0:00:57.173 ****** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [cinder : include_tasks] **************************************************\nTuesday 24 February 2026  01:08:41 +0000 (0:00:00.375)       0:00:57.548 ****** \n\u001b[0;36mincluded: /ansible/roles/cinder/tasks/copy-certs.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [service-cert-copy : cinder | Copying over extra CA certificates] *********\nTuesday 24 February 2026  01:08:42 +0000 (0:00:01.041)       0:00:58.590 ****** \n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-api:2024.2', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8776'], 'timeout': '30'}, 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-api:2024.2', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8776'], 'timeout': '30'}, 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-api:2024.2', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8776'], 'timeout': '30'}, 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-scheduler:2024.2', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-scheduler:2024.2', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-scheduler:2024.2', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-volume:2024.2', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-volume:2024.2', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-volume:2024.2', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-backup:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-backup:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-backup:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}})\u001b[0m\n\nTASK [service-cert-copy : cinder | Copying over backend internal TLS certificate] ***\nTuesday 24 February 2026  01:08:47 +0000 (0:00:04.906)       0:01:03.497 ****** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-api:2024.2', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8776'], 'timeout': '30'}, 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-api:2024.2', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8776'], 'timeout': '30'}, 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-scheduler:2024.2', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-scheduler:2024.2', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-volume:2024.2', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-volume:2024.2', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-backup:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-backup:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-api:2024.2', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8776'], 'timeout': '30'}, 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-scheduler:2024.2', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-volume:2024.2', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-backup:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [service-cert-copy : cinder | Copying over backend internal TLS key] ******\nTuesday 24 February 2026  01:08:48 +0000 (0:00:01.140)       0:01:04.637 ****** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-api:2024.2', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8776'], 'timeout': '30'}, 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-scheduler:2024.2', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-volume:2024.2', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-backup:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-api:2024.2', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8776'], 'timeout': '30'}, 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-scheduler:2024.2', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-volume:2024.2', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-backup:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-api:2024.2', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8776'], 'timeout': '30'}, 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-scheduler:2024.2', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-volume:2024.2', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-backup:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [cinder : Copying over config.json files for services] ********************\nTuesday 24 February 2026  01:08:50 +0000 (0:00:02.056)       0:01:06.694 ****** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-api:2024.2', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8776'], 'timeout': '30'}, 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-api:2024.2', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8776'], 'timeout': '30'}, 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-api:2024.2', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8776'], 'timeout': '30'}, 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-scheduler:2024.2', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-scheduler:2024.2', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-scheduler:2024.2', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-volume:2024.2', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-volume:2024.2', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-volume:2024.2', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-backup:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-backup:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-backup:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}})\u001b[0m\n\nTASK [cinder : Copying over cinder-wsgi.conf] **********************************\nTuesday 24 February 2026  01:08:56 +0000 (0:00:05.870)       0:01:12.565 ****** \n\u001b[0;33mchanged: [testbed-node-1] => (item=/ansible/roles/cinder/templates/cinder-wsgi.conf.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=/ansible/roles/cinder/templates/cinder-wsgi.conf.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=/ansible/roles/cinder/templates/cinder-wsgi.conf.j2)\u001b[0m\n\nTASK [cinder : Copying over cinder.conf] ***************************************\nTuesday 24 February 2026  01:08:59 +0000 (0:00:02.417)       0:01:14.982 ****** \n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-api:2024.2', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8776'], 'timeout': '30'}, 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-api:2024.2', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8776'], 'timeout': '30'}, 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-api:2024.2', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8776'], 'timeout': '30'}, 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-scheduler:2024.2', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-scheduler:2024.2', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-scheduler:2024.2', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-volume:2024.2', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-volume:2024.2', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-volume:2024.2', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-backup:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-backup:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-backup:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}})\u001b[0m\n\nTASK [cinder : Generating 'hostnqn' file for cinder_volume] ********************\nTuesday 24 February 2026  01:09:16 +0000 (0:00:16.882)       0:01:31.865 ****** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [cinder : Copying over existing policy file] ******************************\nTuesday 24 February 2026  01:09:18 +0000 (0:00:02.007)       0:01:33.872 ****** \n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-api:2024.2', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8776'], 'timeout': '30'}, 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-scheduler:2024.2', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-volume:2024.2', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-backup:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-api:2024.2', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8776'], 'timeout': '30'}, 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-scheduler:2024.2', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-volume:2024.2', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-api:2024.2', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8776'], 'timeout': '30'}, 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-backup:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-scheduler:2024.2', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-volume:2024.2', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-backup:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [cinder : Copying over nfs_shares files for cinder_volume] ****************\nTuesday 24 February 2026  01:09:18 +0000 (0:00:00.725)       0:01:34.597 ****** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [cinder : Check cinder containers] ****************************************\nTuesday 24 February 2026  01:09:19 +0000 (0:00:00.353)       0:01:34.950 ****** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-api:2024.2', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8776'], 'timeout': '30'}, 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-api:2024.2', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8776'], 'timeout': '30'}, 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-api:2024.2', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8776'], 'timeout': '30'}, 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-scheduler:2024.2', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-scheduler:2024.2', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-scheduler:2024.2', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-volume:2024.2', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-volume:2024.2', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-volume:2024.2', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-backup:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-backup:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-backup:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}})\u001b[0m\n\nTASK [cinder : include_tasks] **************************************************\nTuesday 24 February 2026  01:09:23 +0000 (0:00:04.005)       0:01:38.956 ****** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [cinder : Creating Cinder database] ***************************************\nTuesday 24 February 2026  01:09:23 +0000 (0:00:00.454)       0:01:39.411 ****** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [cinder : Creating Cinder database user and setting permissions] **********\nTuesday 24 February 2026  01:09:25 +0000 (0:00:02.302)       0:01:41.714 ****** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [cinder : Running Cinder bootstrap container] *****************************\nTuesday 24 February 2026  01:09:28 +0000 (0:00:02.453)       0:01:44.167 ****** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [cinder : Flush handlers] *************************************************\nTuesday 24 February 2026  01:09:50 +0000 (0:00:22.048)       0:02:06.216 ****** \n\nTASK [cinder : Flush handlers] *************************************************\nTuesday 24 February 2026  01:09:50 +0000 (0:00:00.083)       0:02:06.299 ****** \n\nTASK [cinder : Flush handlers] *************************************************\nTuesday 24 February 2026  01:09:50 +0000 (0:00:00.114)       0:02:06.414 ****** \n\nRUNNING HANDLER [cinder : Restart cinder-api container] ************************\nTuesday 24 February 2026  01:09:50 +0000 (0:00:00.127)       0:02:06.541 ****** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [cinder : Restart cinder-scheduler container] ******************\nTuesday 24 February 2026  01:10:16 +0000 (0:00:26.155)       0:02:32.697 ****** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [cinder : Restart cinder-volume container] *********************\nTuesday 24 February 2026  01:10:27 +0000 (0:00:11.017)       0:02:43.714 ****** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\nRUNNING HANDLER [cinder : Restart cinder-backup container] *********************\nTuesday 24 February 2026  01:10:46 +0000 (0:00:18.186)       0:03:01.901 ****** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\nRUNNING HANDLER [cinder : Wait for cinder services to update service versions] ***\nTuesday 24 February 2026  01:10:57 +0000 (0:00:11.005)       0:03:12.907 ****** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=30  \u001b[0m \u001b[0;33mchanged=22  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=9   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=21  \u001b[0m \u001b[0;33mchanged=15  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=7   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=21  \u001b[0m \u001b[0;33mchanged=15  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=7   \u001b[0m rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nTuesday 24 February 2026  01:10:57 +0000 (0:00:00.259)       0:03:13.166 ****** \n=============================================================================== \ncinder : Restart cinder-api container ---------------------------------- 26.16s\ncinder : Running Cinder bootstrap container ---------------------------- 22.05s\ncinder : Restart cinder-volume container ------------------------------- 18.19s\ncinder : Copying over cinder.conf -------------------------------------- 16.88s\ncinder : Restart cinder-scheduler container ---------------------------- 11.02s\ncinder : Restart cinder-backup container ------------------------------- 11.01s\nservice-ks-register : cinder | Granting user roles ---------------------- 8.02s\nservice-ks-register : cinder | Creating endpoints ----------------------- 7.41s\ncinder : Copying over config.json files for services -------------------- 5.87s\ncinder : Copying over multiple ceph.conf for cinder services ------------ 5.05s\nservice-cert-copy : cinder | Copying over extra CA certificates --------- 4.91s\nservice-ks-register : cinder | Creating users --------------------------- 4.49s\ncinder : Ensuring cinder service ceph config subdirs exists ------------- 4.42s\nservice-ks-register : cinder | Creating roles --------------------------- 4.39s\ncinder : Check cinder containers ---------------------------------------- 4.01s\nservice-ks-register : cinder | Creating projects ------------------------ 3.96s\nservice-ks-register : cinder | Creating services ------------------------ 3.95s\ncinder : Copy over Ceph keyring files for cinder-backup ----------------- 3.57s\ncinder : Ensuring config directories exist ------------------------------ 2.80s\ncinder : Copy over Ceph keyring files for cinder-volume ----------------- 2.64s\n\u001b[32m2026-02-24 01:10:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask de6aa5ad-1dfd-4a58-bf00-4755d79ef102 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:10:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 918da279-e500-4c4b-bdd7-1bd26f3de1b1 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:10:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4baf3a0c-387b-4215-b490-5edd972623e9 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:10:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 01:11:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask de6aa5ad-1dfd-4a58-bf00-4755d79ef102 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:11:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 918da279-e500-4c4b-bdd7-1bd26f3de1b1 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:11:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4baf3a0c-387b-4215-b490-5edd972623e9 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:11:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 01:11:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask de6aa5ad-1dfd-4a58-bf00-4755d79ef102 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:11:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 918da279-e500-4c4b-bdd7-1bd26f3de1b1 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:11:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4baf3a0c-387b-4215-b490-5edd972623e9 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:11:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 01:11:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask de6aa5ad-1dfd-4a58-bf00-4755d79ef102 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:11:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 918da279-e500-4c4b-bdd7-1bd26f3de1b1 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:11:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4baf3a0c-387b-4215-b490-5edd972623e9 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:11:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 01:11:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask de6aa5ad-1dfd-4a58-bf00-4755d79ef102 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:11:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 918da279-e500-4c4b-bdd7-1bd26f3de1b1 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:11:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4baf3a0c-387b-4215-b490-5edd972623e9 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:11:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 01:11:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask de6aa5ad-1dfd-4a58-bf00-4755d79ef102 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:11:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 918da279-e500-4c4b-bdd7-1bd26f3de1b1 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:11:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4baf3a0c-387b-4215-b490-5edd972623e9 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:11:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 01:11:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask de6aa5ad-1dfd-4a58-bf00-4755d79ef102 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:11:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 918da279-e500-4c4b-bdd7-1bd26f3de1b1 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:11:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4baf3a0c-387b-4215-b490-5edd972623e9 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:11:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 01:11:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask de6aa5ad-1dfd-4a58-bf00-4755d79ef102 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:11:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 918da279-e500-4c4b-bdd7-1bd26f3de1b1 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:11:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4baf3a0c-387b-4215-b490-5edd972623e9 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:11:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 01:11:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask de6aa5ad-1dfd-4a58-bf00-4755d79ef102 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:11:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 918da279-e500-4c4b-bdd7-1bd26f3de1b1 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:11:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4baf3a0c-387b-4215-b490-5edd972623e9 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:11:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 01:11:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask de6aa5ad-1dfd-4a58-bf00-4755d79ef102 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:11:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 918da279-e500-4c4b-bdd7-1bd26f3de1b1 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:11:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4baf3a0c-387b-4215-b490-5edd972623e9 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:11:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 01:11:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask de6aa5ad-1dfd-4a58-bf00-4755d79ef102 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:11:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 918da279-e500-4c4b-bdd7-1bd26f3de1b1 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:13:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4baf3a0c-387b-4215-b490-5edd972623e9 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:13:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 01:13:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask de6aa5ad-1dfd-4a58-bf00-4755d79ef102 is in state SUCCESS\u001b[0m\n\u001b[32m2026-02-24 01:13:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 918da279-e500-4c4b-bdd7-1bd26f3de1b1 is in state SUCCESS\u001b[0m\n\n\nPLAY [Group hosts based on configuration] **************************************\n\nTASK [Group hosts based on Kolla action] ***************************************\nTuesday 24 February 2026  01:09:51 +0000 (0:00:00.430)       0:00:00.430 ****** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [Group hosts based on enabled services] ***********************************\nTuesday 24 February 2026  01:09:52 +0000 (0:00:00.701)       0:00:01.132 ****** \n\u001b[0;32mok: [testbed-node-0] => (item=enable_nova_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item=enable_nova_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item=enable_nova_True)\u001b[0m\n\nPLAY [Wait for the Nova service] ***********************************************\n\nTASK [Waiting for Nova public port to be UP] ***********************************\nTuesday 24 February 2026  01:09:53 +0000 (0:00:01.455)       0:00:02.587 ****** \n\nSTILL ALIVE [task 'Waiting for Nova public port to be UP' is running] **********\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;32mtestbed-node-0\u001b[0m             : \u001b[0;32mok=3   \u001b[0m changed=0    unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\u001b[0;32mtestbed-node-1\u001b[0m             : \u001b[0;32mok=3   \u001b[0m changed=0    unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\u001b[0;32mtestbed-node-2\u001b[0m             : \u001b[0;32mok=3   \u001b[0m changed=0    unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nTuesday 24 February 2026  01:12:32 +0000 (0:02:38.788)       0:02:41.376 ****** \n=============================================================================== \nWaiting for Nova public port to be UP --------------------------------- 158.79s\nGroup hosts based on enabled services ----------------------------------- 1.45s\nGroup hosts based on Kolla action --------------------------------------- 0.70s\n\n\nPLAY [Group hosts based on configuration] **************************************\n\nTASK [Group hosts based on Kolla action] ***************************************\nTuesday 24 February 2026  01:10:33 +0000 (0:00:00.292)       0:00:00.292 ****** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [Group hosts based on enabled services] ***********************************\nTuesday 24 February 2026  01:10:33 +0000 (0:00:00.324)       0:00:00.616 ****** \n\u001b[0;32mok: [testbed-node-0] => (item=enable_grafana_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item=enable_grafana_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item=enable_grafana_True)\u001b[0m\n\nPLAY [Apply role grafana] ******************************************************\n\nTASK [grafana : include_tasks] *************************************************\nTuesday 24 February 2026  01:10:34 +0000 (0:00:00.486)       0:00:01.103 ****** \n\u001b[0;36mincluded: /ansible/roles/grafana/tasks/deploy.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [grafana : Ensuring config directories exist] *****************************\nTuesday 24 February 2026  01:10:34 +0000 (0:00:00.612)       0:00:01.715 ****** \n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'grafana', 'value': {'container_name': 'grafana', 'group': 'grafana', 'enabled': True, 'image': 'registry.osism.tech/kolla/grafana:2024.2', 'volumes': ['/etc/kolla/grafana/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'grafana_server': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000'}, 'grafana_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'grafana', 'value': {'container_name': 'grafana', 'group': 'grafana', 'enabled': True, 'image': 'registry.osism.tech/kolla/grafana:2024.2', 'volumes': ['/etc/kolla/grafana/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'grafana_server': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000'}, 'grafana_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'grafana', 'value': {'container_name': 'grafana', 'group': 'grafana', 'enabled': True, 'image': 'registry.osism.tech/kolla/grafana:2024.2', 'volumes': ['/etc/kolla/grafana/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'grafana_server': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000'}, 'grafana_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000'}}}})\u001b[0m\n\nTASK [grafana : Check if extra configuration file exists] **********************\nTuesday 24 February 2026  01:10:35 +0000 (0:00:00.778)       0:00:02.494 ****** \n\u001b[1;35m[WARNING]: Skipped '/operations/prometheus/grafana' path due to this access\u001b[0m\n\u001b[1;35missue: '/operations/prometheus/grafana' is not a directory\u001b[0m\n\u001b[0;32mok: [testbed-node-0 -> localhost]\u001b[0m\n\nTASK [grafana : include_tasks] *************************************************\nTuesday 24 February 2026  01:10:36 +0000 (0:00:00.998)       0:00:03.492 ****** \n\u001b[0;36mincluded: /ansible/roles/grafana/tasks/copy-certs.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [service-cert-copy : grafana | Copying over extra CA certificates] ********\nTuesday 24 February 2026  01:10:37 +0000 (0:00:00.919)       0:00:04.412 ****** \n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'grafana', 'value': {'container_name': 'grafana', 'group': 'grafana', 'enabled': True, 'image': 'registry.osism.tech/kolla/grafana:2024.2', 'volumes': ['/etc/kolla/grafana/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'grafana_server': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000'}, 'grafana_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'grafana', 'value': {'container_name': 'grafana', 'group': 'grafana', 'enabled': True, 'image': 'registry.osism.tech/kolla/grafana:2024.2', 'volumes': ['/etc/kolla/grafana/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'grafana_server': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000'}, 'grafana_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'grafana', 'value': {'container_name': 'grafana', 'group': 'grafana', 'enabled': True, 'image': 'registry.osism.tech/kolla/grafana:2024.2', 'volumes': ['/etc/kolla/grafana/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'grafana_server': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000'}, 'grafana_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000'}}}})\u001b[0m\n\nTASK [service-cert-copy : grafana | Copying over backend internal TLS certificate] ***\nTuesday 24 February 2026  01:10:38 +0000 (0:00:01.346)       0:00:05.758 ****** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'grafana', 'value': {'container_name': 'grafana', 'group': 'grafana', 'enabled': True, 'image': 'registry.osism.tech/kolla/grafana:2024.2', 'volumes': ['/etc/kolla/grafana/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'grafana_server': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000'}, 'grafana_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'grafana', 'value': {'container_name': 'grafana', 'group': 'grafana', 'enabled': True, 'image': 'registry.osism.tech/kolla/grafana:2024.2', 'volumes': ['/etc/kolla/grafana/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'grafana_server': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000'}, 'grafana_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'grafana', 'value': {'container_name': 'grafana', 'group': 'grafana', 'enabled': True, 'image': 'registry.osism.tech/kolla/grafana:2024.2', 'volumes': ['/etc/kolla/grafana/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'grafana_server': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000'}, 'grafana_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [service-cert-copy : grafana | Copying over backend internal TLS key] *****\nTuesday 24 February 2026  01:10:39 +0000 (0:00:00.459)       0:00:06.218 ****** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'grafana', 'value': {'container_name': 'grafana', 'group': 'grafana', 'enabled': True, 'image': 'registry.osism.tech/kolla/grafana:2024.2', 'volumes': ['/etc/kolla/grafana/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'grafana_server': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000'}, 'grafana_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'grafana', 'value': {'container_name': 'grafana', 'group': 'grafana', 'enabled': True, 'image': 'registry.osism.tech/kolla/grafana:2024.2', 'volumes': ['/etc/kolla/grafana/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'grafana_server': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000'}, 'grafana_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'grafana', 'value': {'container_name': 'grafana', 'group': 'grafana', 'enabled': True, 'image': 'registry.osism.tech/kolla/grafana:2024.2', 'volumes': ['/etc/kolla/grafana/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'grafana_server': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000'}, 'grafana_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [grafana : Copying over config.json files] ********************************\nTuesday 24 February 2026  01:10:40 +0000 (0:00:00.967)       0:00:07.185 ****** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'grafana', 'value': {'container_name': 'grafana', 'group': 'grafana', 'enabled': True, 'image': 'registry.osism.tech/kolla/grafana:2024.2', 'volumes': ['/etc/kolla/grafana/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'grafana_server': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000'}, 'grafana_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'grafana', 'value': {'container_name': 'grafana', 'group': 'grafana', 'enabled': True, 'image': 'registry.osism.tech/kolla/grafana:2024.2', 'volumes': ['/etc/kolla/grafana/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'grafana_server': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000'}, 'grafana_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'grafana', 'value': {'container_name': 'grafana', 'group': 'grafana', 'enabled': True, 'image': 'registry.osism.tech/kolla/grafana:2024.2', 'volumes': ['/etc/kolla/grafana/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'grafana_server': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000'}, 'grafana_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000'}}}})\u001b[0m\n\nTASK [grafana : Copying over grafana.ini] **************************************\nTuesday 24 February 2026  01:10:41 +0000 (0:00:01.228)       0:00:08.414 ****** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'grafana', 'value': {'container_name': 'grafana', 'group': 'grafana', 'enabled': True, 'image': 'registry.osism.tech/kolla/grafana:2024.2', 'volumes': ['/etc/kolla/grafana/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'grafana_server': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000'}, 'grafana_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'grafana', 'value': {'container_name': 'grafana', 'group': 'grafana', 'enabled': True, 'image': 'registry.osism.tech/kolla/grafana:2024.2', 'volumes': ['/etc/kolla/grafana/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'grafana_server': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000'}, 'grafana_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'grafana', 'value': {'container_name': 'grafana', 'group': 'grafana', 'enabled': True, 'image': 'registry.osism.tech/kolla/grafana:2024.2', 'volumes': ['/etc/kolla/grafana/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'grafana_server': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000'}, 'grafana_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000'}}}})\u001b[0m\n\nTASK [grafana : Copying over extra configuration file] *************************\nTuesday 24 February 2026  01:10:42 +0000 (0:00:01.471)       0:00:09.886 ****** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [grafana : Configuring Prometheus as data source for Grafana] *************\nTuesday 24 February 2026  01:10:43 +0000 (0:00:00.573)       0:00:10.460 ****** \n\u001b[0;33mchanged: [testbed-node-0] => (item=/ansible/roles/grafana/templates/prometheus.yaml.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=/ansible/roles/grafana/templates/prometheus.yaml.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=/ansible/roles/grafana/templates/prometheus.yaml.j2)\u001b[0m\n\nTASK [grafana : Configuring dashboards provisioning] ***************************\nTuesday 24 February 2026  01:10:44 +0000 (0:00:01.188)       0:00:11.648 ****** \n\u001b[0;33mchanged: [testbed-node-0] => (item=/opt/configuration/environments/kolla/files/overlays/grafana/provisioning.yaml)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=/opt/configuration/environments/kolla/files/overlays/grafana/provisioning.yaml)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=/opt/configuration/environments/kolla/files/overlays/grafana/provisioning.yaml)\u001b[0m\n\nTASK [grafana : Find custom grafana dashboards] ********************************\nTuesday 24 February 2026  01:10:45 +0000 (0:00:01.212)       0:00:12.861 ****** \n\u001b[0;32mok: [testbed-node-0 -> localhost]\u001b[0m\n\nTASK [grafana : Find templated grafana dashboards] *****************************\nTuesday 24 February 2026  01:10:46 +0000 (0:00:00.839)       0:00:13.700 ****** \n\u001b[1;35m[WARNING]: Skipped '/etc/kolla/grafana/dashboards' path due to this access\u001b[0m\n\u001b[1;35missue: '/etc/kolla/grafana/dashboards' is not a directory\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\nTASK [grafana : Prune templated Grafana dashboards] ****************************\nTuesday 24 February 2026  01:10:47 +0000 (0:00:00.792)       0:00:14.493 ****** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [grafana : Copying over custom dashboards] ********************************\nTuesday 24 February 2026  01:10:48 +0000 (0:00:00.780)       0:00:15.274 ****** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ceph/ceph-cluster-advanced.json', 'value': {'path': '/operations/grafana/dashboards/ceph/ceph-cluster-advanced.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 121701, 'inode': 1316890, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.147216, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ceph/ceph-cluster-advanced.json', 'value': {'path': '/operations/grafana/dashboards/ceph/ceph-cluster-advanced.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 121701, 'inode': 1316890, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.147216, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ceph/ceph-cluster-advanced.json', 'value': {'path': '/operations/grafana/dashboards/ceph/ceph-cluster-advanced.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 121701, 'inode': 1316890, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.147216, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ceph/cephfsdashboard.json', 'value': {'path': '/operations/grafana/dashboards/ceph/cephfsdashboard.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 143913, 'inode': 1316910, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.155399, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ceph/cephfsdashboard.json', 'value': {'path': '/operations/grafana/dashboards/ceph/cephfsdashboard.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 143913, 'inode': 1316910, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.155399, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ceph/cephfsdashboard.json', 'value': {'path': '/operations/grafana/dashboards/ceph/cephfsdashboard.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 143913, 'inode': 1316910, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.155399, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ceph/rbd-overview.json', 'value': {'path': '/operations/grafana/dashboards/ceph/rbd-overview.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 26019, 'inode': 1316976, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.180783, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ceph/rbd-overview.json', 'value': {'path': '/operations/grafana/dashboards/ceph/rbd-overview.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 26019, 'inode': 1316976, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.180783, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ceph/rbd-overview.json', 'value': {'path': '/operations/grafana/dashboards/ceph/rbd-overview.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 26019, 'inode': 1316976, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.180783, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ceph/ceph_pools.json', 'value': {'path': '/operations/grafana/dashboards/ceph/ceph_pools.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 25279, 'inode': 1316904, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.152629, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ceph/ceph_pools.json', 'value': {'path': '/operations/grafana/dashboards/ceph/ceph_pools.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 25279, 'inode': 1316904, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.152629, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ceph/ceph_pools.json', 'value': {'path': '/operations/grafana/dashboards/ceph/ceph_pools.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 25279, 'inode': 1316904, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.152629, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ceph/rgw-s3-analytics.json', 'value': {'path': '/operations/grafana/dashboards/ceph/rgw-s3-analytics.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 170293, 'inode': 1316977, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.1822166, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ceph/rgw-s3-analytics.json', 'value': {'path': '/operations/grafana/dashboards/ceph/rgw-s3-analytics.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 170293, 'inode': 1316977, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.1822166, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ceph/rgw-s3-analytics.json', 'value': {'path': '/operations/grafana/dashboards/ceph/rgw-s3-analytics.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 170293, 'inode': 1316977, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.1822166, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ceph/ceph-nvmeof-performance.json', 'value': {'path': '/operations/grafana/dashboards/ceph/ceph-nvmeof-performance.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 33297, 'inode': 1316894, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.1502383, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ceph/ceph-nvmeof-performance.json', 'value': {'path': '/operations/grafana/dashboards/ceph/ceph-nvmeof-performance.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 33297, 'inode': 1316894, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.1502383, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ceph/ceph-nvmeof-performance.json', 'value': {'path': '/operations/grafana/dashboards/ceph/ceph-nvmeof-performance.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 33297, 'inode': 1316894, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.1502383, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ceph/osd-device-details.json', 'value': {'path': '/operations/grafana/dashboards/ceph/osd-device-details.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 26346, 'inode': 1316926, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.1593084, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ceph/osd-device-details.json', 'value': {'path': '/operations/grafana/dashboards/ceph/osd-device-details.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 26346, 'inode': 1316926, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.1593084, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ceph/osd-device-details.json', 'value': {'path': '/operations/grafana/dashboards/ceph/osd-device-details.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 26346, 'inode': 1316926, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.1593084, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ceph/radosgw-overview.json', 'value': {'path': '/operations/grafana/dashboards/ceph/radosgw-overview.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 46110, 'inode': 1316969, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.1797712, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ceph/radosgw-overview.json', 'value': {'path': '/operations/grafana/dashboards/ceph/radosgw-overview.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 46110, 'inode': 1316969, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.1797712, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ceph/radosgw-overview.json', 'value': {'path': '/operations/grafana/dashboards/ceph/radosgw-overview.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 46110, 'inode': 1316969, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.1797712, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ceph/README.md', 'value': {'path': '/operations/grafana/dashboards/ceph/README.md', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 84, 'inode': 1316889, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.1467896, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ceph/README.md', 'value': {'path': '/operations/grafana/dashboards/ceph/README.md', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 84, 'inode': 1316889, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.1467896, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ceph/README.md', 'value': {'path': '/operations/grafana/dashboards/ceph/README.md', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 84, 'inode': 1316889, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.1467896, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ceph/ceph-cluster.json', 'value': {'path': '/operations/grafana/dashboards/ceph/ceph-cluster.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 34113, 'inode': 1316891, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.1491747, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ceph/ceph-cluster.json', 'value': {'path': '/operations/grafana/dashboards/ceph/ceph-cluster.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 34113, 'inode': 1316891, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.1491747, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ceph/ceph-cluster.json', 'value': {'path': '/operations/grafana/dashboards/ceph/ceph-cluster.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 34113, 'inode': 1316891, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.1491747, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ceph/cephfs-overview.json', 'value': {'path': '/operations/grafana/dashboards/ceph/cephfs-overview.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 9025, 'inode': 1316907, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.1528525, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ceph/cephfs-overview.json', 'value': {'path': '/operations/grafana/dashboards/ceph/cephfs-overview.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 9025, 'inode': 1316907, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.1528525, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ceph/cephfs-overview.json', 'value': {'path': '/operations/grafana/dashboards/ceph/cephfs-overview.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 9025, 'inode': 1316907, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.1528525, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ceph/pool-detail.json', 'value': {'path': '/operations/grafana/dashboards/ceph/pool-detail.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 19231, 'inode': 1316930, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.160852, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ceph/pool-detail.json', 'value': {'path': '/operations/grafana/dashboards/ceph/pool-detail.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 19231, 'inode': 1316930, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.160852, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ceph/pool-detail.json', 'value': {'path': '/operations/grafana/dashboards/ceph/pool-detail.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 19231, 'inode': 1316930, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.160852, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ceph/rbd-details.json', 'value': {'path': '/operations/grafana/dashboards/ceph/rbd-details.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 13320, 'inode': 1316975, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.1803944, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ceph/rbd-details.json', 'value': {'path': '/operations/grafana/dashboards/ceph/rbd-details.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 13320, 'inode': 1316975, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.1803944, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ceph/rbd-details.json', 'value': {'path': '/operations/grafana/dashboards/ceph/rbd-details.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 13320, 'inode': 1316975, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.1803944, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ceph/ceph_overview.json', 'value': {'path': '/operations/grafana/dashboards/ceph/ceph_overview.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 80386, 'inode': 1316900, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.1517806, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ceph/ceph_overview.json', 'value': {'path': '/operations/grafana/dashboards/ceph/ceph_overview.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 80386, 'inode': 1316900, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.1517806, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ceph/ceph_overview.json', 'value': {'path': '/operations/grafana/dashboards/ceph/ceph_overview.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 80386, 'inode': 1316900, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.1517806, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ceph/radosgw-detail.json', 'value': {'path': '/operations/grafana/dashboards/ceph/radosgw-detail.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 20042, 'inode': 1316935, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.1782167, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ceph/radosgw-detail.json', 'value': {'path': '/operations/grafana/dashboards/ceph/radosgw-detail.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 20042, 'inode': 1316935, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.1782167, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ceph/radosgw-detail.json', 'value': {'path': '/operations/grafana/dashboards/ceph/radosgw-detail.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 20042, 'inode': 1316935, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.1782167, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ceph/smb-overview.json', 'value': {'path': '/operations/grafana/dashboards/ceph/smb-overview.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 29877, 'inode': 1316980, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.1822166, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ceph/smb-overview.json', 'value': {'path': '/operations/grafana/dashboards/ceph/smb-overview.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 29877, 'inode': 1316980, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.1822166, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ceph/smb-overview.json', 'value': {'path': '/operations/grafana/dashboards/ceph/smb-overview.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 29877, 'inode': 1316980, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.1822166, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ceph/osds-overview.json', 'value': {'path': '/operations/grafana/dashboards/ceph/osds-overview.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 38375, 'inode': 1316927, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.160412, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ceph/osds-overview.json', 'value': {'path': '/operations/grafana/dashboards/ceph/osds-overview.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 38375, 'inode': 1316927, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.160412, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ceph/osds-overview.json', 'value': {'path': '/operations/grafana/dashboards/ceph/osds-overview.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 38375, 'inode': 1316927, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.160412, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ceph/multi-cluster-overview.json', 'value': {'path': '/operations/grafana/dashboards/ceph/multi-cluster-overview.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 63043, 'inode': 1316922, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.1593084, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ceph/multi-cluster-overview.json', 'value': {'path': '/operations/grafana/dashboards/ceph/multi-cluster-overview.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 63043, 'inode': 1316922, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.1593084, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ceph/multi-cluster-overview.json', 'value': {'path': '/operations/grafana/dashboards/ceph/multi-cluster-overview.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 63043, 'inode': 1316922, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.1593084, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ceph/hosts-overview.json', 'value': {'path': '/operations/grafana/dashboards/ceph/hosts-overview.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 27387, 'inode': 1316919, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.1575785, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ceph/hosts-overview.json', 'value': {'path': '/operations/grafana/dashboards/ceph/hosts-overview.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 27387, 'inode': 1316919, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.1575785, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ceph/hosts-overview.json', 'value': {'path': '/operations/grafana/dashboards/ceph/hosts-overview.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 27387, 'inode': 1316919, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.1575785, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ceph/pool-overview.json', 'value': {'path': '/operations/grafana/dashboards/ceph/pool-overview.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 49016, 'inode': 1316932, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.1612163, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ceph/pool-overview.json', 'value': {'path': '/operations/grafana/dashboards/ceph/pool-overview.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 49016, 'inode': 1316932, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.1612163, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ceph/pool-overview.json', 'value': {'path': '/operations/grafana/dashboards/ceph/pool-overview.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 49016, 'inode': 1316932, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.1612163, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ceph/host-details.json', 'value': {'path': '/operations/grafana/dashboards/ceph/host-details.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 43303, 'inode': 1316916, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.1567206, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ceph/host-details.json', 'value': {'path': '/operations/grafana/dashboards/ceph/host-details.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 43303, 'inode': 1316916, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.1567206, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ceph/host-details.json', 'value': {'path': '/operations/grafana/dashboards/ceph/host-details.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 43303, 'inode': 1316916, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.1567206, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ceph/radosgw-sync-overview.json', 'value': {'path': '/operations/grafana/dashboards/ceph/radosgw-sync-overview.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 16614, 'inode': 1316973, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.1800752, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ceph/radosgw-sync-overview.json', 'value': {'path': '/operations/grafana/dashboards/ceph/radosgw-sync-overview.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 16614, 'inode': 1316973, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.1800752, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ceph/radosgw-sync-overview.json', 'value': {'path': '/operations/grafana/dashboards/ceph/radosgw-sync-overview.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 16614, 'inode': 1316973, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.1800752, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ceph/ceph-nvmeof.json', 'value': {'path': '/operations/grafana/dashboards/ceph/ceph-nvmeof.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 52667, 'inode': 1316898, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.1502383, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ceph/ceph-nvmeof.json', 'value': {'path': '/operations/grafana/dashboards/ceph/ceph-nvmeof.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 52667, 'inode': 1316898, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.1502383, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ceph/ceph-nvmeof.json', 'value': {'path': '/operations/grafana/dashboards/ceph/ceph-nvmeof.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 52667, 'inode': 1316898, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.1502383, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'openstack/openstack.json', 'value': {'path': '/operations/grafana/dashboards/openstack/openstack.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 57270, 'inode': 1317236, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.3583703, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'openstack/openstack.json', 'value': {'path': '/operations/grafana/dashboards/openstack/openstack.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 57270, 'inode': 1317236, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.3583703, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'openstack/openstack.json', 'value': {'path': '/operations/grafana/dashboards/openstack/openstack.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 57270, 'inode': 1317236, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.3583703, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'infrastructure/haproxy.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/haproxy.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 410814, 'inode': 1317001, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.200857, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'infrastructure/haproxy.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/haproxy.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 410814, 'inode': 1317001, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.200857, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'infrastructure/haproxy.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/haproxy.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 410814, 'inode': 1317001, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.200857, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'infrastructure/database.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/database.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 30898, 'inode': 1316990, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.1872168, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'infrastructure/database.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/database.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 30898, 'inode': 1316990, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.1872168, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'infrastructure/database.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/database.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 30898, 'inode': 1316990, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.1872168, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'infrastructure/node-rsrc-use.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/node-rsrc-use.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 15767, 'inode': 1317017, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.2022812, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'infrastructure/node-rsrc-use.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/node-rsrc-use.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 15767, 'inode': 1317017, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.2022812, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'infrastructure/node-rsrc-use.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/node-rsrc-use.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 15767, 'inode': 1317017, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.2022812, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'infrastructure/alertmanager-overview.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/alertmanager-overview.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 9645, 'inode': 1316984, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.1841078, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'infrastructure/alertmanager-overview.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/alertmanager-overview.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 9645, 'inode': 1316984, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.1841078, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'infrastructure/alertmanager-overview.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/alertmanager-overview.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 9645, 'inode': 1316984, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.1841078, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'infrastructure/opensearch.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/opensearch.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 65458, 'inode': 1317050, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.2272177, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'infrastructure/opensearch.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/opensearch.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 65458, 'inode': 1317050, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.2272177, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'infrastructure/opensearch.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/opensearch.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 65458, 'inode': 1317050, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.2272177, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'infrastructure/node_exporter_full.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/node_exporter_full.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 682774, 'inode': 1317019, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.2222176, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'infrastructure/node_exporter_full.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/node_exporter_full.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 682774, 'inode': 1317019, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.2222176, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'infrastructure/node_exporter_full.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/node_exporter_full.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 682774, 'inode': 1317019, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.2222176, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'infrastructure/prometheus-remote-write.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/prometheus-remote-write.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 22303, 'inode': 1317051, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.2282176, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'infrastructure/prometheus-remote-write.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/prometheus-remote-write.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 22303, 'inode': 1317051, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.2282176, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'infrastructure/prometheus-remote-write.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/prometheus-remote-write.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 22303, 'inode': 1317051, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.2282176, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'infrastructure/redfish.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/redfish.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 38087, 'inode': 1317055, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.2342176, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'infrastructure/redfish.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/redfish.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 38087, 'inode': 1317055, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.2342176, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'infrastructure/redfish.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/redfish.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 38087, 'inode': 1317055, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.2342176, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'infrastructure/nodes.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/nodes.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 21194, 'inode': 1317049, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.2242174, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'infrastructure/nodes.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/nodes.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 21194, 'inode': 1317049, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.2242174, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'infrastructure/nodes.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/nodes.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 21194, 'inode': 1317049, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.2242174, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'infrastructure/memcached.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/memcached.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 24243, 'inode': 1317013, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.2016113, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'infrastructure/memcached.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/memcached.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 24243, 'inode': 1317013, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.2016113, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'infrastructure/memcached.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/memcached.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 24243, 'inode': 1317013, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.2016113, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'infrastructure/fluentd.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/fluentd.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 82960, 'inode': 1316998, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.194241, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'infrastructure/fluentd.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/fluentd.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 82960, 'inode': 1316998, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.194241, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'infrastructure/fluentd.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/fluentd.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 82960, 'inode': 1316998, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.194241, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'infrastructure/libvirt.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/libvirt.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 29672, 'inode': 1317011, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.2011416, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'infrastructure/libvirt.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/libvirt.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 29672, 'inode': 1317011, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.2011416, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'infrastructure/libvirt.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/libvirt.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 29672, 'inode': 1317011, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.2011416, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'infrastructure/elasticsearch.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/elasticsearch.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 187864, 'inode': 1316991, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.1930285, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'infrastructure/elasticsearch.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/elasticsearch.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 187864, 'inode': 1316991, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.1930285, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'infrastructure/elasticsearch.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/elasticsearch.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 187864, 'inode': 1316991, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.1930285, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'infrastructure/node-cluster-rsrc-use.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/node-cluster-rsrc-use.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 15957, 'inode': 1317015, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.2019198, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'infrastructure/node-cluster-rsrc-use.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/node-cluster-rsrc-use.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 15957, 'inode': 1317015, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.2019198, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'infrastructure/node-cluster-rsrc-use.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/node-cluster-rsrc-use.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 15957, 'inode': 1317015, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.2019198, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'infrastructure/rabbitmq.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/rabbitmq.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 222049, 'inode': 1317054, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.2332177, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'infrastructure/rabbitmq.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/rabbitmq.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 222049, 'inode': 1317054, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.2332177, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'infrastructure/rabbitmq.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/rabbitmq.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 222049, 'inode': 1317054, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.2332177, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'infrastructure/prometheus_alertmanager.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/prometheus_alertmanager.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 115472, 'inode': 1317053, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.2312176, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'infrastructure/prometheus_alertmanager.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/prometheus_alertmanager.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 115472, 'inode': 1317053, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.2312176, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'infrastructure/prometheus_alertmanager.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/prometheus_alertmanager.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 115472, 'inode': 1317053, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.2312176, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'infrastructure/blackbox.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/blackbox.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 31128, 'inode': 1316985, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.1844537, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'infrastructure/blackbox.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/blackbox.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 31128, 'inode': 1316985, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.1844537, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'infrastructure/blackbox.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/blackbox.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 31128, 'inode': 1316985, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.1844537, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'infrastructure/cadvisor.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/cadvisor.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 53882, 'inode': 1316987, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.1867146, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'infrastructure/cadvisor.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/cadvisor.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 53882, 'inode': 1316987, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.1867146, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'infrastructure/cadvisor.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/cadvisor.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 53882, 'inode': 1316987, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.1867146, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'infrastructure/node_exporter_side_by_side.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/node_exporter_side_by_side.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 70691, 'inode': 1317048, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.2232175, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'infrastructure/node_exporter_side_by_side.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/node_exporter_side_by_side.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 70691, 'inode': 1317048, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.2232175, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'infrastructure/node_exporter_side_by_side.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/node_exporter_side_by_side.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 70691, 'inode': 1317048, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.2232175, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'infrastructure/prometheus.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/prometheus.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 21951, 'inode': 1317052, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.2292175, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'infrastructure/prometheus.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/prometheus.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 21951, 'inode': 1317052, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.2292175, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'infrastructure/prometheus.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/prometheus.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 21951, 'inode': 1317052, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.2292175, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m\n\nTASK [grafana : Check grafana containers] **************************************\nTuesday 24 February 2026  01:11:29 +0000 (0:00:41.241)       0:00:56.515 ****** \n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'grafana', 'value': {'container_name': 'grafana', 'group': 'grafana', 'enabled': True, 'image': 'registry.osism.tech/kolla/grafana:2024.2', 'volumes': ['/etc/kolla/grafana/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'grafana_server': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000'}, 'grafana_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'grafana', 'value': {'container_name': 'grafana', 'group': 'grafana', 'enabled': True, 'image': 'registry.osism.tech/kolla/grafana:2024.2', 'volumes': ['/etc/kolla/grafana/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'grafana_server': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000'}, 'grafana_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'grafana', 'value': {'container_name': 'grafana', 'group': 'grafana', 'enabled': True, 'image': 'registry.osism.tech/kolla/grafana:2024.2', 'volumes': ['/etc/kolla/grafana/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'grafana_server': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000'}, 'grafana_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000'}}}})\u001b[0m\n\nTASK [grafana : Creating grafana database] *************************************\nTuesday 24 February 2026  01:11:30 +0000 (0:00:01.099)       0:00:57.614 ****** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [grafana : Creating grafana database user and setting permissions] ********\nTuesday 24 February 2026  01:11:33 +0000 (0:00:02.638)       0:01:00.253 ****** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [grafana : Flush handlers] ************************************************\nTuesday 24 February 2026  01:11:35 +0000 (0:00:02.735)       0:01:02.989 ****** \n\nTASK [grafana : Flush handlers] ************************************************\nTuesday 24 February 2026  01:11:35 +0000 (0:00:00.066)       0:01:03.055 ****** \n\nTASK [grafana : Flush handlers] ************************************************\nTuesday 24 February 2026  01:11:36 +0000 (0:00:00.257)       0:01:03.313 ****** \n\nRUNNING HANDLER [grafana : Restart first grafana container] ********************\nTuesday 24 February 2026  01:11:36 +0000 (0:00:00.072)       0:01:03.385 ****** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nRUNNING HANDLER [grafana : Waiting for grafana to start on first node] *********\nTuesday 24 February 2026  01:11:38 +0000 (0:00:02.118)       0:01:05.504 ****** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[1;30mFAILED - RETRYING: [testbed-node-0]: Waiting for grafana to start on first node (12 retries left).\u001b[0m\n\u001b[1;30mFAILED - RETRYING: [testbed-node-0]: Waiting for grafana to start on first node (11 retries left).\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nRUNNING HANDLER [grafana : Restart remaining grafana containers] ***************\nTuesday 24 February 2026  01:12:05 +0000 (0:00:27.152)       0:01:32.656 ****** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\nTASK [grafana : Wait for grafana application ready] ****************************\nTuesday 24 February 2026  01:12:30 +0000 (0:00:24.871)       0:01:57.527 ****** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [grafana : Remove old grafana docker volume] ******************************\nTuesday 24 February 2026  01:12:32 +0000 (0:00:02.381)       0:01:59.909 ****** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [grafana : Enable grafana datasources] ************************************\nTuesday 24 February 2026  01:12:33 +0000 (0:00:00.600)       0:02:00.510 ****** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'influxdb', 'value': {'enabled': False, 'data': {'isDefault': True, 'database': 'telegraf', 'name': 'telegraf', 'type': 'influxdb', 'url': 'https://api-int.testbed.osism.xyz:8086', 'access': 'proxy', 'basicAuth': False}}}) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'opensearch', 'value': {'enabled': True, 'data': {'name': 'opensearch', 'type': 'grafana-opensearch-datasource', 'access': 'proxy', 'url': 'https://api-int.testbed.osism.xyz:9200', 'jsonData': {'flavor': 'OpenSearch', 'database': 'flog-*', 'version': '2.11.1', 'timeField': '@timestamp', 'logLevelField': 'log_level'}}}})\u001b[0m\n\nTASK [grafana : Disable Getting Started panel] *********************************\nTuesday 24 February 2026  01:12:36 +0000 (0:00:02.886)       0:02:03.396 ****** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=21  \u001b[0m \u001b[0;33mchanged=12  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=7   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=14  \u001b[0m \u001b[0;33mchanged=9   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=7   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=14  \u001b[0m \u001b[0;33mchanged=9   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=7   \u001b[0m rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nTuesday 24 February 2026  01:12:36 +0000 (0:00:00.295)       0:02:03.692 ****** \n=============================================================================== \ngrafana : Copying over custom dashboards ------------------------------- 41.24s\ngrafana : Waiting for grafana to start on first node ------------------- 27.15s\ngrafana : Restart remaining grafana containers ------------------------- 24.87s\ngrafana : Enable grafana datasources ------------------------------------ 2.89s\ngrafana : Creating grafana database user and setting permissions -------- 2.74s\ngrafana : Creating grafana database ------------------------------------- 2.64s\ngrafana : Wait for grafana application ready ---------------------------- 2.38s\ngrafana : Restart first grafana container ------------------------------- 2.12s\ngrafana : Copying over grafana.ini -------------------------------------- 1.47s\nservice-cert-copy : grafana | Copying over extra CA certificates -------- 1.35s\ngrafana : Copying over config.json files -------------------------------- 1.23s\ngrafana : Configuring dashboards provisioning --------------------------- 1.21s\ngrafana : Configuring Prometheus as data source for Grafana ------------- 1.19s\ngrafana : Check grafana containers -------------------------------------- 1.10s\ngrafana : Check if extra configuration file exists ---------------------- 1.00s\nservice-cert-copy : grafana | Copying over backend internal TLS key ----- 0.97s\ngrafana : include_tasks ------------------------------------------------- 0.92s\ngrafana : Find custom grafana dashboards -------------------------------- 0.84s\ngrafana : Find templated grafana dashboards ----------------------------- 0.79s\ngrafana : Prune templated Grafana dashboards ---------------------------- 0.78s\n\u001b[32m2026-02-24 01:13:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 79f56a03-a0d2-400b-aa03-9d9321f018b2 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:13:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4baf3a0c-387b-4215-b490-5edd972623e9 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:13:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 01:13:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 79f56a03-a0d2-400b-aa03-9d9321f018b2 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:13:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4baf3a0c-387b-4215-b490-5edd972623e9 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:13:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 01:13:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 79f56a03-a0d2-400b-aa03-9d9321f018b2 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:13:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4baf3a0c-387b-4215-b490-5edd972623e9 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:13:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 01:13:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 79f56a03-a0d2-400b-aa03-9d9321f018b2 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:13:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4baf3a0c-387b-4215-b490-5edd972623e9 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:13:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 01:13:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 79f56a03-a0d2-400b-aa03-9d9321f018b2 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:13:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4baf3a0c-387b-4215-b490-5edd972623e9 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:13:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 01:13:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 79f56a03-a0d2-400b-aa03-9d9321f018b2 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:13:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4baf3a0c-387b-4215-b490-5edd972623e9 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:13:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 01:13:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 79f56a03-a0d2-400b-aa03-9d9321f018b2 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:13:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4baf3a0c-387b-4215-b490-5edd972623e9 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:13:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 01:13:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 79f56a03-a0d2-400b-aa03-9d9321f018b2 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:13:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4baf3a0c-387b-4215-b490-5edd972623e9 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:13:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 01:13:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 79f56a03-a0d2-400b-aa03-9d9321f018b2 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:13:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4baf3a0c-387b-4215-b490-5edd972623e9 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:13:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 01:13:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 79f56a03-a0d2-400b-aa03-9d9321f018b2 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:13:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4baf3a0c-387b-4215-b490-5edd972623e9 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:13:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 01:14:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 79f56a03-a0d2-400b-aa03-9d9321f018b2 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:14:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4baf3a0c-387b-4215-b490-5edd972623e9 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:14:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 01:14:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 79f56a03-a0d2-400b-aa03-9d9321f018b2 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:14:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4baf3a0c-387b-4215-b490-5edd972623e9 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:14:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 01:14:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 79f56a03-a0d2-400b-aa03-9d9321f018b2 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:14:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4baf3a0c-387b-4215-b490-5edd972623e9 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:14:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 01:14:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 79f56a03-a0d2-400b-aa03-9d9321f018b2 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:14:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4baf3a0c-387b-4215-b490-5edd972623e9 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:14:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 01:14:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 79f56a03-a0d2-400b-aa03-9d9321f018b2 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:14:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4baf3a0c-387b-4215-b490-5edd972623e9 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:14:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 01:14:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 79f56a03-a0d2-400b-aa03-9d9321f018b2 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:14:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4baf3a0c-387b-4215-b490-5edd972623e9 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:14:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 01:14:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 79f56a03-a0d2-400b-aa03-9d9321f018b2 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:14:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4baf3a0c-387b-4215-b490-5edd972623e9 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:14:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 01:14:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 79f56a03-a0d2-400b-aa03-9d9321f018b2 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:14:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4baf3a0c-387b-4215-b490-5edd972623e9 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:14:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 01:14:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 79f56a03-a0d2-400b-aa03-9d9321f018b2 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:14:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4baf3a0c-387b-4215-b490-5edd972623e9 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:14:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 01:14:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 79f56a03-a0d2-400b-aa03-9d9321f018b2 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:14:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4baf3a0c-387b-4215-b490-5edd972623e9 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:14:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 01:14:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 79f56a03-a0d2-400b-aa03-9d9321f018b2 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:14:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4baf3a0c-387b-4215-b490-5edd972623e9 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:14:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 01:14:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 79f56a03-a0d2-400b-aa03-9d9321f018b2 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:14:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4baf3a0c-387b-4215-b490-5edd972623e9 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:14:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 01:14:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 79f56a03-a0d2-400b-aa03-9d9321f018b2 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:14:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4baf3a0c-387b-4215-b490-5edd972623e9 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:14:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 01:14:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 79f56a03-a0d2-400b-aa03-9d9321f018b2 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:14:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4baf3a0c-387b-4215-b490-5edd972623e9 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:14:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 01:14:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 79f56a03-a0d2-400b-aa03-9d9321f018b2 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:14:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4baf3a0c-387b-4215-b490-5edd972623e9 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:14:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 01:14:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 79f56a03-a0d2-400b-aa03-9d9321f018b2 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:14:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4baf3a0c-387b-4215-b490-5edd972623e9 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:14:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 01:14:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 79f56a03-a0d2-400b-aa03-9d9321f018b2 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:14:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4baf3a0c-387b-4215-b490-5edd972623e9 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:14:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 01:14:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 79f56a03-a0d2-400b-aa03-9d9321f018b2 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:14:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4baf3a0c-387b-4215-b490-5edd972623e9 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:14:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 01:14:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 79f56a03-a0d2-400b-aa03-9d9321f018b2 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:14:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4baf3a0c-387b-4215-b490-5edd972623e9 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:14:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 01:15:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 79f56a03-a0d2-400b-aa03-9d9321f018b2 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:15:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4baf3a0c-387b-4215-b490-5edd972623e9 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:15:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 01:15:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 79f56a03-a0d2-400b-aa03-9d9321f018b2 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:15:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4baf3a0c-387b-4215-b490-5edd972623e9 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:15:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 01:15:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 79f56a03-a0d2-400b-aa03-9d9321f018b2 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:15:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4baf3a0c-387b-4215-b490-5edd972623e9 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:15:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 01:15:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 79f56a03-a0d2-400b-aa03-9d9321f018b2 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:15:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4baf3a0c-387b-4215-b490-5edd972623e9 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:15:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 01:15:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 79f56a03-a0d2-400b-aa03-9d9321f018b2 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:15:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4baf3a0c-387b-4215-b490-5edd972623e9 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:15:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 01:15:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 79f56a03-a0d2-400b-aa03-9d9321f018b2 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:15:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4baf3a0c-387b-4215-b490-5edd972623e9 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:15:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 01:15:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 79f56a03-a0d2-400b-aa03-9d9321f018b2 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:15:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4baf3a0c-387b-4215-b490-5edd972623e9 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:15:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 01:15:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 79f56a03-a0d2-400b-aa03-9d9321f018b2 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:15:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4baf3a0c-387b-4215-b490-5edd972623e9 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:15:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 01:15:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 79f56a03-a0d2-400b-aa03-9d9321f018b2 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:15:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4baf3a0c-387b-4215-b490-5edd972623e9 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:15:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 01:15:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 79f56a03-a0d2-400b-aa03-9d9321f018b2 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:15:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4baf3a0c-387b-4215-b490-5edd972623e9 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:15:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 01:15:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 79f56a03-a0d2-400b-aa03-9d9321f018b2 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:15:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4baf3a0c-387b-4215-b490-5edd972623e9 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:15:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 01:15:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 79f56a03-a0d2-400b-aa03-9d9321f018b2 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:15:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4baf3a0c-387b-4215-b490-5edd972623e9 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:15:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 01:15:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 79f56a03-a0d2-400b-aa03-9d9321f018b2 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:15:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4baf3a0c-387b-4215-b490-5edd972623e9 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:15:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 01:15:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 79f56a03-a0d2-400b-aa03-9d9321f018b2 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:15:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4baf3a0c-387b-4215-b490-5edd972623e9 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:15:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 01:15:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 79f56a03-a0d2-400b-aa03-9d9321f018b2 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:15:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4baf3a0c-387b-4215-b490-5edd972623e9 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:15:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 01:15:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 79f56a03-a0d2-400b-aa03-9d9321f018b2 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:15:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4baf3a0c-387b-4215-b490-5edd972623e9 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:15:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 01:15:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 79f56a03-a0d2-400b-aa03-9d9321f018b2 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:15:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4baf3a0c-387b-4215-b490-5edd972623e9 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:15:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 01:15:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 79f56a03-a0d2-400b-aa03-9d9321f018b2 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:15:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4baf3a0c-387b-4215-b490-5edd972623e9 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:15:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 01:15:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 79f56a03-a0d2-400b-aa03-9d9321f018b2 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:15:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4baf3a0c-387b-4215-b490-5edd972623e9 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:15:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 01:15:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 79f56a03-a0d2-400b-aa03-9d9321f018b2 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:15:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4baf3a0c-387b-4215-b490-5edd972623e9 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:15:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 01:16:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 79f56a03-a0d2-400b-aa03-9d9321f018b2 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:16:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4baf3a0c-387b-4215-b490-5edd972623e9 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:16:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 01:16:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 79f56a03-a0d2-400b-aa03-9d9321f018b2 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:16:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4baf3a0c-387b-4215-b490-5edd972623e9 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:16:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 01:16:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 79f56a03-a0d2-400b-aa03-9d9321f018b2 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:16:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4baf3a0c-387b-4215-b490-5edd972623e9 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:16:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 01:16:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 79f56a03-a0d2-400b-aa03-9d9321f018b2 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:16:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4baf3a0c-387b-4215-b490-5edd972623e9 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:16:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 01:16:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 79f56a03-a0d2-400b-aa03-9d9321f018b2 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:16:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4baf3a0c-387b-4215-b490-5edd972623e9 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:16:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 01:16:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 79f56a03-a0d2-400b-aa03-9d9321f018b2 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:16:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4baf3a0c-387b-4215-b490-5edd972623e9 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:16:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 01:16:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 79f56a03-a0d2-400b-aa03-9d9321f018b2 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:16:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4baf3a0c-387b-4215-b490-5edd972623e9 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:16:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 01:16:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 79f56a03-a0d2-400b-aa03-9d9321f018b2 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:16:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4baf3a0c-387b-4215-b490-5edd972623e9 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:16:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 01:16:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 79f56a03-a0d2-400b-aa03-9d9321f018b2 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:16:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4baf3a0c-387b-4215-b490-5edd972623e9 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:16:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 01:16:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 79f56a03-a0d2-400b-aa03-9d9321f018b2 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:16:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4baf3a0c-387b-4215-b490-5edd972623e9 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:16:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 01:16:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 79f56a03-a0d2-400b-aa03-9d9321f018b2 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:16:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4baf3a0c-387b-4215-b490-5edd972623e9 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:16:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 01:16:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 79f56a03-a0d2-400b-aa03-9d9321f018b2 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:16:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4baf3a0c-387b-4215-b490-5edd972623e9 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:16:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 01:16:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 79f56a03-a0d2-400b-aa03-9d9321f018b2 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:16:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4baf3a0c-387b-4215-b490-5edd972623e9 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:16:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 01:16:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 79f56a03-a0d2-400b-aa03-9d9321f018b2 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:16:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4baf3a0c-387b-4215-b490-5edd972623e9 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:16:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 01:16:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 79f56a03-a0d2-400b-aa03-9d9321f018b2 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:16:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4baf3a0c-387b-4215-b490-5edd972623e9 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:16:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 01:16:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 79f56a03-a0d2-400b-aa03-9d9321f018b2 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:16:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4baf3a0c-387b-4215-b490-5edd972623e9 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:16:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 01:16:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 79f56a03-a0d2-400b-aa03-9d9321f018b2 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:16:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4baf3a0c-387b-4215-b490-5edd972623e9 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:16:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 01:16:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 79f56a03-a0d2-400b-aa03-9d9321f018b2 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:16:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4baf3a0c-387b-4215-b490-5edd972623e9 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:16:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 01:16:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 79f56a03-a0d2-400b-aa03-9d9321f018b2 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:16:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4baf3a0c-387b-4215-b490-5edd972623e9 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:16:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 01:16:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 79f56a03-a0d2-400b-aa03-9d9321f018b2 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:16:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4baf3a0c-387b-4215-b490-5edd972623e9 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:16:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 01:17:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 79f56a03-a0d2-400b-aa03-9d9321f018b2 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:17:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4baf3a0c-387b-4215-b490-5edd972623e9 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:17:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 01:17:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 79f56a03-a0d2-400b-aa03-9d9321f018b2 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:17:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4baf3a0c-387b-4215-b490-5edd972623e9 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:17:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 01:17:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 79f56a03-a0d2-400b-aa03-9d9321f018b2 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:17:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4baf3a0c-387b-4215-b490-5edd972623e9 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:17:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 01:17:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 79f56a03-a0d2-400b-aa03-9d9321f018b2 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:17:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4baf3a0c-387b-4215-b490-5edd972623e9 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:17:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 01:17:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 79f56a03-a0d2-400b-aa03-9d9321f018b2 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:17:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4baf3a0c-387b-4215-b490-5edd972623e9 is in state SUCCESS\u001b[0m\n\n\nPLAY [Group hosts based on configuration] **************************************\n\nTASK [Group hosts based on OpenStack release] **********************************\nTuesday 24 February 2026  01:07:57 +0000 (0:00:00.344)       0:00:00.344 ****** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nTASK [Group hosts based on Kolla action] ***************************************\nTuesday 24 February 2026  01:07:58 +0000 (0:00:01.277)       0:00:01.622 ****** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nTASK [Group hosts based on enabled services] ***********************************\nTuesday 24 February 2026  01:07:59 +0000 (0:00:00.952)       0:00:02.574 ****** \n\u001b[0;33mchanged: [testbed-manager] => (item=enable_nova_True)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=enable_nova_True)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=enable_nova_True)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=enable_nova_True)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item=enable_nova_True)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=enable_nova_True)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=enable_nova_True)\u001b[0m\n\nPLAY [Bootstrap nova API databases] ********************************************\n\nTASK [Bootstrap deploy] ********************************************************\nTuesday 24 February 2026  01:08:00 +0000 (0:00:00.931)       0:00:03.506 ****** \n\u001b[0;36mincluded: nova for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [nova : Creating Nova databases] ******************************************\nTuesday 24 February 2026  01:08:01 +0000 (0:00:01.042)       0:00:04.549 ****** \n\u001b[0;33mchanged: [testbed-node-0] => (item=nova_cell0)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=nova_api)\u001b[0m\n\nTASK [nova : Creating Nova databases user and setting permissions] *************\nTuesday 24 February 2026  01:08:06 +0000 (0:00:04.784)       0:00:09.333 ****** \n\u001b[0;33mchanged: [testbed-node-0] => (item=None)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=None)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [nova : Ensuring config directories exist] ********************************\nTuesday 24 February 2026  01:08:11 +0000 (0:00:04.789)       0:00:14.123 ****** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [nova : Copying over config.json files for nova-api-bootstrap] ************\nTuesday 24 February 2026  01:08:12 +0000 (0:00:01.198)       0:00:15.322 ****** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [nova : Copying over nova.conf for nova-api-bootstrap] ********************\nTuesday 24 February 2026  01:08:14 +0000 (0:00:01.801)       0:00:17.123 ****** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [nova : include_tasks] ****************************************************\nTuesday 24 February 2026  01:08:18 +0000 (0:00:04.441)       0:00:21.564 ****** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [nova : Running Nova API bootstrap container] *****************************\nTuesday 24 February 2026  01:08:19 +0000 (0:00:00.463)       0:00:22.027 ****** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [nova : Create cell0 mappings] ********************************************\nTuesday 24 February 2026  01:08:54 +0000 (0:00:35.166)       0:00:57.194 ****** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [nova-cell : Get a list of existing cells] ********************************\nTuesday 24 February 2026  01:09:11 +0000 (0:00:17.691)       0:01:14.886 ****** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [nova-cell : Extract current cell settings from list] *********************\nTuesday 24 February 2026  01:09:26 +0000 (0:00:14.691)       0:01:29.577 ****** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [nova : Update cell0 mappings] ********************************************\nTuesday 24 February 2026  01:09:27 +0000 (0:00:00.979)       0:01:30.557 ****** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [nova : include_tasks] ****************************************************\nTuesday 24 February 2026  01:09:27 +0000 (0:00:00.422)       0:01:30.980 ****** \n\u001b[0;36mincluded: /ansible/roles/nova/tasks/bootstrap_service.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [nova : Running Nova API bootstrap container] *****************************\nTuesday 24 February 2026  01:09:28 +0000 (0:00:00.543)       0:01:31.523 ****** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [Bootstrap upgrade] *******************************************************\nTuesday 24 February 2026  01:09:51 +0000 (0:00:23.084)       0:01:54.607 ****** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nPLAY [Bootstrap nova cell databases] *******************************************\n\nTASK [Bootstrap deploy] ********************************************************\nTuesday 24 February 2026  01:09:52 +0000 (0:00:00.516)       0:01:55.123 ****** \n\u001b[0;36mincluded: nova-cell for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [nova-cell : Creating Nova cell database] *********************************\nTuesday 24 February 2026  01:09:53 +0000 (0:00:01.123)       0:01:56.247 ****** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [nova-cell : Creating Nova cell database user and setting permissions] ****\nTuesday 24 February 2026  01:09:55 +0000 (0:00:02.466)       0:01:58.713 ****** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [service-rabbitmq : nova | Ensure RabbitMQ vhosts exist] ******************\nTuesday 24 February 2026  01:09:58 +0000 (0:00:02.859)       0:02:01.573 ****** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [service-rabbitmq : nova | Ensure RabbitMQ users exist] *******************\nTuesday 24 February 2026  01:09:59 +0000 (0:00:00.542)       0:02:02.115 ****** \n\u001b[0;36mskipping: [testbed-node-2] => (item=None) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=None) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-0] => (item=None)\u001b[0m\n\u001b[0;32mok: [testbed-node-0 -> {{ service_rabbitmq_delegate_host }}]\u001b[0m\n\nTASK [service-rabbitmq : nova | Ensure RabbitMQ vhosts exist] ******************\nTuesday 24 February 2026  01:10:07 +0000 (0:00:08.820)       0:02:10.935 ****** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [service-rabbitmq : nova | Ensure RabbitMQ users exist] *******************\nTuesday 24 February 2026  01:10:08 +0000 (0:00:00.397)       0:02:11.333 ****** \n\u001b[0;36mskipping: [testbed-node-0] => (item=None) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=None) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=None) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [nova-cell : Ensuring config directories exist] ***************************\nTuesday 24 February 2026  01:10:08 +0000 (0:00:00.612)       0:02:11.946 ****** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [nova-cell : Copying over config.json files for nova-cell-bootstrap] ******\nTuesday 24 February 2026  01:10:09 +0000 (0:00:00.714)       0:02:12.660 ****** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [nova-cell : Copying over nova.conf for nova-cell-bootstrap] **************\nTuesday 24 February 2026  01:10:10 +0000 (0:00:01.001)       0:02:13.662 ****** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [nova-cell : Running Nova cell bootstrap container] ***********************\nTuesday 24 February 2026  01:10:13 +0000 (0:00:02.410)       0:02:16.073 ****** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [nova-cell : Get a list of existing cells] ********************************\nTuesday 24 February 2026  01:10:35 +0000 (0:00:22.161)       0:02:38.234 ****** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [nova-cell : Extract current cell settings from list] *********************\nTuesday 24 February 2026  01:10:48 +0000 (0:00:13.546)       0:02:51.780 ****** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [nova-cell : Create cell] *************************************************\nTuesday 24 February 2026  01:10:49 +0000 (0:00:00.963)       0:02:52.743 ****** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [nova-cell : Update cell] *************************************************\nTuesday 24 February 2026  01:11:04 +0000 (0:00:14.398)       0:03:07.141 ****** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [Bootstrap upgrade] *******************************************************\nTuesday 24 February 2026  01:11:05 +0000 (0:00:01.307)       0:03:08.449 ****** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nPLAY [Apply role nova] *********************************************************\n\nTASK [nova : include_tasks] ****************************************************\nTuesday 24 February 2026  01:11:06 +0000 (0:00:00.691)       0:03:09.141 ****** \n\u001b[0;36mincluded: /ansible/roles/nova/tasks/deploy.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [service-ks-register : nova | Creating services] **************************\nTuesday 24 February 2026  01:11:06 +0000 (0:00:00.637)       0:03:09.778 ****** \n\u001b[0;36mskipping: [testbed-node-0] => (item=nova_legacy (compute_legacy)) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=nova (compute))\u001b[0m\n\nTASK [service-ks-register : nova | Creating endpoints] *************************\nTuesday 24 February 2026  01:11:10 +0000 (0:00:03.678)       0:03:13.457 ****** \n\u001b[0;36mskipping: [testbed-node-0] => (item=nova_legacy -> https://api-int.testbed.osism.xyz:8774/v2/%(tenant_id)s -> internal) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=nova_legacy -> https://api.testbed.osism.xyz:8774/v2/%(tenant_id)s -> public) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=nova -> https://api-int.testbed.osism.xyz:8774/v2.1 -> internal)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=nova -> https://api.testbed.osism.xyz:8774/v2.1 -> public)\u001b[0m\n\nTASK [service-ks-register : nova | Creating projects] **************************\nTuesday 24 February 2026  01:11:17 +0000 (0:00:07.286)       0:03:20.743 ****** \n\u001b[0;32mok: [testbed-node-0] => (item=service)\u001b[0m\n\nTASK [service-ks-register : nova | Creating users] *****************************\nTuesday 24 February 2026  01:11:21 +0000 (0:00:03.290)       0:03:24.034 ****** \n\u001b[0;33mchanged: [testbed-node-0] => (item=nova -> service)\u001b[0m\n\u001b[1;35m[WARNING]: Module did not set no_log for update_password\u001b[0m\n\nTASK [service-ks-register : nova | Creating roles] *****************************\nTuesday 24 February 2026  01:11:24 +0000 (0:00:03.880)       0:03:27.915 ****** \n\u001b[0;32mok: [testbed-node-0] => (item=admin)\u001b[0m\n\nTASK [service-ks-register : nova | Granting user roles] ************************\nTuesday 24 February 2026  01:11:28 +0000 (0:00:03.797)       0:03:31.712 ****** \n\u001b[0;33mchanged: [testbed-node-0] => (item=nova -> service -> admin)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=nova -> service -> service)\u001b[0m\n\nTASK [nova : Ensuring config directories exist] ********************************\nTuesday 24 February 2026  01:11:37 +0000 (0:00:08.533)       0:03:40.246 ****** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/nova-api:2024.2', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8774 '], 'timeout': '30'}, 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/nova-api:2024.2', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8774 '], 'timeout': '30'}, 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/nova-api:2024.2', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8774 '], 'timeout': '30'}, 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/nova-scheduler:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/nova-scheduler:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/nova-scheduler:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}})\u001b[0m\n\nTASK [nova : Check if policies shall be overwritten] ***************************\nTuesday 24 February 2026  01:11:38 +0000 (0:00:01.522)       0:03:41.768 ****** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [nova : Set nova policy file] *********************************************\nTuesday 24 February 2026  01:11:38 +0000 (0:00:00.135)       0:03:41.904 ****** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [nova : Check for vendordata file] ****************************************\nTuesday 24 February 2026  01:11:39 +0000 (0:00:00.338)       0:03:42.242 ****** \n\u001b[0;32mok: [testbed-node-0 -> localhost]\u001b[0m\n\nTASK [nova : Set vendordata file path] *****************************************\nTuesday 24 February 2026  01:11:40 +0000 (0:00:01.181)       0:03:43.424 ****** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [nova : include_tasks] ****************************************************\nTuesday 24 February 2026  01:11:40 +0000 (0:00:00.352)       0:03:43.776 ****** \n\u001b[0;36mincluded: /ansible/roles/nova/tasks/copy-certs.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [service-cert-copy : nova | Copying over extra CA certificates] ***********\nTuesday 24 February 2026  01:11:41 +0000 (0:00:00.613)       0:03:44.390 ****** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/nova-api:2024.2', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8774 '], 'timeout': '30'}, 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/nova-api:2024.2', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8774 '], 'timeout': '30'}, 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/nova-api:2024.2', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8774 '], 'timeout': '30'}, 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/nova-scheduler:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/nova-scheduler:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/nova-scheduler:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}})\u001b[0m\n\nTASK [service-cert-copy : nova | Copying over backend internal TLS certificate] ***\nTuesday 24 February 2026  01:11:44 +0000 (0:00:03.219)       0:03:47.609 ****** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/nova-api:2024.2', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8774 '], 'timeout': '30'}, 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/nova-scheduler:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/nova-api:2024.2', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8774 '], 'timeout': '30'}, 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/nova-scheduler:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/nova-api:2024.2', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8774 '], 'timeout': '30'}, 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/nova-scheduler:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [service-cert-copy : nova | Copying over backend internal TLS key] ********\nTuesday 24 February 2026  01:11:45 +0000 (0:00:00.685)       0:03:48.295 ****** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/nova-api:2024.2', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8774 '], 'timeout': '30'}, 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/nova-scheduler:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/nova-api:2024.2', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8774 '], 'timeout': '30'}, 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/nova-scheduler:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/nova-api:2024.2', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8774 '], 'timeout': '30'}, 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/nova-scheduler:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [nova : Copying over config.json files for services] **********************\nTuesday 24 February 2026  01:11:46 +0000 (0:00:00.966)       0:03:49.261 ****** \n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/nova-api:2024.2', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8774 '], 'timeout': '30'}, 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/nova-api:2024.2', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8774 '], 'timeout': '30'}, 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/nova-api:2024.2', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8774 '], 'timeout': '30'}, 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/nova-scheduler:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/nova-scheduler:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/nova-scheduler:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}})\u001b[0m\n\nTASK [nova : Copying over nova.conf] *******************************************\nTuesday 24 February 2026  01:11:49 +0000 (0:00:02.883)       0:03:52.145 ****** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/nova-api:2024.2', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8774 '], 'timeout': '30'}, 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/nova-api:2024.2', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8774 '], 'timeout': '30'}, 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/nova-api:2024.2', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8774 '], 'timeout': '30'}, 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/nova-scheduler:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/nova-scheduler:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/nova-scheduler:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}})\u001b[0m\n\nTASK [nova : Copying over existing policy file] ********************************\nTuesday 24 February 2026  01:11:55 +0000 (0:00:06.354)       0:03:58.499 ****** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/nova-api:2024.2', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8774 '], 'timeout': '30'}, 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/nova-scheduler:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/nova-api:2024.2', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8774 '], 'timeout': '30'}, 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/nova-scheduler:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/nova-api:2024.2', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8774 '], 'timeout': '30'}, 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/nova-scheduler:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [nova : Copying over nova-api-wsgi.conf] **********************************\nTuesday 24 February 2026  01:11:56 +0000 (0:00:00.635)       0:03:59.134 ****** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [nova : Copying over vendordata file] *************************************\nTuesday 24 February 2026  01:11:57 +0000 (0:00:01.659)       0:04:00.794 ****** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [nova : Check nova containers] ********************************************\nTuesday 24 February 2026  01:11:58 +0000 (0:00:00.327)       0:04:01.121 ****** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/nova-api:2024.2', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8774 '], 'timeout': '30'}, 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/nova-api:2024.2', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8774 '], 'timeout': '30'}, 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/nova-api:2024.2', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8774 '], 'timeout': '30'}, 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/nova-scheduler:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/nova-scheduler:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/nova-scheduler:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}})\u001b[0m\n\nTASK [nova : Flush handlers] ***************************************************\nTuesday 24 February 2026  01:12:00 +0000 (0:00:01.987)       0:04:03.108 ****** \n\nTASK [nova : Flush handlers] ***************************************************\nTuesday 24 February 2026  01:12:00 +0000 (0:00:00.134)       0:04:03.243 ****** \n\nTASK [nova : Flush handlers] ***************************************************\nTuesday 24 February 2026  01:12:00 +0000 (0:00:00.123)       0:04:03.366 ****** \n\nRUNNING HANDLER [nova : Restart nova-scheduler container] **********************\nTuesday 24 February 2026  01:12:00 +0000 (0:00:00.124)       0:04:03.491 ****** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [nova : Restart nova-api container] ****************************\nTuesday 24 February 2026  01:12:23 +0000 (0:00:22.540)       0:04:26.032 ****** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\nPLAY [Apply role nova-cell] ****************************************************\n\nTASK [nova-cell : include_tasks] ***********************************************\nTuesday 24 February 2026  01:12:29 +0000 (0:00:06.843)       0:04:32.875 ****** \n\u001b[0;36mincluded: /ansible/roles/nova-cell/tasks/deploy.yml for testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [nova-cell : include_tasks] ***********************************************\nTuesday 24 February 2026  01:12:31 +0000 (0:00:01.427)       0:04:34.303 ****** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [Load and persist br_netfilter module] ************************************\nTuesday 24 February 2026  01:12:31 +0000 (0:00:00.686)       0:04:34.989 ****** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mincluded: module-load for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m\n\nTASK [module-load : Load modules] **********************************************\nTuesday 24 February 2026  01:12:33 +0000 (0:00:01.138)       0:04:36.128 ****** \n\u001b[0;32mok: [testbed-node-3] => (item=br_netfilter)\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item=br_netfilter)\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item=br_netfilter)\u001b[0m\n\nTASK [module-load : Persist modules via modules-load.d] ************************\nTuesday 24 February 2026  01:12:33 +0000 (0:00:00.751)       0:04:36.879 ****** \n\u001b[0;33mchanged: [testbed-node-4] => (item=br_netfilter)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item=br_netfilter)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=br_netfilter)\u001b[0m\n\nTASK [module-load : Drop module persistence] ***********************************\nTuesday 24 February 2026  01:12:35 +0000 (0:00:01.655)       0:04:38.534 ****** \n\u001b[0;36mskipping: [testbed-node-3] => (item=br_netfilter) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=br_netfilter) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=br_netfilter) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [nova-cell : Enable bridge-nf-call sysctl variables] **********************\nTuesday 24 February 2026  01:12:36 +0000 (0:00:00.628)       0:04:39.163 ****** \n\u001b[0;36mskipping: [testbed-node-0] => (item=net.bridge.bridge-nf-call-iptables) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=net.bridge.bridge-nf-call-ip6tables) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=net.bridge.bridge-nf-call-iptables) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=net.bridge.bridge-nf-call-ip6tables) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=net.bridge.bridge-nf-call-iptables)\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item=net.bridge.bridge-nf-call-iptables)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=net.bridge.bridge-nf-call-iptables)\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=net.bridge.bridge-nf-call-iptables) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=net.bridge.bridge-nf-call-ip6tables) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=net.bridge.bridge-nf-call-ip6tables)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=net.bridge.bridge-nf-call-ip6tables)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item=net.bridge.bridge-nf-call-ip6tables)\u001b[0m\n\nTASK [nova-cell : Install udev kolla kvm rules] ********************************\nTuesday 24 February 2026  01:12:37 +0000 (0:00:01.306)       0:04:40.470 ****** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nTASK [nova-cell : Mask qemu-kvm service] ***************************************\nTuesday 24 February 2026  01:12:38 +0000 (0:00:01.329)       0:04:41.799 ****** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\nTASK [nova-cell : Ensuring config directories exist] ***************************\nTuesday 24 February 2026  01:12:40 +0000 (0:00:02.076)       0:04:43.876 ****** \n\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'nova-libvirt', 'value': {'container_name': 'nova_libvirt', 'group': 'compute', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-libvirt:2024.2', 'pid_mode': 'host', 'cgroupns_mode': 'host', 'privileged': True, 'volumes': ['/etc/kolla/nova-libvirt/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', '', '/sys/fs/cgroup:/sys/fs/cgroup', 'kolla_logs:/var/log/kolla/', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', 'nova_libvirt_qemu:/etc/libvirt/qemu', ''], 'dimensions': {'ulimits': {'memlock': {'soft': 67108864, 'hard': 67108864}}}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'virsh version --daemon'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'nova-libvirt', 'value': {'container_name': 'nova_libvirt', 'group': 'compute', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-libvirt:2024.2', 'pid_mode': 'host', 'cgroupns_mode': 'host', 'privileged': True, 'volumes': ['/etc/kolla/nova-libvirt/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', '', '/sys/fs/cgroup:/sys/fs/cgroup', 'kolla_logs:/var/log/kolla/', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', 'nova_libvirt_qemu:/etc/libvirt/qemu', ''], 'dimensions': {'ulimits': {'memlock': {'soft': 67108864, 'hard': 67108864}}}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'virsh version --daemon'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'nova-ssh', 'value': {'container_name': 'nova_ssh', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-ssh:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla', 'nova_compute:/var/lib/nova', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8022'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'nova-libvirt', 'value': {'container_name': 'nova_libvirt', 'group': 'compute', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-libvirt:2024.2', 'pid_mode': 'host', 'cgroupns_mode': 'host', 'privileged': True, 'volumes': ['/etc/kolla/nova-libvirt/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', '', '/sys/fs/cgroup:/sys/fs/cgroup', 'kolla_logs:/var/log/kolla/', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', 'nova_libvirt_qemu:/etc/libvirt/qemu', ''], 'dimensions': {'ulimits': {'memlock': {'soft': 67108864, 'hard': 67108864}}}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'virsh version --daemon'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'nova-novncproxy', 'value': {'container_name': 'nova_novncproxy', 'group': 'nova-novncproxy', 'image': 'registry.osism.tech/kolla/nova-novncproxy:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-novncproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:6080/vnc_lite.html'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'nova-ssh', 'value': {'container_name': 'nova_ssh', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-ssh:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla', 'nova_compute:/var/lib/nova', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8022'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'nova-ssh', 'value': {'container_name': 'nova_ssh', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-ssh:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla', 'nova_compute:/var/lib/nova', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8022'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'nova-compute', 'value': {'container_name': 'nova_compute', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-compute:2024.2', 'environment': {'LIBGUESTFS_BACKEND': 'direct'}, 'privileged': True, 'enabled': True, 'ipc_mode': 'host', 'volumes': ['/etc/kolla/nova-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', 'kolla_logs:/var/log/kolla/', '', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-compute 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'nova-novncproxy', 'value': {'container_name': 'nova_novncproxy', 'group': 'nova-novncproxy', 'image': 'registry.osism.tech/kolla/nova-novncproxy:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-novncproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:6080/vnc_lite.html'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'nova-novncproxy', 'value': {'container_name': 'nova_novncproxy', 'group': 'nova-novncproxy', 'image': 'registry.osism.tech/kolla/nova-novncproxy:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-novncproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:6080/vnc_lite.html'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'nova-compute', 'value': {'container_name': 'nova_compute', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-compute:2024.2', 'environment': {'LIBGUESTFS_BACKEND': 'direct'}, 'privileged': True, 'enabled': True, 'ipc_mode': 'host', 'volumes': ['/etc/kolla/nova-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', 'kolla_logs:/var/log/kolla/', '', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-compute 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'nova-conductor', 'value': {'container_name': 'nova_conductor', 'group': 'nova-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-conductor:2024.2', 'volumes': ['/etc/kolla/nova-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'nova-compute', 'value': {'container_name': 'nova_compute', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-compute:2024.2', 'environment': {'LIBGUESTFS_BACKEND': 'direct'}, 'privileged': True, 'enabled': True, 'ipc_mode': 'host', 'volumes': ['/etc/kolla/nova-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', 'kolla_logs:/var/log/kolla/', '', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-compute 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'nova-conductor', 'value': {'container_name': 'nova_conductor', 'group': 'nova-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-conductor:2024.2', 'volumes': ['/etc/kolla/nova-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'nova-conductor', 'value': {'container_name': 'nova_conductor', 'group': 'nova-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-conductor:2024.2', 'volumes': ['/etc/kolla/nova-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}})\u001b[0m\n\nTASK [nova-cell : include_tasks] ***********************************************\nTuesday 24 February 2026  01:12:43 +0000 (0:00:02.268)       0:04:46.144 ****** \n\u001b[0;36mincluded: /ansible/roles/nova-cell/tasks/copy-certs.yml for testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [service-cert-copy : nova | Copying over extra CA certificates] ***********\nTuesday 24 February 2026  01:12:44 +0000 (0:00:01.361)       0:04:47.505 ****** \n\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'nova-libvirt', 'value': {'container_name': 'nova_libvirt', 'group': 'compute', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-libvirt:2024.2', 'pid_mode': 'host', 'cgroupns_mode': 'host', 'privileged': True, 'volumes': ['/etc/kolla/nova-libvirt/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', '', '/sys/fs/cgroup:/sys/fs/cgroup', 'kolla_logs:/var/log/kolla/', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', 'nova_libvirt_qemu:/etc/libvirt/qemu', ''], 'dimensions': {'ulimits': {'memlock': {'soft': 67108864, 'hard': 67108864}}}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'virsh version --daemon'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'nova-libvirt', 'value': {'container_name': 'nova_libvirt', 'group': 'compute', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-libvirt:2024.2', 'pid_mode': 'host', 'cgroupns_mode': 'host', 'privileged': True, 'volumes': ['/etc/kolla/nova-libvirt/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', '', '/sys/fs/cgroup:/sys/fs/cgroup', 'kolla_logs:/var/log/kolla/', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', 'nova_libvirt_qemu:/etc/libvirt/qemu', ''], 'dimensions': {'ulimits': {'memlock': {'soft': 67108864, 'hard': 67108864}}}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'virsh version --daemon'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'nova-libvirt', 'value': {'container_name': 'nova_libvirt', 'group': 'compute', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-libvirt:2024.2', 'pid_mode': 'host', 'cgroupns_mode': 'host', 'privileged': True, 'volumes': ['/etc/kolla/nova-libvirt/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', '', '/sys/fs/cgroup:/sys/fs/cgroup', 'kolla_logs:/var/log/kolla/', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', 'nova_libvirt_qemu:/etc/libvirt/qemu', ''], 'dimensions': {'ulimits': {'memlock': {'soft': 67108864, 'hard': 67108864}}}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'virsh version --daemon'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'nova-novncproxy', 'value': {'container_name': 'nova_novncproxy', 'group': 'nova-novncproxy', 'image': 'registry.osism.tech/kolla/nova-novncproxy:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-novncproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:6080/vnc_lite.html'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'nova-ssh', 'value': {'container_name': 'nova_ssh', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-ssh:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla', 'nova_compute:/var/lib/nova', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8022'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'nova-novncproxy', 'value': {'container_name': 'nova_novncproxy', 'group': 'nova-novncproxy', 'image': 'registry.osism.tech/kolla/nova-novncproxy:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-novncproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:6080/vnc_lite.html'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'nova-novncproxy', 'value': {'container_name': 'nova_novncproxy', 'group': 'nova-novncproxy', 'image': 'registry.osism.tech/kolla/nova-novncproxy:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-novncproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:6080/vnc_lite.html'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'nova-ssh', 'value': {'container_name': 'nova_ssh', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-ssh:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla', 'nova_compute:/var/lib/nova', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8022'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'nova-ssh', 'value': {'container_name': 'nova_ssh', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-ssh:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla', 'nova_compute:/var/lib/nova', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8022'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'nova-conductor', 'value': {'container_name': 'nova_conductor', 'group': 'nova-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-conductor:2024.2', 'volumes': ['/etc/kolla/nova-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'nova-compute', 'value': {'container_name': 'nova_compute', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-compute:2024.2', 'environment': {'LIBGUESTFS_BACKEND': 'direct'}, 'privileged': True, 'enabled': True, 'ipc_mode': 'host', 'volumes': ['/etc/kolla/nova-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', 'kolla_logs:/var/log/kolla/', '', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-compute 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'nova-conductor', 'value': {'container_name': 'nova_conductor', 'group': 'nova-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-conductor:2024.2', 'volumes': ['/etc/kolla/nova-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'nova-conductor', 'value': {'container_name': 'nova_conductor', 'group': 'nova-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-conductor:2024.2', 'volumes': ['/etc/kolla/nova-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'nova-compute', 'value': {'container_name': 'nova_compute', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-compute:2024.2', 'environment': {'LIBGUESTFS_BACKEND': 'direct'}, 'privileged': True, 'enabled': True, 'ipc_mode': 'host', 'volumes': ['/etc/kolla/nova-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', 'kolla_logs:/var/log/kolla/', '', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-compute 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'nova-compute', 'value': {'container_name': 'nova_compute', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-compute:2024.2', 'environment': {'LIBGUESTFS_BACKEND': 'direct'}, 'privileged': True, 'enabled': True, 'ipc_mode': 'host', 'volumes': ['/etc/kolla/nova-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', 'kolla_logs:/var/log/kolla/', '', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-compute 5672'], 'timeout': '30'}}})\u001b[0m\n\nTASK [service-cert-copy : nova | Copying over backend internal TLS certificate] ***\nTuesday 24 February 2026  01:12:48 +0000 (0:00:03.878)       0:04:51.383 ****** \n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'nova-libvirt', 'value': {'container_name': 'nova_libvirt', 'group': 'compute', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-libvirt:2024.2', 'pid_mode': 'host', 'cgroupns_mode': 'host', 'privileged': True, 'volumes': ['/etc/kolla/nova-libvirt/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', '', '/sys/fs/cgroup:/sys/fs/cgroup', 'kolla_logs:/var/log/kolla/', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', 'nova_libvirt_qemu:/etc/libvirt/qemu', ''], 'dimensions': {'ulimits': {'memlock': {'soft': 67108864, 'hard': 67108864}}}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'virsh version --daemon'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'nova-ssh', 'value': {'container_name': 'nova_ssh', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-ssh:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla', 'nova_compute:/var/lib/nova', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8022'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'nova-compute', 'value': {'container_name': 'nova_compute', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-compute:2024.2', 'environment': {'LIBGUESTFS_BACKEND': 'direct'}, 'privileged': True, 'enabled': True, 'ipc_mode': 'host', 'volumes': ['/etc/kolla/nova-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', 'kolla_logs:/var/log/kolla/', '', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-compute 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'nova-libvirt', 'value': {'container_name': 'nova_libvirt', 'group': 'compute', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-libvirt:2024.2', 'pid_mode': 'host', 'cgroupns_mode': 'host', 'privileged': True, 'volumes': ['/etc/kolla/nova-libvirt/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', '', '/sys/fs/cgroup:/sys/fs/cgroup', 'kolla_logs:/var/log/kolla/', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', 'nova_libvirt_qemu:/etc/libvirt/qemu', ''], 'dimensions': {'ulimits': {'memlock': {'soft': 67108864, 'hard': 67108864}}}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'virsh version --daemon'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'nova-ssh', 'value': {'container_name': 'nova_ssh', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-ssh:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla', 'nova_compute:/var/lib/nova', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8022'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'nova-compute', 'value': {'container_name': 'nova_compute', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-compute:2024.2', 'environment': {'LIBGUESTFS_BACKEND': 'direct'}, 'privileged': True, 'enabled': True, 'ipc_mode': 'host', 'volumes': ['/etc/kolla/nova-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', 'kolla_logs:/var/log/kolla/', '', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-compute 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'nova-libvirt', 'value': {'container_name': 'nova_libvirt', 'group': 'compute', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-libvirt:2024.2', 'pid_mode': 'host', 'cgroupns_mode': 'host', 'privileged': True, 'volumes': ['/etc/kolla/nova-libvirt/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', '', '/sys/fs/cgroup:/sys/fs/cgroup', 'kolla_logs:/var/log/kolla/', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', 'nova_libvirt_qemu:/etc/libvirt/qemu', ''], 'dimensions': {'ulimits': {'memlock': {'soft': 67108864, 'hard': 67108864}}}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'virsh version --daemon'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'nova-ssh', 'value': {'container_name': 'nova_ssh', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-ssh:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla', 'nova_compute:/var/lib/nova', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8022'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'nova-compute', 'value': {'container_name': 'nova_compute', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-compute:2024.2', 'environment': {'LIBGUESTFS_BACKEND': 'direct'}, 'privileged': True, 'enabled': True, 'ipc_mode': 'host', 'volumes': ['/etc/kolla/nova-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', 'kolla_logs:/var/log/kolla/', '', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-compute 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova-novncproxy', 'value': {'container_name': 'nova_novncproxy', 'group': 'nova-novncproxy', 'image': 'registry.osism.tech/kolla/nova-novncproxy:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-novncproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:6080/vnc_lite.html'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova-conductor', 'value': {'container_name': 'nova_conductor', 'group': 'nova-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-conductor:2024.2', 'volumes': ['/etc/kolla/nova-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova-novncproxy', 'value': {'container_name': 'nova_novncproxy', 'group': 'nova-novncproxy', 'image': 'registry.osism.tech/kolla/nova-novncproxy:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-novncproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:6080/vnc_lite.html'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova-conductor', 'value': {'container_name': 'nova_conductor', 'group': 'nova-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-conductor:2024.2', 'volumes': ['/etc/kolla/nova-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova-novncproxy', 'value': {'container_name': 'nova_novncproxy', 'group': 'nova-novncproxy', 'image': 'registry.osism.tech/kolla/nova-novncproxy:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-novncproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:6080/vnc_lite.html'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova-conductor', 'value': {'container_name': 'nova_conductor', 'group': 'nova-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-conductor:2024.2', 'volumes': ['/etc/kolla/nova-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [service-cert-copy : nova | Copying over backend internal TLS key] ********\nTuesday 24 February 2026  01:12:50 +0000 (0:00:01.752)       0:04:53.136 ****** \n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'nova-libvirt', 'value': {'container_name': 'nova_libvirt', 'group': 'compute', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-libvirt:2024.2', 'pid_mode': 'host', 'cgroupns_mode': 'host', 'privileged': True, 'volumes': ['/etc/kolla/nova-libvirt/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', '', '/sys/fs/cgroup:/sys/fs/cgroup', 'kolla_logs:/var/log/kolla/', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', 'nova_libvirt_qemu:/etc/libvirt/qemu', ''], 'dimensions': {'ulimits': {'memlock': {'soft': 67108864, 'hard': 67108864}}}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'virsh version --daemon'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'nova-ssh', 'value': {'container_name': 'nova_ssh', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-ssh:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla', 'nova_compute:/var/lib/nova', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8022'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'nova-compute', 'value': {'container_name': 'nova_compute', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-compute:2024.2', 'environment': {'LIBGUESTFS_BACKEND': 'direct'}, 'privileged': True, 'enabled': True, 'ipc_mode': 'host', 'volumes': ['/etc/kolla/nova-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', 'kolla_logs:/var/log/kolla/', '', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-compute 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'nova-libvirt', 'value': {'container_name': 'nova_libvirt', 'group': 'compute', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-libvirt:2024.2', 'pid_mode': 'host', 'cgroupns_mode': 'host', 'privileged': True, 'volumes': ['/etc/kolla/nova-libvirt/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', '', '/sys/fs/cgroup:/sys/fs/cgroup', 'kolla_logs:/var/log/kolla/', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', 'nova_libvirt_qemu:/etc/libvirt/qemu', ''], 'dimensions': {'ulimits': {'memlock': {'soft': 67108864, 'hard': 67108864}}}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'virsh version --daemon'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'nova-ssh', 'value': {'container_name': 'nova_ssh', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-ssh:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla', 'nova_compute:/var/lib/nova', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8022'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'nova-compute', 'value': {'container_name': 'nova_compute', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-compute:2024.2', 'environment': {'LIBGUESTFS_BACKEND': 'direct'}, 'privileged': True, 'enabled': True, 'ipc_mode': 'host', 'volumes': ['/etc/kolla/nova-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', 'kolla_logs:/var/log/kolla/', '', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-compute 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'nova-libvirt', 'value': {'container_name': 'nova_libvirt', 'group': 'compute', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-libvirt:2024.2', 'pid_mode': 'host', 'cgroupns_mode': 'host', 'privileged': True, 'volumes': ['/etc/kolla/nova-libvirt/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', '', '/sys/fs/cgroup:/sys/fs/cgroup', 'kolla_logs:/var/log/kolla/', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', 'nova_libvirt_qemu:/etc/libvirt/qemu', ''], 'dimensions': {'ulimits': {'memlock': {'soft': 67108864, 'hard': 67108864}}}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'virsh version --daemon'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'nova-ssh', 'value': {'container_name': 'nova_ssh', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-ssh:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla', 'nova_compute:/var/lib/nova', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8022'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova-novncproxy', 'value': {'container_name': 'nova_novncproxy', 'group': 'nova-novncproxy', 'image': 'registry.osism.tech/kolla/nova-novncproxy:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-novncproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:6080/vnc_lite.html'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova-conductor', 'value': {'container_name': 'nova_conductor', 'group': 'nova-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-conductor:2024.2', 'volumes': ['/etc/kolla/nova-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'nova-compute', 'value': {'container_name': 'nova_compute', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-compute:2024.2', 'environment': {'LIBGUESTFS_BACKEND': 'direct'}, 'privileged': True, 'enabled': True, 'ipc_mode': 'host', 'volumes': ['/etc/kolla/nova-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', 'kolla_logs:/var/log/kolla/', '', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-compute 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova-novncproxy', 'value': {'container_name': 'nova_novncproxy', 'group': 'nova-novncproxy', 'image': 'registry.osism.tech/kolla/nova-novncproxy:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-novncproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:6080/vnc_lite.html'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova-conductor', 'value': {'container_name': 'nova_conductor', 'group': 'nova-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-conductor:2024.2', 'volumes': ['/etc/kolla/nova-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova-novncproxy', 'value': {'container_name': 'nova_novncproxy', 'group': 'nova-novncproxy', 'image': 'registry.osism.tech/kolla/nova-novncproxy:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-novncproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:6080/vnc_lite.html'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova-conductor', 'value': {'container_name': 'nova_conductor', 'group': 'nova-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-conductor:2024.2', 'volumes': ['/etc/kolla/nova-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [nova-cell : include_tasks] ***********************************************\nTuesday 24 February 2026  01:12:52 +0000 (0:00:02.331)       0:04:55.467 ****** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mincluded: /ansible/roles/nova-cell/tasks/external_ceph.yml for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m\n\nTASK [nova-cell : Check nova keyring file] *************************************\nTuesday 24 February 2026  01:12:53 +0000 (0:00:01.241)       0:04:56.709 ****** \n\u001b[0;32mok: [testbed-node-3 -> localhost]\u001b[0m\n\u001b[0;32mok: [testbed-node-4 -> localhost]\u001b[0m\n\u001b[0;32mok: [testbed-node-5 -> localhost]\u001b[0m\n\nTASK [nova-cell : Check cinder keyring file] ***********************************\nTuesday 24 February 2026  01:12:54 +0000 (0:00:01.242)       0:04:57.951 ****** \n\u001b[0;32mok: [testbed-node-3 -> localhost]\u001b[0m\n\u001b[0;32mok: [testbed-node-4 -> localhost]\u001b[0m\n\u001b[0;32mok: [testbed-node-5 -> localhost]\u001b[0m\n\nTASK [nova-cell : Extract nova key from file] **********************************\nTuesday 24 February 2026  01:12:56 +0000 (0:00:01.241)       0:04:59.192 ****** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [nova-cell : Extract cinder key from file] ********************************\nTuesday 24 February 2026  01:12:56 +0000 (0:00:00.555)       0:04:59.748 ****** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [nova-cell : Copy over ceph nova keyring file] ****************************\nTuesday 24 February 2026  01:12:57 +0000 (0:00:00.897)       0:05:00.645 ****** \n\u001b[0;33mchanged: [testbed-node-3] => (item=nova-compute)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=nova-compute)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=nova-compute)\u001b[0m\n\nTASK [nova-cell : Copy over ceph cinder keyring file] **************************\nTuesday 24 February 2026  01:12:58 +0000 (0:00:01.291)       0:05:01.937 ****** \n\u001b[0;33mchanged: [testbed-node-3] => (item=nova-compute)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=nova-compute)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=nova-compute)\u001b[0m\n\nTASK [nova-cell : Copy over ceph.conf] *****************************************\nTuesday 24 February 2026  01:13:00 +0000 (0:00:01.244)       0:05:03.182 ****** \n\u001b[0;33mchanged: [testbed-node-3] => (item=nova-compute)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=nova-compute)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=nova-compute)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item=nova-libvirt)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=nova-libvirt)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=nova-libvirt)\u001b[0m\n\nTASK [nova-cell : Ensure /etc/ceph directory exists (host libvirt)] ************\nTuesday 24 February 2026  01:13:04 +0000 (0:00:04.219)       0:05:07.401 ****** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [nova-cell : Copy over ceph.conf (host libvirt)] **************************\nTuesday 24 February 2026  01:13:05 +0000 (0:00:00.669)       0:05:08.071 ****** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [nova-cell : Ensuring libvirt secrets directory exists] *******************\nTuesday 24 February 2026  01:13:05 +0000 (0:00:00.372)       0:05:08.443 ****** \n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nTASK [nova-cell : Pushing nova secret xml for libvirt] *************************\nTuesday 24 February 2026  01:13:06 +0000 (0:00:01.365)       0:05:09.809 ****** \n\u001b[0;33mchanged: [testbed-node-3] => (item={'uuid': '5a2bf0bf-e1ab-4a6a-bc32-404bb6ba91fd', 'name': 'client.nova secret', 'enabled': True})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'uuid': '5a2bf0bf-e1ab-4a6a-bc32-404bb6ba91fd', 'name': 'client.nova secret', 'enabled': True})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'uuid': '5a2bf0bf-e1ab-4a6a-bc32-404bb6ba91fd', 'name': 'client.nova secret', 'enabled': True})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'uuid': '63dd366f-e403-41f2-beff-dad9980a1637', 'name': 'client.cinder secret', 'enabled': 'yes'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'uuid': '63dd366f-e403-41f2-beff-dad9980a1637', 'name': 'client.cinder secret', 'enabled': 'yes'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'uuid': '63dd366f-e403-41f2-beff-dad9980a1637', 'name': 'client.cinder secret', 'enabled': 'yes'})\u001b[0m\n\nTASK [nova-cell : Pushing secrets key for libvirt] *****************************\nTuesday 24 February 2026  01:13:10 +0000 (0:00:03.628)       0:05:13.438 ****** \n\u001b[0;33mchanged: [testbed-node-3] => (item=None)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=None)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=None)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item=None)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item=None)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item=None)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nTASK [nova-cell : Include tasks from qemu_wrapper.yml] *************************\nTuesday 24 February 2026  01:13:14 +0000 (0:00:03.975)       0:05:17.413 ****** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mincluded: /ansible/roles/nova-cell/tasks/qemu_wrapper.yml for testbed-node-3, testbed-node-5, testbed-node-4\u001b[0m\n\nTASK [nova-cell : Check qemu wrapper file] *************************************\nTuesday 24 February 2026  01:13:16 +0000 (0:00:01.687)       0:05:19.100 ****** \n\u001b[0;32mok: [testbed-node-3 -> localhost]\u001b[0m\n\u001b[0;32mok: [testbed-node-4 -> localhost]\u001b[0m\n\u001b[0;32mok: [testbed-node-5 -> localhost]\u001b[0m\n\nTASK [nova-cell : Copy qemu wrapper] *******************************************\nTuesday 24 February 2026  01:13:17 +0000 (0:00:01.379)       0:05:20.479 ****** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [nova-cell : Check if policies shall be overwritten] **********************\nTuesday 24 February 2026  01:13:17 +0000 (0:00:00.340)       0:05:20.820 ****** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [nova-cell : Set nova policy file] ****************************************\nTuesday 24 February 2026  01:13:17 +0000 (0:00:00.155)       0:05:20.976 ****** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [nova-cell : Check for vendordata file] ***********************************\nTuesday 24 February 2026  01:13:18 +0000 (0:00:00.618)       0:05:21.594 ****** \n\u001b[0;32mok: [testbed-node-3 -> localhost]\u001b[0m\n\nTASK [nova-cell : Set vendordata file path] ************************************\nTuesday 24 February 2026  01:13:19 +0000 (0:00:01.121)       0:05:22.716 ****** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [nova-cell : Copying over config.json files for services] *****************\nTuesday 24 February 2026  01:13:20 +0000 (0:00:00.714)       0:05:23.430 ****** \n\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'nova-libvirt', 'value': {'container_name': 'nova_libvirt', 'group': 'compute', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-libvirt:2024.2', 'pid_mode': 'host', 'cgroupns_mode': 'host', 'privileged': True, 'volumes': ['/etc/kolla/nova-libvirt/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', '', '/sys/fs/cgroup:/sys/fs/cgroup', 'kolla_logs:/var/log/kolla/', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', 'nova_libvirt_qemu:/etc/libvirt/qemu', ''], 'dimensions': {'ulimits': {'memlock': {'soft': 67108864, 'hard': 67108864}}}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'virsh version --daemon'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'nova-libvirt', 'value': {'container_name': 'nova_libvirt', 'group': 'compute', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-libvirt:2024.2', 'pid_mode': 'host', 'cgroupns_mode': 'host', 'privileged': True, 'volumes': ['/etc/kolla/nova-libvirt/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', '', '/sys/fs/cgroup:/sys/fs/cgroup', 'kolla_logs:/var/log/kolla/', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', 'nova_libvirt_qemu:/etc/libvirt/qemu', ''], 'dimensions': {'ulimits': {'memlock': {'soft': 67108864, 'hard': 67108864}}}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'virsh version --daemon'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'nova-libvirt', 'value': {'container_name': 'nova_libvirt', 'group': 'compute', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-libvirt:2024.2', 'pid_mode': 'host', 'cgroupns_mode': 'host', 'privileged': True, 'volumes': ['/etc/kolla/nova-libvirt/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', '', '/sys/fs/cgroup:/sys/fs/cgroup', 'kolla_logs:/var/log/kolla/', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', 'nova_libvirt_qemu:/etc/libvirt/qemu', ''], 'dimensions': {'ulimits': {'memlock': {'soft': 67108864, 'hard': 67108864}}}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'virsh version --daemon'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'nova-novncproxy', 'value': {'container_name': 'nova_novncproxy', 'group': 'nova-novncproxy', 'image': 'registry.osism.tech/kolla/nova-novncproxy:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-novncproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:6080/vnc_lite.html'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'nova-novncproxy', 'value': {'container_name': 'nova_novncproxy', 'group': 'nova-novncproxy', 'image': 'registry.osism.tech/kolla/nova-novncproxy:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-novncproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:6080/vnc_lite.html'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'nova-ssh', 'value': {'container_name': 'nova_ssh', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-ssh:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla', 'nova_compute:/var/lib/nova', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8022'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'nova-novncproxy', 'value': {'container_name': 'nova_novncproxy', 'group': 'nova-novncproxy', 'image': 'registry.osism.tech/kolla/nova-novncproxy:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-novncproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:6080/vnc_lite.html'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'nova-ssh', 'value': {'container_name': 'nova_ssh', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-ssh:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla', 'nova_compute:/var/lib/nova', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8022'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'nova-ssh', 'value': {'container_name': 'nova_ssh', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-ssh:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla', 'nova_compute:/var/lib/nova', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8022'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'nova-conductor', 'value': {'container_name': 'nova_conductor', 'group': 'nova-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-conductor:2024.2', 'volumes': ['/etc/kolla/nova-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'nova-conductor', 'value': {'container_name': 'nova_conductor', 'group': 'nova-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-conductor:2024.2', 'volumes': ['/etc/kolla/nova-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'nova-conductor', 'value': {'container_name': 'nova_conductor', 'group': 'nova-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-conductor:2024.2', 'volumes': ['/etc/kolla/nova-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'nova-compute', 'value': {'container_name': 'nova_compute', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-compute:2024.2', 'environment': {'LIBGUESTFS_BACKEND': 'direct'}, 'privileged': True, 'enabled': True, 'ipc_mode': 'host', 'volumes': ['/etc/kolla/nova-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', 'kolla_logs:/var/log/kolla/', '', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-compute 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'nova-compute', 'value': {'container_name': 'nova_compute', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-compute:2024.2', 'environment': {'LIBGUESTFS_BACKEND': 'direct'}, 'privileged': True, 'enabled': True, 'ipc_mode': 'host', 'volumes': ['/etc/kolla/nova-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', 'kolla_logs:/var/log/kolla/', '', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-compute 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'nova-compute', 'value': {'container_name': 'nova_compute', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-compute:2024.2', 'environment': {'LIBGUESTFS_BACKEND': 'direct'}, 'privileged': True, 'enabled': True, 'ipc_mode': 'host', 'volumes': ['/etc/kolla/nova-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', 'kolla_logs:/var/log/kolla/', '', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-compute 5672'], 'timeout': '30'}}})\u001b[0m\n\nTASK [nova-cell : Copying over nova.conf] **************************************\nTuesday 24 February 2026  01:13:24 +0000 (0:00:04.272)       0:05:27.702 ****** \n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'nova-libvirt', 'value': {'container_name': 'nova_libvirt', 'group': 'compute', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-libvirt:2024.2', 'pid_mode': 'host', 'cgroupns_mode': 'host', 'privileged': True, 'volumes': ['/etc/kolla/nova-libvirt/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', '', '/sys/fs/cgroup:/sys/fs/cgroup', 'kolla_logs:/var/log/kolla/', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', 'nova_libvirt_qemu:/etc/libvirt/qemu', ''], 'dimensions': {'ulimits': {'memlock': {'soft': 67108864, 'hard': 67108864}}}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'virsh version --daemon'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'nova-ssh', 'value': {'container_name': 'nova_ssh', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-ssh:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla', 'nova_compute:/var/lib/nova', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8022'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'nova-libvirt', 'value': {'container_name': 'nova_libvirt', 'group': 'compute', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-libvirt:2024.2', 'pid_mode': 'host', 'cgroupns_mode': 'host', 'privileged': True, 'volumes': ['/etc/kolla/nova-libvirt/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', '', '/sys/fs/cgroup:/sys/fs/cgroup', 'kolla_logs:/var/log/kolla/', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', 'nova_libvirt_qemu:/etc/libvirt/qemu', ''], 'dimensions': {'ulimits': {'memlock': {'soft': 67108864, 'hard': 67108864}}}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'virsh version --daemon'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'nova-ssh', 'value': {'container_name': 'nova_ssh', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-ssh:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla', 'nova_compute:/var/lib/nova', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8022'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'nova-libvirt', 'value': {'container_name': 'nova_libvirt', 'group': 'compute', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-libvirt:2024.2', 'pid_mode': 'host', 'cgroupns_mode': 'host', 'privileged': True, 'volumes': ['/etc/kolla/nova-libvirt/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', '', '/sys/fs/cgroup:/sys/fs/cgroup', 'kolla_logs:/var/log/kolla/', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', 'nova_libvirt_qemu:/etc/libvirt/qemu', ''], 'dimensions': {'ulimits': {'memlock': {'soft': 67108864, 'hard': 67108864}}}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'virsh version --daemon'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'nova-ssh', 'value': {'container_name': 'nova_ssh', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-ssh:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla', 'nova_compute:/var/lib/nova', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8022'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'nova-compute', 'value': {'container_name': 'nova_compute', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-compute:2024.2', 'environment': {'LIBGUESTFS_BACKEND': 'direct'}, 'privileged': True, 'enabled': True, 'ipc_mode': 'host', 'volumes': ['/etc/kolla/nova-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', 'kolla_logs:/var/log/kolla/', '', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-compute 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'nova-novncproxy', 'value': {'container_name': 'nova_novncproxy', 'group': 'nova-novncproxy', 'image': 'registry.osism.tech/kolla/nova-novncproxy:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-novncproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:6080/vnc_lite.html'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'nova-compute', 'value': {'container_name': 'nova_compute', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-compute:2024.2', 'environment': {'LIBGUESTFS_BACKEND': 'direct'}, 'privileged': True, 'enabled': True, 'ipc_mode': 'host', 'volumes': ['/etc/kolla/nova-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', 'kolla_logs:/var/log/kolla/', '', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-compute 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'nova-compute', 'value': {'container_name': 'nova_compute', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-compute:2024.2', 'environment': {'LIBGUESTFS_BACKEND': 'direct'}, 'privileged': True, 'enabled': True, 'ipc_mode': 'host', 'volumes': ['/etc/kolla/nova-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', 'kolla_logs:/var/log/kolla/', '', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-compute 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'nova-novncproxy', 'value': {'container_name': 'nova_novncproxy', 'group': 'nova-novncproxy', 'image': 'registry.osism.tech/kolla/nova-novncproxy:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-novncproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:6080/vnc_lite.html'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'nova-novncproxy', 'value': {'container_name': 'nova_novncproxy', 'group': 'nova-novncproxy', 'image': 'registry.osism.tech/kolla/nova-novncproxy:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-novncproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:6080/vnc_lite.html'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'nova-conductor', 'value': {'container_name': 'nova_conductor', 'group': 'nova-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-conductor:2024.2', 'volumes': ['/etc/kolla/nova-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'nova-conductor', 'value': {'container_name': 'nova_conductor', 'group': 'nova-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-conductor:2024.2', 'volumes': ['/etc/kolla/nova-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'nova-conductor', 'value': {'container_name': 'nova_conductor', 'group': 'nova-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-conductor:2024.2', 'volumes': ['/etc/kolla/nova-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}})\u001b[0m\n\nTASK [nova-cell : Copying over Nova compute provider config] *******************\nTuesday 24 February 2026  01:13:31 +0000 (0:00:07.196)       0:05:34.899 ****** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [nova-cell : Copying over libvirt configuration] **************************\nTuesday 24 February 2026  01:13:33 +0000 (0:00:01.789)       0:05:36.688 ****** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'src': 'qemu.conf.j2', 'dest': 'qemu.conf'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'src': 'qemu.conf.j2', 'dest': 'qemu.conf'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'src': 'qemu.conf.j2', 'dest': 'qemu.conf'}) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'src': 'qemu.conf.j2', 'dest': 'qemu.conf'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'src': 'qemu.conf.j2', 'dest': 'qemu.conf'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'src': 'qemu.conf.j2', 'dest': 'qemu.conf'})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'src': 'libvirtd.conf.j2', 'dest': 'libvirtd.conf'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'src': 'libvirtd.conf.j2', 'dest': 'libvirtd.conf'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'src': 'libvirtd.conf.j2', 'dest': 'libvirtd.conf'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'src': 'libvirtd.conf.j2', 'dest': 'libvirtd.conf'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'src': 'libvirtd.conf.j2', 'dest': 'libvirtd.conf'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'src': 'libvirtd.conf.j2', 'dest': 'libvirtd.conf'})\u001b[0m\n\nTASK [nova-cell : Copying over libvirt TLS keys] *******************************\nTuesday 24 February 2026  01:13:37 +0000 (0:00:04.058)       0:05:40.747 ****** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [nova-cell : Copying over libvirt SASL configuration] *********************\nTuesday 24 February 2026  01:13:38 +0000 (0:00:00.674)       0:05:41.421 ****** \n\u001b[0;36mskipping: [testbed-node-2] => (item={'src': 'auth.conf.j2', 'dest': 'auth.conf', 'service': 'nova-compute'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'src': 'auth.conf.j2', 'dest': 'auth.conf', 'service': 'nova-compute'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'src': 'auth.conf.j2', 'dest': 'auth.conf', 'service': 'nova-compute'}) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'src': 'auth.conf.j2', 'dest': 'auth.conf', 'service': 'nova-compute'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'src': 'auth.conf.j2', 'dest': 'auth.conf', 'service': 'nova-compute'})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'src': 'auth.conf.j2', 'dest': 'auth.conf', 'service': 'nova-libvirt'}) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'src': 'auth.conf.j2', 'dest': 'auth.conf', 'service': 'nova-compute'})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'src': 'auth.conf.j2', 'dest': 'auth.conf', 'service': 'nova-libvirt'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'src': 'auth.conf.j2', 'dest': 'auth.conf', 'service': 'nova-libvirt'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'src': 'sasl.conf.j2', 'dest': 'sasl.conf', 'service': 'nova-libvirt'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'src': 'sasl.conf.j2', 'dest': 'sasl.conf', 'service': 'nova-libvirt'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'src': 'sasl.conf.j2', 'dest': 'sasl.conf', 'service': 'nova-libvirt'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'src': 'auth.conf.j2', 'dest': 'auth.conf', 'service': 'nova-libvirt'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'src': 'auth.conf.j2', 'dest': 'auth.conf', 'service': 'nova-libvirt'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'src': 'auth.conf.j2', 'dest': 'auth.conf', 'service': 'nova-libvirt'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'src': 'sasl.conf.j2', 'dest': 'sasl.conf', 'service': 'nova-libvirt'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'src': 'sasl.conf.j2', 'dest': 'sasl.conf', 'service': 'nova-libvirt'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'src': 'sasl.conf.j2', 'dest': 'sasl.conf', 'service': 'nova-libvirt'})\u001b[0m\n\nTASK [nova-cell : Copying files for nova-ssh] **********************************\nTuesday 24 February 2026  01:13:44 +0000 (0:00:05.897)       0:05:47.319 ****** \n\u001b[0;36mskipping: [testbed-node-1] => (item={'src': 'sshd_config.j2', 'dest': 'sshd_config'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'src': 'sshd_config.j2', 'dest': 'sshd_config'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'src': 'sshd_config.j2', 'dest': 'sshd_config'}) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'src': 'sshd_config.j2', 'dest': 'sshd_config'})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'src': 'id_rsa', 'dest': 'id_rsa'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'src': 'id_rsa', 'dest': 'id_rsa'}) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'src': 'sshd_config.j2', 'dest': 'sshd_config'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'src': 'sshd_config.j2', 'dest': 'sshd_config'})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'src': 'id_rsa.pub', 'dest': 'id_rsa.pub'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'src': 'id_rsa.pub', 'dest': 'id_rsa.pub'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'src': 'id_rsa', 'dest': 'id_rsa'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'src': 'ssh_config.j2', 'dest': 'ssh_config'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'src': 'ssh_config.j2', 'dest': 'ssh_config'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'src': 'id_rsa.pub', 'dest': 'id_rsa.pub'}) \u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'src': 'id_rsa', 'dest': 'id_rsa'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'src': 'id_rsa', 'dest': 'id_rsa'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'src': 'id_rsa', 'dest': 'id_rsa'})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'src': 'ssh_config.j2', 'dest': 'ssh_config'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'src': 'id_rsa.pub', 'dest': 'id_rsa.pub'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'src': 'id_rsa.pub', 'dest': 'id_rsa.pub'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'src': 'id_rsa.pub', 'dest': 'id_rsa.pub'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'src': 'ssh_config.j2', 'dest': 'ssh_config'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'src': 'ssh_config.j2', 'dest': 'ssh_config'})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'src': 'ssh_config.j2', 'dest': 'ssh_config'})\u001b[0m\n\nTASK [nova-cell : Copying VMware vCenter CA file] ******************************\nTuesday 24 February 2026  01:13:52 +0000 (0:00:07.804)       0:05:55.123 ****** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [nova-cell : Copying 'release' file for nova_compute] *********************\nTuesday 24 February 2026  01:13:53 +0000 (0:00:00.924)       0:05:56.048 ****** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [nova-cell : Generating 'hostnqn' file for nova_compute] ******************\nTuesday 24 February 2026  01:13:53 +0000 (0:00:00.635)       0:05:56.683 ****** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\nTASK [nova-cell : Copying over existing policy file] ***************************\nTuesday 24 February 2026  01:13:55 +0000 (0:00:02.193)       0:05:58.877 ****** \n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'nova-libvirt', 'value': {'container_name': 'nova_libvirt', 'group': 'compute', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-libvirt:2024.2', 'pid_mode': 'host', 'cgroupns_mode': 'host', 'privileged': True, 'volumes': ['/etc/kolla/nova-libvirt/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', '', '/sys/fs/cgroup:/sys/fs/cgroup', 'kolla_logs:/var/log/kolla/', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', 'nova_libvirt_qemu:/etc/libvirt/qemu', ''], 'dimensions': {'ulimits': {'memlock': {'soft': 67108864, 'hard': 67108864}}}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'virsh version --daemon'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'nova-ssh', 'value': {'container_name': 'nova_ssh', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-ssh:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla', 'nova_compute:/var/lib/nova', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8022'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'nova-compute', 'value': {'container_name': 'nova_compute', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-compute:2024.2', 'environment': {'LIBGUESTFS_BACKEND': 'direct'}, 'privileged': True, 'enabled': True, 'ipc_mode': 'host', 'volumes': ['/etc/kolla/nova-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', 'kolla_logs:/var/log/kolla/', '', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-compute 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'nova-libvirt', 'value': {'container_name': 'nova_libvirt', 'group': 'compute', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-libvirt:2024.2', 'pid_mode': 'host', 'cgroupns_mode': 'host', 'privileged': True, 'volumes': ['/etc/kolla/nova-libvirt/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', '', '/sys/fs/cgroup:/sys/fs/cgroup', 'kolla_logs:/var/log/kolla/', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', 'nova_libvirt_qemu:/etc/libvirt/qemu', ''], 'dimensions': {'ulimits': {'memlock': {'soft': 67108864, 'hard': 67108864}}}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'virsh version --daemon'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'nova-ssh', 'value': {'container_name': 'nova_ssh', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-ssh:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla', 'nova_compute:/var/lib/nova', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8022'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'nova-compute', 'value': {'container_name': 'nova_compute', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-compute:2024.2', 'environment': {'LIBGUESTFS_BACKEND': 'direct'}, 'privileged': True, 'enabled': True, 'ipc_mode': 'host', 'volumes': ['/etc/kolla/nova-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', 'kolla_logs:/var/log/kolla/', '', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-compute 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'nova-libvirt', 'value': {'container_name': 'nova_libvirt', 'group': 'compute', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-libvirt:2024.2', 'pid_mode': 'host', 'cgroupns_mode': 'host', 'privileged': True, 'volumes': ['/etc/kolla/nova-libvirt/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', '', '/sys/fs/cgroup:/sys/fs/cgroup', 'kolla_logs:/var/log/kolla/', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', 'nova_libvirt_qemu:/etc/libvirt/qemu', ''], 'dimensions': {'ulimits': {'memlock': {'soft': 67108864, 'hard': 67108864}}}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'virsh version --daemon'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'nova-ssh', 'value': {'container_name': 'nova_ssh', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-ssh:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla', 'nova_compute:/var/lib/nova', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8022'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'nova-compute', 'value': {'container_name': 'nova_compute', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-compute:2024.2', 'environment': {'LIBGUESTFS_BACKEND': 'direct'}, 'privileged': True, 'enabled': True, 'ipc_mode': 'host', 'volumes': ['/etc/kolla/nova-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', 'kolla_logs:/var/log/kolla/', '', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-compute 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova-novncproxy', 'value': {'container_name': 'nova_novncproxy', 'group': 'nova-novncproxy', 'image': 'registry.osism.tech/kolla/nova-novncproxy:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-novncproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:6080/vnc_lite.html'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova-conductor', 'value': {'container_name': 'nova_conductor', 'group': 'nova-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-conductor:2024.2', 'volumes': ['/etc/kolla/nova-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova-novncproxy', 'value': {'container_name': 'nova_novncproxy', 'group': 'nova-novncproxy', 'image': 'registry.osism.tech/kolla/nova-novncproxy:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-novncproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:6080/vnc_lite.html'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova-conductor', 'value': {'container_name': 'nova_conductor', 'group': 'nova-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-conductor:2024.2', 'volumes': ['/etc/kolla/nova-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova-novncproxy', 'value': {'container_name': 'nova_novncproxy', 'group': 'nova-novncproxy', 'image': 'registry.osism.tech/kolla/nova-novncproxy:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-novncproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:6080/vnc_lite.html'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova-conductor', 'value': {'container_name': 'nova_conductor', 'group': 'nova-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-conductor:2024.2', 'volumes': ['/etc/kolla/nova-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [nova-cell : Copying over vendordata file to containers] ******************\nTuesday 24 February 2026  01:13:57 +0000 (0:00:01.834)       0:06:00.711 ****** \n\u001b[0;36mskipping: [testbed-node-3] => (item=nova-compute) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=nova-compute-ironic) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=nova-compute) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=nova-compute-ironic) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=nova-compute) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=nova-compute-ironic) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=nova-compute) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=nova-compute-ironic) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=nova-compute) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=nova-compute-ironic) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=nova-compute) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=nova-compute-ironic) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [nova-cell : Check nova-cell containers] **********************************\nTuesday 24 February 2026  01:13:58 +0000 (0:00:01.001)       0:06:01.713 ****** \n\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'nova-libvirt', 'value': {'container_name': 'nova_libvirt', 'group': 'compute', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-libvirt:2024.2', 'pid_mode': 'host', 'cgroupns_mode': 'host', 'privileged': True, 'volumes': ['/etc/kolla/nova-libvirt/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', '', '/sys/fs/cgroup:/sys/fs/cgroup', 'kolla_logs:/var/log/kolla/', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', 'nova_libvirt_qemu:/etc/libvirt/qemu', ''], 'dimensions': {'ulimits': {'memlock': {'soft': 67108864, 'hard': 67108864}}}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'virsh version --daemon'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'nova-libvirt', 'value': {'container_name': 'nova_libvirt', 'group': 'compute', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-libvirt:2024.2', 'pid_mode': 'host', 'cgroupns_mode': 'host', 'privileged': True, 'volumes': ['/etc/kolla/nova-libvirt/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', '', '/sys/fs/cgroup:/sys/fs/cgroup', 'kolla_logs:/var/log/kolla/', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', 'nova_libvirt_qemu:/etc/libvirt/qemu', ''], 'dimensions': {'ulimits': {'memlock': {'soft': 67108864, 'hard': 67108864}}}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'virsh version --daemon'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'nova-libvirt', 'value': {'container_name': 'nova_libvirt', 'group': 'compute', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-libvirt:2024.2', 'pid_mode': 'host', 'cgroupns_mode': 'host', 'privileged': True, 'volumes': ['/etc/kolla/nova-libvirt/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', '', '/sys/fs/cgroup:/sys/fs/cgroup', 'kolla_logs:/var/log/kolla/', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', 'nova_libvirt_qemu:/etc/libvirt/qemu', ''], 'dimensions': {'ulimits': {'memlock': {'soft': 67108864, 'hard': 67108864}}}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'virsh version --daemon'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'nova-novncproxy', 'value': {'container_name': 'nova_novncproxy', 'group': 'nova-novncproxy', 'image': 'registry.osism.tech/kolla/nova-novncproxy:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-novncproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:6080/vnc_lite.html'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'nova-novncproxy', 'value': {'container_name': 'nova_novncproxy', 'group': 'nova-novncproxy', 'image': 'registry.osism.tech/kolla/nova-novncproxy:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-novncproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:6080/vnc_lite.html'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'nova-ssh', 'value': {'container_name': 'nova_ssh', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-ssh:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla', 'nova_compute:/var/lib/nova', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8022'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'nova-ssh', 'value': {'container_name': 'nova_ssh', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-ssh:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla', 'nova_compute:/var/lib/nova', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8022'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'nova-ssh', 'value': {'container_name': 'nova_ssh', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-ssh:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla', 'nova_compute:/var/lib/nova', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8022'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'nova-compute', 'value': {'container_name': 'nova_compute', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-compute:2024.2', 'environment': {'LIBGUESTFS_BACKEND': 'direct'}, 'privileged': True, 'enabled': True, 'ipc_mode': 'host', 'volumes': ['/etc/kolla/nova-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', 'kolla_logs:/var/log/kolla/', '', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-compute 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'nova-novncproxy', 'value': {'container_name': 'nova_novncproxy', 'group': 'nova-novncproxy', 'image': 'registry.osism.tech/kolla/nova-novncproxy:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-novncproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:6080/vnc_lite.html'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'nova-conductor', 'value': {'container_name': 'nova_conductor', 'group': 'nova-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-conductor:2024.2', 'volumes': ['/etc/kolla/nova-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'nova-conductor', 'value': {'container_name': 'nova_conductor', 'group': 'nova-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-conductor:2024.2', 'volumes': ['/etc/kolla/nova-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'nova-compute', 'value': {'container_name': 'nova_compute', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-compute:2024.2', 'environment': {'LIBGUESTFS_BACKEND': 'direct'}, 'privileged': True, 'enabled': True, 'ipc_mode': 'host', 'volumes': ['/etc/kolla/nova-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', 'kolla_logs:/var/log/kolla/', '', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-compute 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'nova-compute', 'value': {'container_name': 'nova_compute', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-compute:2024.2', 'environment': {'LIBGUESTFS_BACKEND': 'direct'}, 'privileged': True, 'enabled': True, 'ipc_mode': 'host', 'volumes': ['/etc/kolla/nova-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', 'kolla_logs:/var/log/kolla/', '', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-compute 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'nova-conductor', 'value': {'container_name': 'nova_conductor', 'group': 'nova-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-conductor:2024.2', 'volumes': ['/etc/kolla/nova-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}})\u001b[0m\n\nTASK [nova-cell : include_tasks] ***********************************************\nTuesday 24 February 2026  01:14:01 +0000 (0:00:03.157)       0:06:04.871 ****** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [nova-cell : Flush handlers] **********************************************\nTuesday 24 February 2026  01:14:02 +0000 (0:00:00.891)       0:06:05.762 ****** \n\nTASK [nova-cell : Flush handlers] **********************************************\nTuesday 24 February 2026  01:14:02 +0000 (0:00:00.154)       0:06:05.917 ****** \n\nTASK [nova-cell : Flush handlers] **********************************************\nTuesday 24 February 2026  01:14:03 +0000 (0:00:00.141)       0:06:06.058 ****** \n\nTASK [nova-cell : Flush handlers] **********************************************\nTuesday 24 February 2026  01:14:03 +0000 (0:00:00.140)       0:06:06.199 ****** \n\nTASK [nova-cell : Flush handlers] **********************************************\nTuesday 24 February 2026  01:14:03 +0000 (0:00:00.133)       0:06:06.332 ****** \n\nTASK [nova-cell : Flush handlers] **********************************************\nTuesday 24 February 2026  01:14:03 +0000 (0:00:00.345)       0:06:06.678 ****** \n\nRUNNING HANDLER [nova-cell : Restart nova-conductor container] *****************\nTuesday 24 February 2026  01:14:03 +0000 (0:00:00.154)       0:06:06.832 ****** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [nova-cell : Restart nova-novncproxy container] ****************\nTuesday 24 February 2026  01:14:11 +0000 (0:00:07.882)       0:06:14.715 ****** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [nova-cell : Restart nova-ssh container] ***********************\nTuesday 24 February 2026  01:14:30 +0000 (0:00:18.894)       0:06:33.610 ****** \n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\nRUNNING HANDLER [nova-cell : Restart nova-libvirt container] *******************\nTuesday 24 February 2026  01:14:53 +0000 (0:00:23.141)       0:06:56.751 ****** \n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\nRUNNING HANDLER [nova-cell : Checking libvirt container is ready] **************\nTuesday 24 February 2026  01:15:26 +0000 (0:00:32.606)       0:07:29.357 ****** \n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nRUNNING HANDLER [nova-cell : Create libvirt SASL user] *************************\nTuesday 24 February 2026  01:15:27 +0000 (0:00:00.821)       0:07:30.178 ****** \n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nRUNNING HANDLER [nova-cell : Restart nova-compute container] *******************\nTuesday 24 February 2026  01:15:27 +0000 (0:00:00.807)       0:07:30.986 ****** \n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\nRUNNING HANDLER [nova-cell : Wait for nova-compute services to update service versions] ***\nTuesday 24 February 2026  01:15:53 +0000 (0:00:25.952)       0:07:56.938 ****** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [nova-cell : Waiting for nova-compute services to register themselves] ****\nTuesday 24 February 2026  01:15:54 +0000 (0:00:00.184)       0:07:57.123 ****** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[1;30mFAILED - RETRYING: [testbed-node-3 -> testbed-node-0]: Waiting for nova-compute services to register themselves (20 retries left).\u001b[0m\n\u001b[0;32mok: [testbed-node-3 -> testbed-node-0(192.168.16.10)]\u001b[0m\n\nTASK [nova-cell : Fail if nova-compute service failed to register] *************\nTuesday 24 February 2026  01:16:17 +0000 (0:00:23.609)       0:08:20.732 ****** \n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\nTASK [nova-cell : Include discover_computes.yml] *******************************\nTuesday 24 February 2026  01:16:30 +0000 (0:00:12.676)       0:08:33.408 ****** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mincluded: /ansible/roles/nova-cell/tasks/discover_computes.yml for testbed-node-3\u001b[0m\n\nTASK [nova-cell : Get a list of existing cells] ********************************\nTuesday 24 February 2026  01:16:34 +0000 (0:00:04.225)       0:08:37.634 ****** \n\u001b[0;32mok: [testbed-node-3 -> testbed-node-0(192.168.16.10)]\u001b[0m\n\nTASK [nova-cell : Extract current cell settings from list] *********************\nTuesday 24 February 2026  01:16:49 +0000 (0:00:14.715)       0:08:52.349 ****** \n\u001b[0;32mok: [testbed-node-3 -> testbed-node-0(192.168.16.10)]\u001b[0m\n\nTASK [nova-cell : Fail if cell settings not found] *****************************\nTuesday 24 February 2026  01:16:51 +0000 (0:00:01.656)       0:08:54.006 ****** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [nova-cell : Discover nova hosts] *****************************************\nTuesday 24 February 2026  01:16:52 +0000 (0:00:01.577)       0:08:55.583 ****** \n\u001b[0;32mok: [testbed-node-3 -> testbed-node-0(192.168.16.10)]\u001b[0m\n\nTASK [nova-cell : Remove old nova_libvirt_secrets container volume] ************\nTuesday 24 February 2026  01:17:05 +0000 (0:00:12.692)       0:09:08.275 ****** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nPLAY [Refresh nova scheduler cell cache] ***************************************\n\nTASK [nova : Refresh cell cache in nova scheduler] *****************************\nTuesday 24 February 2026  01:17:07 +0000 (0:00:02.057)       0:09:10.333 ****** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nPLAY [Reload global Nova super conductor services] *****************************\n\nTASK [nova : Reload nova super conductor services to remove RPC version pin] ***\nTuesday 24 February 2026  01:17:08 +0000 (0:00:01.342)       0:09:11.675 ****** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nPLAY [Reload Nova cell services] ***********************************************\n\nTASK [nova-cell : Reload nova cell services to remove RPC version cap] *********\nTuesday 24 February 2026  01:17:09 +0000 (0:00:00.689)       0:09:12.365 ****** \n\u001b[0;36mskipping: [testbed-node-3] => (item=nova-conductor) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=nova-compute) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=nova-compute-ironic) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=nova-novncproxy) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=nova-serialproxy) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item=nova-spicehtml5proxy) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=nova-conductor) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=nova-compute) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=nova-compute-ironic) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=nova-novncproxy) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=nova-serialproxy) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item=nova-spicehtml5proxy) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=nova-conductor) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=nova-compute) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=nova-compute-ironic) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=nova-novncproxy) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=nova-serialproxy) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item=nova-spicehtml5proxy) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=nova-conductor) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=nova-compute) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=nova-compute-ironic) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=nova-novncproxy) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=nova-serialproxy) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=nova-spicehtml5proxy) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=nova-conductor) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=nova-compute) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=nova-compute-ironic) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=nova-conductor) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=nova-compute) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=nova-novncproxy) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=nova-compute-ironic) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=nova-serialproxy) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=nova-novncproxy) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=nova-spicehtml5proxy) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=nova-serialproxy) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=nova-spicehtml5proxy) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nPLAY [Reload global Nova API services] *****************************************\n\nTASK [nova : Reload nova API services to remove RPC version pin] ***************\nTuesday 24 February 2026  01:17:11 +0000 (0:00:01.956)       0:09:14.321 ****** \n\u001b[0;36mskipping: [testbed-node-0] => (item=nova-scheduler) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item=nova-api) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=nova-scheduler) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item=nova-api) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=nova-scheduler) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item=nova-api) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nPLAY [Run Nova API online data migrations] *************************************\n\nTASK [nova : Run Nova API online database migrations] **************************\nTuesday 24 February 2026  01:17:12 +0000 (0:00:00.916)       0:09:15.238 ****** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nPLAY [Run Nova cell online data migrations] ************************************\n\nTASK [nova-cell : Run Nova cell online database migrations] ********************\nTuesday 24 February 2026  01:17:13 +0000 (0:00:01.028)       0:09:16.267 ****** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;33mtestbed-manager\u001b[0m            : \u001b[0;32mok=3   \u001b[0m \u001b[0;33mchanged=3   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=54  \u001b[0m \u001b[0;33mchanged=35  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=45  \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=27  \u001b[0m \u001b[0;33mchanged=19  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=52  \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=27  \u001b[0m \u001b[0;33mchanged=19  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=52  \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-3\u001b[0m             : \u001b[0;32mok=45  \u001b[0m \u001b[0;33mchanged=27  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=21  \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-4\u001b[0m             : \u001b[0;32mok=39  \u001b[0m \u001b[0;33mchanged=27  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=20  \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-5\u001b[0m             : \u001b[0;32mok=39  \u001b[0m \u001b[0;33mchanged=27  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=20  \u001b[0m rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nTuesday 24 February 2026  01:17:14 +0000 (0:00:00.757)       0:09:17.024 ****** \n=============================================================================== \nnova : Running Nova API bootstrap container ---------------------------- 35.17s\nnova-cell : Restart nova-libvirt container ----------------------------- 32.61s\nnova-cell : Restart nova-compute container ----------------------------- 25.95s\nnova-cell : Waiting for nova-compute services to register themselves --- 23.61s\nnova-cell : Restart nova-ssh container --------------------------------- 23.14s\nnova : Running Nova API bootstrap container ---------------------------- 23.08s\nnova : Restart nova-scheduler container -------------------------------- 22.54s\nnova-cell : Running Nova cell bootstrap container ---------------------- 22.16s\nnova-cell : Restart nova-novncproxy container -------------------------- 18.89s\nnova : Create cell0 mappings ------------------------------------------- 17.69s\nnova-cell : Get a list of existing cells ------------------------------- 14.72s\nnova-cell : Get a list of existing cells ------------------------------- 14.69s\nnova-cell : Create cell ------------------------------------------------ 14.40s\nnova-cell : Get a list of existing cells ------------------------------- 13.55s\nnova-cell : Discover nova hosts ---------------------------------------- 12.69s\nnova-cell : Fail if nova-compute service failed to register ------------ 12.68s\nservice-rabbitmq : nova | Ensure RabbitMQ users exist ------------------- 8.82s\nservice-ks-register : nova | Granting user roles ------------------------ 8.53s\nnova-cell : Restart nova-conductor container ---------------------------- 7.88s\nnova-cell : Copying files for nova-ssh ---------------------------------- 7.80s\n\u001b[32m2026-02-24 01:17:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 01:17:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 79f56a03-a0d2-400b-aa03-9d9321f018b2 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:17:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 01:17:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 79f56a03-a0d2-400b-aa03-9d9321f018b2 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:17:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 01:17:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 79f56a03-a0d2-400b-aa03-9d9321f018b2 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:17:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 01:17:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 79f56a03-a0d2-400b-aa03-9d9321f018b2 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:17:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 01:17:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 79f56a03-a0d2-400b-aa03-9d9321f018b2 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:17:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 01:17:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 79f56a03-a0d2-400b-aa03-9d9321f018b2 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:17:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 01:17:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 79f56a03-a0d2-400b-aa03-9d9321f018b2 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:17:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 01:17:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 79f56a03-a0d2-400b-aa03-9d9321f018b2 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:17:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\u001b[32m2026-02-24 01:17:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 79f56a03-a0d2-400b-aa03-9d9321f018b2 is in state STARTED\u001b[0m\n\u001b[32m2026-02-24 01:17:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m\n\n\nPLAY [Group hosts based on configuration] **************************************\n\nTASK [Group hosts based on Kolla action] ***************************************\nTuesday 24 February 2026  01:12:37 +0000 (0:00:00.286)       0:00:00.286 ****** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [Group hosts based on enabled services] ***********************************\nTuesday 24 February 2026  01:12:38 +0000 (0:00:00.385)       0:00:00.672 ****** \n\u001b[0;32mok: [testbed-node-0] => (item=enable_octavia_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item=enable_octavia_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item=enable_octavia_True)\u001b[0m\n\nPLAY [Apply role octavia] ******************************************************\n\nTASK [octavia : include_tasks] *************************************************\nTuesday 24 February 2026  01:12:38 +0000 (0:00:00.502)       0:00:01.174 ****** \n\u001b[0;36mincluded: /ansible/roles/octavia/tasks/deploy.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [service-ks-register : octavia | Creating services] ***********************\nTuesday 24 February 2026  01:12:39 +0000 (0:00:00.591)       0:00:01.766 ****** \n\u001b[0;33mchanged: [testbed-node-0] => (item=octavia (load-balancer))\u001b[0m\n\nTASK [service-ks-register : octavia | Creating endpoints] **********************\nTuesday 24 February 2026  01:12:43 +0000 (0:00:03.714)       0:00:05.480 ****** \n\u001b[0;33mchanged: [testbed-node-0] => (item=octavia -> https://api-int.testbed.osism.xyz:9876 -> internal)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=octavia -> https://api.testbed.osism.xyz:9876 -> public)\u001b[0m\n\nTASK [service-ks-register : octavia | Creating projects] ***********************\nTuesday 24 February 2026  01:12:49 +0000 (0:00:06.856)       0:00:12.337 ****** \n\u001b[0;32mok: [testbed-node-0] => (item=service)\u001b[0m\n\nTASK [service-ks-register : octavia | Creating users] **************************\nTuesday 24 February 2026  01:12:53 +0000 (0:00:03.144)       0:00:15.482 ****** \n\u001b[0;33mchanged: [testbed-node-0] => (item=octavia -> service)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=octavia -> service)\u001b[0m\n\u001b[1;35m[WARNING]: Module did not set no_log for update_password\u001b[0m\n\nTASK [service-ks-register : octavia | Creating roles] **************************\nTuesday 24 February 2026  01:13:02 +0000 (0:00:08.997)       0:00:24.479 ****** \n\u001b[0;32mok: [testbed-node-0] => (item=admin)\u001b[0m\n\nTASK [service-ks-register : octavia | Granting user roles] *********************\nTuesday 24 February 2026  01:13:06 +0000 (0:00:04.127)       0:00:28.607 ****** \n\u001b[0;33mchanged: [testbed-node-0] => (item=octavia -> service -> admin)\u001b[0m\n\u001b[0;32mok: [testbed-node-0] => (item=octavia -> service -> admin)\u001b[0m\n\nTASK [octavia : Adding octavia related roles] **********************************\nTuesday 24 February 2026  01:13:13 +0000 (0:00:07.061)       0:00:35.668 ****** \n\u001b[0;33mchanged: [testbed-node-0] => (item=load-balancer_observer)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=load-balancer_global_observer)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=load-balancer_member)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=load-balancer_admin)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=load-balancer_quota_admin)\u001b[0m\n\nTASK [octavia : include_tasks] *************************************************\nTuesday 24 February 2026  01:13:30 +0000 (0:00:17.369)       0:00:53.038 ****** \n\u001b[0;36mincluded: /ansible/roles/octavia/tasks/prepare.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [octavia : Create amphora flavor] *****************************************\nTuesday 24 February 2026  01:13:31 +0000 (0:00:00.715)       0:00:53.753 ****** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [octavia : Create nova keypair for amphora] *******************************\nTuesday 24 February 2026  01:13:37 +0000 (0:00:05.860)       0:00:59.614 ****** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [octavia : Get service project id] ****************************************\nTuesday 24 February 2026  01:13:42 +0000 (0:00:05.274)       0:01:04.888 ****** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [octavia : Create security groups for octavia] ****************************\nTuesday 24 February 2026  01:13:46 +0000 (0:00:03.939)       0:01:08.827 ****** \n\u001b[0;33mchanged: [testbed-node-0] => (item=lb-mgmt-sec-grp)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=lb-health-mgr-sec-grp)\u001b[0m\n\nTASK [octavia : Add rules for security groups] *********************************\nTuesday 24 February 2026  01:13:57 +0000 (0:00:11.020)       0:01:19.848 ****** \n\u001b[0;33mchanged: [testbed-node-0] => (item=[{'name': 'lb-mgmt-sec-grp', 'enabled': True}, {'protocol': 'icmp'}])\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=[{'name': 'lb-mgmt-sec-grp', 'enabled': True}, {'protocol': 'tcp', 'src_port': 22, 'dst_port': 22}])\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=[{'name': 'lb-mgmt-sec-grp', 'enabled': True}, {'protocol': 'tcp', 'src_port': '9443', 'dst_port': '9443'}])\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=[{'name': 'lb-health-mgr-sec-grp', 'enabled': True}, {'protocol': 'udp', 'src_port': '5555', 'dst_port': '5555'}])\u001b[0m\n\nTASK [octavia : Create loadbalancer management network] ************************\nTuesday 24 February 2026  01:14:12 +0000 (0:00:15.234)       0:01:35.083 ****** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [octavia : Create loadbalancer management subnet] *************************\nTuesday 24 February 2026  01:14:16 +0000 (0:00:04.275)       0:01:39.359 ****** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [octavia : Create loadbalancer management router for IPv6] ****************\nTuesday 24 February 2026  01:14:23 +0000 (0:00:06.702)       0:01:46.062 ****** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [octavia : Update loadbalancer management subnet] *************************\nTuesday 24 February 2026  01:14:23 +0000 (0:00:00.253)       0:01:46.315 ****** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [octavia : include_tasks] *************************************************\nTuesday 24 February 2026  01:14:27 +0000 (0:00:03.925)       0:01:50.240 ****** \n\u001b[0;36mincluded: /ansible/roles/octavia/tasks/hm-interface.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [octavia : Create ports for Octavia health-manager nodes] *****************\nTuesday 24 February 2026  01:14:29 +0000 (0:00:01.228)       0:01:51.468 ****** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\nTASK [octavia : Update Octavia health manager port host_id] ********************\nTuesday 24 February 2026  01:14:35 +0000 (0:00:06.029)       0:01:57.498 ****** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\nTASK [octavia : Add Octavia port to openvswitch br-int] ************************\nTuesday 24 February 2026  01:14:40 +0000 (0:00:05.274)       0:02:02.772 ****** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [octavia : Install isc-dhcp-client package] *******************************\nTuesday 24 February 2026  01:14:41 +0000 (0:00:00.823)       0:02:03.595 ****** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\nTASK [octavia : Create octavia dhclient conf] **********************************\nTuesday 24 February 2026  01:14:44 +0000 (0:00:02.781)       0:02:06.377 ****** \n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [octavia : Create octavia-interface service] ******************************\nTuesday 24 February 2026  01:14:45 +0000 (0:00:01.414)       0:02:07.791 ****** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [octavia : Restart octavia-interface.service if required] *****************\nTuesday 24 February 2026  01:14:46 +0000 (0:00:01.272)       0:02:09.064 ****** \n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [octavia : Enable and start octavia-interface.service] ********************\nTuesday 24 February 2026  01:14:48 +0000 (0:00:02.238)       0:02:11.303 ****** \n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [octavia : Wait for interface ohm0 ip appear] *****************************\nTuesday 24 February 2026  01:14:51 +0000 (0:00:02.075)       0:02:13.378 ****** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [octavia : Gather facts] **************************************************\nTuesday 24 February 2026  01:14:51 +0000 (0:00:00.676)       0:02:14.055 ****** \n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [octavia : include_tasks] *************************************************\nTuesday 24 February 2026  01:14:55 +0000 (0:00:03.875)       0:02:17.930 ****** \n\u001b[0;36mincluded: /ansible/roles/octavia/tasks/get_resources_info.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [octavia : Get amphora flavor info] ***************************************\nTuesday 24 February 2026  01:14:56 +0000 (0:00:00.832)       0:02:18.763 ****** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [octavia : Get service project id] ****************************************\nTuesday 24 February 2026  01:15:00 +0000 (0:00:03.884)       0:02:22.647 ****** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [octavia : Get security groups for octavia] *******************************\nTuesday 24 February 2026  01:15:04 +0000 (0:00:04.068)       0:02:26.716 ****** \n\u001b[0;32mok: [testbed-node-0] => (item=lb-mgmt-sec-grp)\u001b[0m\n\u001b[0;32mok: [testbed-node-0] => (item=lb-health-mgr-sec-grp)\u001b[0m\n\nTASK [octavia : Get loadbalancer management network] ***************************\nTuesday 24 February 2026  01:15:11 +0000 (0:00:07.173)       0:02:33.890 ****** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [octavia : Set octavia resources facts] ***********************************\nTuesday 24 February 2026  01:15:15 +0000 (0:00:03.806)       0:02:37.696 ****** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [octavia : Ensuring config directories exist] *****************************\nTuesday 24 February 2026  01:15:15 +0000 (0:00:00.374)       0:02:38.071 ****** \n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-api:2024.2', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-api:2024.2', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-api:2024.2', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-driver-agent:2024.2', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-driver-agent:2024.2', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-driver-agent:2024.2', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-health-manager:2024.2', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-health-manager:2024.2', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-health-manager:2024.2', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-housekeeping:2024.2', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-housekeeping:2024.2', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-housekeeping:2024.2', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-worker:2024.2', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-worker:2024.2', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-worker:2024.2', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}})\u001b[0m\n\nTASK [octavia : Check if policies shall be overwritten] ************************\nTuesday 24 February 2026  01:15:18 +0000 (0:00:02.697)       0:02:40.768 ****** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [octavia : Set octavia policy file] ***************************************\nTuesday 24 February 2026  01:15:18 +0000 (0:00:00.147)       0:02:40.916 ****** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [octavia : Copying over existing policy file] *****************************\nTuesday 24 February 2026  01:15:19 +0000 (0:00:00.637)       0:02:41.553 ****** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-api:2024.2', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-driver-agent:2024.2', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-health-manager:2024.2', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-housekeeping:2024.2', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-worker:2024.2', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-api:2024.2', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-driver-agent:2024.2', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-health-manager:2024.2', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-housekeeping:2024.2', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-worker:2024.2', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-api:2024.2', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-driver-agent:2024.2', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-health-manager:2024.2', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-housekeeping:2024.2', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-worker:2024.2', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [octavia : include_tasks] *************************************************\nTuesday 24 February 2026  01:15:19 +0000 (0:00:00.795)       0:02:42.348 ****** \n\u001b[0;36mincluded: /ansible/roles/octavia/tasks/copy-certs.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [service-cert-copy : octavia | Copying over extra CA certificates] ********\nTuesday 24 February 2026  01:15:20 +0000 (0:00:00.647)       0:02:42.996 ****** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-api:2024.2', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-api:2024.2', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9876'], 'timeout': '30'}, 'hapr\u001b[32m2026-02-24 01:17:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 79f56a03-a0d2-400b-aa03-9d9321f018b2 is in state SUCCESS\u001b[0m\noxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-api:2024.2', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-driver-agent:2024.2', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-driver-agent:2024.2', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-driver-agent:2024.2', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-health-manager:2024.2', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-health-manager:2024.2', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-health-manager:2024.2', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-housekeeping:2024.2', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-housekeeping:2024.2', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-housekeeping:2024.2', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-worker:2024.2', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-worker:2024.2', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-worker:2024.2', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}})\u001b[0m\n\nTASK [service-cert-copy : octavia | Copying over backend internal TLS certificate] ***\nTuesday 24 February 2026  01:15:26 +0000 (0:00:06.004)       0:02:49.001 ****** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-api:2024.2', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-driver-agent:2024.2', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-health-manager:2024.2', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-housekeeping:2024.2', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-worker:2024.2', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-api:2024.2', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-driver-agent:2024.2', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-health-manager:2024.2', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-housekeeping:2024.2', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-worker:2024.2', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-api:2024.2', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-driver-agent:2024.2', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-health-manager:2024.2', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-housekeeping:2024.2', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-worker:2024.2', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [service-cert-copy : octavia | Copying over backend internal TLS key] *****\nTuesday 24 February 2026  01:15:27 +0000 (0:00:00.760)       0:02:49.762 ****** \n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-api:2024.2', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-driver-agent:2024.2', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-health-manager:2024.2', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-housekeeping:2024.2', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-worker:2024.2', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-api:2024.2', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-driver-agent:2024.2', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-health-manager:2024.2', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-housekeeping:2024.2', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-worker:2024.2', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-api:2024.2', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-driver-agent:2024.2', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-health-manager:2024.2', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-housekeeping:2024.2', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-worker:2024.2', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [octavia : Copying over config.json files for services] *******************\nTuesday 24 February 2026  01:15:28 +0000 (0:00:01.303)       0:02:51.066 ****** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-api:2024.2', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-api:2024.2', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-api:2024.2', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-driver-agent:2024.2', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-driver-agent:2024.2', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-driver-agent:2024.2', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[32m2026-02-24 01:17:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until refresh of running tasks\u001b[0m\n => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-health-manager:2024.2', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-health-manager:2024.2', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-health-manager:2024.2', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-housekeeping:2024.2', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-housekeeping:2024.2', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-housekeeping:2024.2', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-worker:2024.2', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-worker:2024.2', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-worker:2024.2', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}})\u001b[0m\n\nTASK [octavia : Copying over octavia-wsgi.conf] ********************************\nTuesday 24 February 2026  01:15:35 +0000 (0:00:06.622)       0:02:57.689 ****** \n\u001b[0;33mchanged: [testbed-node-0] => (item=/ansible/roles/octavia/templates/octavia-wsgi.conf.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=/ansible/roles/octavia/templates/octavia-wsgi.conf.j2)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=/ansible/roles/octavia/templates/octavia-wsgi.conf.j2)\u001b[0m\n\nTASK [octavia : Copying over octavia.conf] *************************************\nTuesday 24 February 2026  01:15:37 +0000 (0:00:02.575)       0:03:00.264 ****** \n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-api:2024.2', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-api:2024.2', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-api:2024.2', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-driver-agent:2024.2', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-driver-agent:2024.2', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-driver-agent:2024.2', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-health-manager:2024.2', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-health-manager:2024.2', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-health-manager:2024.2', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-housekeeping:2024.2', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-housekeeping:2024.2', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-housekeeping:2024.2', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-worker:2024.2', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-worker:2024.2', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-worker:2024.2', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}})\u001b[0m\n\nTASK [octavia : Copying over Octavia SSH key] **********************************\nTuesday 24 February 2026  01:16:00 +0000 (0:00:22.424)       0:03:22.689 ****** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nTASK [octavia : Copying certificate files for octavia-worker] ******************\nTuesday 24 February 2026  01:16:02 +0000 (0:00:01.718)       0:03:24.407 ****** \n\u001b[0;33mchanged: [testbed-node-0] => (item=client.cert-and-key.pem)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=client.cert-and-key.pem)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=client.cert-and-key.pem)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=client_ca.cert.pem)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=client_ca.cert.pem)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=client_ca.cert.pem)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=server_ca.cert.pem)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=server_ca.cert.pem)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=server_ca.cert.pem)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=server_ca.key.pem)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=server_ca.key.pem)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=server_ca.key.pem)\u001b[0m\n\nTASK [octavia : Copying certificate files for octavia-housekeeping] ************\nTuesday 24 February 2026  01:16:07 +0000 (0:00:05.867)       0:03:30.275 ****** \n\u001b[0;33mchanged: [testbed-node-0] => (item=client.cert-and-key.pem)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=client.cert-and-key.pem)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=client.cert-and-key.pem)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=client_ca.cert.pem)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=client_ca.cert.pem)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=client_ca.cert.pem)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=server_ca.cert.pem)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=server_ca.cert.pem)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=server_ca.cert.pem)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=server_ca.key.pem)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=server_ca.key.pem)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=server_ca.key.pem)\u001b[0m\n\nTASK [octavia : Copying certificate files for octavia-health-manager] **********\nTuesday 24 February 2026  01:16:14 +0000 (0:00:06.159)       0:03:36.434 ****** \n\u001b[0;33mchanged: [testbed-node-0] => (item=client.cert-and-key.pem)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=client.cert-and-key.pem)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=client.cert-and-key.pem)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=client_ca.cert.pem)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=client_ca.cert.pem)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=client_ca.cert.pem)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=server_ca.cert.pem)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=server_ca.cert.pem)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=server_ca.cert.pem)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item=server_ca.key.pem)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item=server_ca.key.pem)\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item=server_ca.key.pem)\u001b[0m\n\nTASK [octavia : Check octavia containers] **************************************\nTuesday 24 February 2026  01:16:20 +0000 (0:00:06.732)       0:03:43.167 ****** \n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-api:2024.2', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-api:2024.2', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-api:2024.2', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-driver-agent:2024.2', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-driver-agent:2024.2', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-driver-agent:2024.2', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-health-manager:2024.2', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-health-manager:2024.2', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-health-manager:2024.2', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-housekeeping:2024.2', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-housekeeping:2024.2', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-housekeeping:2024.2', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-worker:2024.2', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-worker:2024.2', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}})\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-worker:2024.2', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}})\u001b[0m\n\nTASK [octavia : include_tasks] *************************************************\nTuesday 24 February 2026  01:16:26 +0000 (0:00:05.296)       0:03:48.463 ****** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [octavia : Creating Octavia database] *************************************\nTuesday 24 February 2026  01:16:26 +0000 (0:00:00.770)       0:03:49.233 ****** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [octavia : Creating Octavia persistence database] *************************\nTuesday 24 February 2026  01:16:29 +0000 (0:00:02.658)       0:03:51.892 ****** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [octavia : Creating Octavia database user and setting permissions] ********\nTuesday 24 February 2026  01:16:32 +0000 (0:00:02.602)       0:03:54.494 ****** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [octavia : Creating Octavia persistence database user and setting permissions] ***\nTuesday 24 February 2026  01:16:34 +0000 (0:00:02.733)       0:03:57.228 ****** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [octavia : Running Octavia bootstrap container] ***************************\nTuesday 24 February 2026  01:16:37 +0000 (0:00:03.109)       0:04:00.337 ****** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [octavia : Flush handlers] ************************************************\nTuesday 24 February 2026  01:16:58 +0000 (0:00:21.009)       0:04:21.347 ****** \n\nTASK [octavia : Flush handlers] ************************************************\nTuesday 24 February 2026  01:16:59 +0000 (0:00:00.093)       0:04:21.441 ****** \n\nTASK [octavia : Flush handlers] ************************************************\nTuesday 24 February 2026  01:16:59 +0000 (0:00:00.069)       0:04:21.510 ****** \n\nRUNNING HANDLER [octavia : Restart octavia-api container] **********************\nTuesday 24 February 2026  01:16:59 +0000 (0:00:00.109)       0:04:21.619 ****** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nRUNNING HANDLER [octavia : Restart octavia-driver-agent container] *************\nTuesday 24 February 2026  01:17:09 +0000 (0:00:09.924)       0:04:31.544 ****** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\nRUNNING HANDLER [octavia : Restart octavia-health-manager container] ***********\nTuesday 24 February 2026  01:17:15 +0000 (0:00:06.526)       0:04:38.071 ****** \n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\nRUNNING HANDLER [octavia : Restart octavia-housekeeping container] *************\nTuesday 24 February 2026  01:17:26 +0000 (0:00:11.115)       0:04:49.186 ****** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\nRUNNING HANDLER [octavia : Restart octavia-worker container] *******************\nTuesday 24 February 2026  01:17:37 +0000 (0:00:10.444)       0:04:59.630 ****** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-1]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-2]\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=57  \u001b[0m \u001b[0;33mchanged=38  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=7   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=33  \u001b[0m \u001b[0;33mchanged=22  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=5   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=33  \u001b[0m \u001b[0;33mchanged=22  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=5   \u001b[0m rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nTuesday 24 February 2026  01:17:43 +0000 (0:00:06.392)       0:05:06.023 ****** \n=============================================================================== \noctavia : Copying over octavia.conf ------------------------------------ 22.42s\noctavia : Running Octavia bootstrap container -------------------------- 21.01s\noctavia : Adding octavia related roles --------------------------------- 17.37s\noctavia : Add rules for security groups -------------------------------- 15.23s\noctavia : Restart octavia-health-manager container --------------------- 11.12s\noctavia : Create security groups for octavia --------------------------- 11.02s\noctavia : Restart octavia-housekeeping container ----------------------- 10.44s\noctavia : Restart octavia-api container --------------------------------- 9.92s\nservice-ks-register : octavia | Creating users -------------------------- 9.00s\noctavia : Get security groups for octavia ------------------------------- 7.17s\nservice-ks-register : octavia | Granting user roles --------------------- 7.06s\nservice-ks-register : octavia | Creating endpoints ---------------------- 6.86s\noctavia : Copying certificate files for octavia-health-manager ---------- 6.73s\noctavia : Create loadbalancer management subnet ------------------------- 6.70s\noctavia : Copying over config.json files for services ------------------- 6.62s\noctavia : Restart octavia-driver-agent container ------------------------ 6.53s\noctavia : Restart octavia-worker container ------------------------------ 6.39s\noctavia : Copying certificate files for octavia-housekeeping ------------ 6.16s\noctavia : Create ports for Octavia health-manager nodes ----------------- 6.03s\nservice-cert-copy : octavia | Copying over extra CA certificates -------- 6.00s\n\u001b[32m2026-02-24 01:17:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until refresh of running tasks\u001b[0m\n\u001b[32m2026-02-24 01:17:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until refresh of running tasks\u001b[0m\n\u001b[32m2026-02-24 01:17:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until refresh of running tasks\u001b[0m\n\u001b[32m2026-02-24 01:17:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until refresh of running tasks\u001b[0m\n\u001b[32m2026-02-24 01:18:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until refresh of running tasks\u001b[0m\n\u001b[32m2026-02-24 01:18:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until refresh of running tasks\u001b[0m\n\u001b[32m2026-02-24 01:18:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until refresh of running tasks\u001b[0m\n\u001b[32m2026-02-24 01:18:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until refresh of running tasks\u001b[0m\n\u001b[32m2026-02-24 01:18:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until refresh of running tasks\u001b[0m\n\u001b[32m2026-02-24 01:18:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until refresh of running tasks\u001b[0m\n\u001b[32m2026-02-24 01:18:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until refresh of running tasks\u001b[0m\n\u001b[32m2026-02-24 01:18:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until refresh of running tasks\u001b[0m\n\u001b[32m2026-02-24 01:18:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until refresh of running tasks\u001b[0m\n\u001b[32m2026-02-24 01:18:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until refresh of running tasks\u001b[0m\n\u001b[32m2026-02-24 01:18:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until refresh of running tasks\u001b[0m\n\u001b[32m2026-02-24 01:18:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until refresh of running tasks\u001b[0m\n\u001b[32m2026-02-24 01:18:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until refresh of running tasks\u001b[0m\n\u001b[32m2026-02-24 01:18:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until refresh of running tasks\u001b[0m\n\u001b[32m2026-02-24 01:18:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until refresh of running tasks\u001b[0m\n\n\n--> DEPLOY IN A NUTSHELL -- END -- Tue Feb 24 01:18:47 UTC 2026",
                            "stdout_lines": [
                                "+ set -e",
                                "+ source /opt/configuration/scripts/include.sh",
                                "++ export INTERACTIVE=false",
                                "",
                                "# PULL IMAGES",
                                "",
                                "++ INTERACTIVE=false",
                                "++ export OSISM_APPLY_RETRY=1",
                                "++ OSISM_APPLY_RETRY=1",
                                "+ source /opt/manager-vars.sh",
                                "++ export NUMBER_OF_NODES=6",
                                "++ NUMBER_OF_NODES=6",
                                "++ export CEPH_VERSION=reef",
                                "++ CEPH_VERSION=reef",
                                "++ export CONFIGURATION_VERSION=main",
                                "++ CONFIGURATION_VERSION=main",
                                "++ export MANAGER_VERSION=latest",
                                "++ MANAGER_VERSION=latest",
                                "++ export OPENSTACK_VERSION=2024.2",
                                "++ OPENSTACK_VERSION=2024.2",
                                "++ export ARA=false",
                                "++ ARA=false",
                                "++ export DEPLOY_MODE=manager",
                                "++ DEPLOY_MODE=manager",
                                "++ export TEMPEST=true",
                                "++ TEMPEST=true",
                                "++ export IS_ZUUL=true",
                                "++ IS_ZUUL=true",
                                "++ export MANAGER_PUBLIC_IP_ADDRESS=81.163.192.195",
                                "++ MANAGER_PUBLIC_IP_ADDRESS=81.163.192.195",
                                "++ export EXTERNAL_API=false",
                                "++ EXTERNAL_API=false",
                                "++ export IMAGE_USER=ubuntu",
                                "++ IMAGE_USER=ubuntu",
                                "++ export IMAGE_NODE_USER=ubuntu",
                                "++ IMAGE_NODE_USER=ubuntu",
                                "++ export CEPH_STACK=ceph-ansible",
                                "++ CEPH_STACK=ceph-ansible",
                                "+ echo",
                                "+ echo '# PULL IMAGES'",
                                "+ echo",
                                "++ semver latest 7.0.0",
                                "+ [[ -1 -ge 0 ]]",
                                "+ [[ latest == \\l\\a\\t\\e\\s\\t ]]",
                                "+ osism apply --no-wait -r 2 -e custom pull-images",
                                "\u001b[32m2026-02-24 00:42:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTrying to run play pull-images in environment custom\u001b[0m",
                                "\u001b[32m2026-02-24 00:42:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mPrepare task for execution of pull-images.\u001b[0m",
                                "\u001b[32m2026-02-24 00:42:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 0b17a71b-4591-4989-ae1c-e5307c45d989 (pull-images) was prepared for execution.\u001b[0m",
                                "\u001b[32m2026-02-24 00:42:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 0b17a71b-4591-4989-ae1c-e5307c45d989 is running in background. No more output. Check ARA for logs.\u001b[0m",
                                "\u001b[32m2026-02-24 00:42:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTrying to run play wipe-partitions in environment custom\u001b[0m",
                                "\u001b[32m2026-02-24 00:42:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mPrepare task for execution of wipe-partitions.\u001b[0m",
                                "\u001b[32m2026-02-24 00:42:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 94d06f06-2939-4095-b22c-bd35654caca4 (wipe-partitions) was prepared for execution.\u001b[0m",
                                "\u001b[32m2026-02-24 00:42:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task 94d06f06-2939-4095-b22c-bd35654caca4 (wipe-partitions) has been started and output is visible here.\u001b[0m",
                                "",
                                "PLAY [Wipe partitions] *********************************************************",
                                "",
                                "TASK [Find all logical devices owned by UID 167] *******************************",
                                "Tuesday 24 February 2026  00:42:35 +0000 (0:00:00.120)       0:00:00.120 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Remove all rook related logical devices] *********************************",
                                "Tuesday 24 February 2026  00:42:35 +0000 (0:00:00.598)       0:00:00.719 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Find all logical devices with prefix ceph] *******************************",
                                "Tuesday 24 February 2026  00:42:36 +0000 (0:00:00.343)       0:00:01.063 ****** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Remove all ceph related logical devices] *********************************",
                                "Tuesday 24 February 2026  00:42:36 +0000 (0:00:00.557)       0:00:01.620 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Check device availability] ***********************************************",
                                "Tuesday 24 February 2026  00:42:37 +0000 (0:00:00.250)       0:00:01.871 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=/dev/sdb)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=/dev/sdb)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=/dev/sdb)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=/dev/sdc)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=/dev/sdc)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=/dev/sdc)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=/dev/sdd)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=/dev/sdd)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=/dev/sdd)\u001b[0m",
                                "",
                                "TASK [Wipe partitions with wipefs] *********************************************",
                                "Tuesday 24 February 2026  00:42:38 +0000 (0:00:01.290)       0:00:03.162 ****** ",
                                "\u001b[0;32mok: [testbed-node-3] => (item=/dev/sdb)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item=/dev/sdb)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item=/dev/sdb)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item=/dev/sdc)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item=/dev/sdc)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item=/dev/sdc)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item=/dev/sdd)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item=/dev/sdd)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item=/dev/sdd)\u001b[0m",
                                "",
                                "TASK [Overwrite first 32M with zeros] ******************************************",
                                "Tuesday 24 February 2026  00:42:39 +0000 (0:00:01.499)       0:00:04.661 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=/dev/sdb)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=/dev/sdb)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=/dev/sdb)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=/dev/sdc)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=/dev/sdc)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=/dev/sdc)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=/dev/sdd)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=/dev/sdd)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=/dev/sdd)\u001b[0m",
                                "",
                                "TASK [Reload udev rules] *******************************************************",
                                "Tuesday 24 February 2026  00:42:41 +0000 (0:00:02.088)       0:00:06.750 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Request device events from the kernel] ***********************************",
                                "Tuesday 24 February 2026  00:42:42 +0000 (0:00:00.643)       0:00:07.393 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;33mtestbed-node-3\u001b[0m             : \u001b[0;32mok=7   \u001b[0m \u001b[0;33mchanged=5   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-4\u001b[0m             : \u001b[0;32mok=7   \u001b[0m \u001b[0;33mchanged=5   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-5\u001b[0m             : \u001b[0;32mok=7   \u001b[0m \u001b[0;33mchanged=5   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Tuesday 24 February 2026  00:42:43 +0000 (0:00:00.693)       0:00:08.087 ****** ",
                                "=============================================================================== ",
                                "Overwrite first 32M with zeros ------------------------------------------ 2.09s",
                                "Wipe partitions with wipefs --------------------------------------------- 1.50s",
                                "Check device availability ----------------------------------------------- 1.29s",
                                "Request device events from the kernel ----------------------------------- 0.69s",
                                "Reload udev rules ------------------------------------------------------- 0.64s",
                                "Find all logical devices owned by UID 167 ------------------------------- 0.60s",
                                "Find all logical devices with prefix ceph ------------------------------- 0.56s",
                                "Remove all rook related logical devices --------------------------------- 0.34s",
                                "Remove all ceph related logical devices --------------------------------- 0.25s",
                                "\u001b[32m2026-02-24 00:42:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mPrepare task for execution of facts.\u001b[0m",
                                "\u001b[32m2026-02-24 00:42:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3ce96cb6-2139-4792-bb10-c5cbb803795e (facts) was prepared for execution.\u001b[0m",
                                "\u001b[32m2026-02-24 00:42:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task 3ce96cb6-2139-4792-bb10-c5cbb803795e (facts) has been started and output is visible here.\u001b[0m",
                                "",
                                "PLAY [Apply role facts] ********************************************************",
                                "",
                                "TASK [osism.commons.facts : Create custom facts directory] *********************",
                                "Tuesday 24 February 2026  00:43:00 +0000 (0:00:00.263)       0:00:00.263 ****** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.commons.facts : Copy fact files] ***********************************",
                                "Tuesday 24 February 2026  00:43:01 +0000 (0:00:01.079)       0:00:01.342 ****** ",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "PLAY [Gather facts for all hosts] **********************************************",
                                "",
                                "TASK [Gathers facts about hosts] ***********************************************",
                                "Tuesday 24 February 2026  00:43:03 +0000 (0:00:01.231)       0:00:02.574 ****** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "PLAY [Gather facts for all hosts if using --limit] *****************************",
                                "",
                                "TASK [Gather facts for all hosts] **********************************************",
                                "Tuesday 24 February 2026  00:43:08 +0000 (0:00:04.968)       0:00:07.542 ****** ",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;32mtestbed-manager\u001b[0m            : \u001b[0;32mok=2   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;32mtestbed-node-0\u001b[0m             : \u001b[0;32mok=2   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;32mtestbed-node-1\u001b[0m             : \u001b[0;32mok=2   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;32mtestbed-node-2\u001b[0m             : \u001b[0;32mok=2   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;32mtestbed-node-3\u001b[0m             : \u001b[0;32mok=2   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;32mtestbed-node-4\u001b[0m             : \u001b[0;32mok=2   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;32mtestbed-node-5\u001b[0m             : \u001b[0;32mok=2   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Tuesday 24 February 2026  00:43:08 +0000 (0:00:00.521)       0:00:08.064 ****** ",
                                "=============================================================================== ",
                                "Gathers facts about hosts ----------------------------------------------- 4.97s",
                                "osism.commons.facts : Copy fact files ----------------------------------- 1.23s",
                                "osism.commons.facts : Create custom facts directory --------------------- 1.08s",
                                "Gather facts for all hosts ---------------------------------------------- 0.52s",
                                "\u001b[32m2026-02-24 00:43:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mPrepare task for execution of ceph-configure-lvm-volumes.\u001b[0m",
                                "\u001b[32m2026-02-24 00:43:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 567e2c8a-8f76-4540-a3d2-6eaea431f2ce (ceph-configure-lvm-volumes) was prepared for execution.\u001b[0m",
                                "\u001b[32m2026-02-24 00:43:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task 567e2c8a-8f76-4540-a3d2-6eaea431f2ce (ceph-configure-lvm-volumes) has been started and output is visible here.\u001b[0m",
                                "\u001b[1;35m[WARNING]: Collection community.general does not support Ansible version\u001b[0m",
                                "\u001b[1;35m2.16.14\u001b[0m",
                                "",
                                "PLAY [Ceph configure LVM] ******************************************************",
                                "",
                                "TASK [Get extra vars for Ceph configuration] ***********************************",
                                "Tuesday 24 February 2026  00:43:15 +0000 (0:00:00.240)       0:00:00.240 ****** ",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-manager(192.168.16.5)]\u001b[0m",
                                "",
                                "TASK [Get initial list of available block devices] *****************************",
                                "Tuesday 24 February 2026  00:43:15 +0000 (0:00:00.208)       0:00:00.448 ****** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Tuesday 24 February 2026  00:43:15 +0000 (0:00:00.197)       0:00:00.646 ****** ",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-3 => (item=loop0)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-3 => (item=loop1)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-3 => (item=loop2)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-3 => (item=loop3)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-3 => (item=loop4)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-3 => (item=loop5)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-3 => (item=loop6)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-3 => (item=loop7)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-3 => (item=sda)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-3 => (item=sdb)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-3 => (item=sdc)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-3 => (item=sdd)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-3 => (item=sr0)\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Tuesday 24 February 2026  00:43:16 +0000 (0:00:00.403)       0:00:01.050 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Tuesday 24 February 2026  00:43:16 +0000 (0:00:00.191)       0:00:01.241 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Tuesday 24 February 2026  00:43:16 +0000 (0:00:00.156)       0:00:01.397 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Tuesday 24 February 2026  00:43:16 +0000 (0:00:00.185)       0:00:01.583 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Tuesday 24 February 2026  00:43:17 +0000 (0:00:00.180)       0:00:01.764 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Tuesday 24 February 2026  00:43:17 +0000 (0:00:00.186)       0:00:01.950 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Tuesday 24 February 2026  00:43:17 +0000 (0:00:00.194)       0:00:02.145 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Tuesday 24 February 2026  00:43:17 +0000 (0:00:00.176)       0:00:02.321 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Tuesday 24 February 2026  00:43:17 +0000 (0:00:00.242)       0:00:02.564 ****** ",
                                "\u001b[0;32mok: [testbed-node-3] => (item=scsi-0QEMU_QEMU_HARDDISK_f90a4e80-1f36-4d14-b7da-de7b20e9fbca)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item=scsi-SQEMU_QEMU_HARDDISK_f90a4e80-1f36-4d14-b7da-de7b20e9fbca)\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Tuesday 24 February 2026  00:43:18 +0000 (0:00:00.371)       0:00:02.936 ****** ",
                                "\u001b[0;32mok: [testbed-node-3] => (item=scsi-0QEMU_QEMU_HARDDISK_9c6b6763-9704-41ec-9b6d-97eb80a5145c)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item=scsi-SQEMU_QEMU_HARDDISK_9c6b6763-9704-41ec-9b6d-97eb80a5145c)\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Tuesday 24 February 2026  00:43:18 +0000 (0:00:00.576)       0:00:03.512 ****** ",
                                "\u001b[0;32mok: [testbed-node-3] => (item=scsi-0QEMU_QEMU_HARDDISK_6644f191-69e9-4f3a-baff-cffe4152595b)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item=scsi-SQEMU_QEMU_HARDDISK_6644f191-69e9-4f3a-baff-cffe4152595b)\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Tuesday 24 February 2026  00:43:19 +0000 (0:00:00.630)       0:00:04.143 ****** ",
                                "\u001b[0;32mok: [testbed-node-3] => (item=scsi-0QEMU_QEMU_HARDDISK_4534650b-23a0-46c8-8bf1-44520caee637)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item=scsi-SQEMU_QEMU_HARDDISK_4534650b-23a0-46c8-8bf1-44520caee637)\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Tuesday 24 February 2026  00:43:20 +0000 (0:00:00.899)       0:00:05.042 ****** ",
                                "\u001b[0;32mok: [testbed-node-3] => (item=ata-QEMU_DVD-ROM_QM00001)\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Tuesday 24 February 2026  00:43:20 +0000 (0:00:00.330)       0:00:05.372 ****** ",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-3 => (item=loop0)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-3 => (item=loop1)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-3 => (item=loop2)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-3 => (item=loop3)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-3 => (item=loop4)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-3 => (item=loop5)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-3 => (item=loop6)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-3 => (item=loop7)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-3 => (item=sda)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-3 => (item=sdb)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-3 => (item=sdc)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-3 => (item=sdd)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-3 => (item=sr0)\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Tuesday 24 February 2026  00:43:20 +0000 (0:00:00.377)       0:00:05.750 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Tuesday 24 February 2026  00:43:21 +0000 (0:00:00.237)       0:00:05.988 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Tuesday 24 February 2026  00:43:21 +0000 (0:00:00.206)       0:00:06.195 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Tuesday 24 February 2026  00:43:21 +0000 (0:00:00.214)       0:00:06.409 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Tuesday 24 February 2026  00:43:21 +0000 (0:00:00.197)       0:00:06.606 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Tuesday 24 February 2026  00:43:22 +0000 (0:00:00.209)       0:00:06.816 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Tuesday 24 February 2026  00:43:22 +0000 (0:00:00.239)       0:00:07.056 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Tuesday 24 February 2026  00:43:22 +0000 (0:00:00.193)       0:00:07.249 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Tuesday 24 February 2026  00:43:22 +0000 (0:00:00.196)       0:00:07.446 ****** ",
                                "\u001b[0;32mok: [testbed-node-3] => (item=sda1)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item=sda14)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item=sda15)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item=sda16)\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Tuesday 24 February 2026  00:43:23 +0000 (0:00:01.088)       0:00:08.534 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Tuesday 24 February 2026  00:43:23 +0000 (0:00:00.209)       0:00:08.744 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Tuesday 24 February 2026  00:43:24 +0000 (0:00:00.229)       0:00:08.974 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Tuesday 24 February 2026  00:43:24 +0000 (0:00:00.205)       0:00:09.180 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Set UUIDs for OSD VGs/LVs] ***********************************************",
                                "Tuesday 24 February 2026  00:43:24 +0000 (0:00:00.203)       0:00:09.384 ****** ",
                                "\u001b[0;32mok: [testbed-node-3] => (item={'key': 'sdb', 'value': None})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item={'key': 'sdc', 'value': None})\u001b[0m",
                                "",
                                "TASK [Generate WAL VG names] ***************************************************",
                                "Tuesday 24 February 2026  00:43:24 +0000 (0:00:00.222)       0:00:09.606 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Generate DB VG names] ****************************************************",
                                "Tuesday 24 February 2026  00:43:24 +0000 (0:00:00.144)       0:00:09.751 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Generate shared DB/WAL VG names] *****************************************",
                                "Tuesday 24 February 2026  00:43:25 +0000 (0:00:00.156)       0:00:09.907 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Define lvm_volumes structures] *******************************************",
                                "Tuesday 24 February 2026  00:43:25 +0000 (0:00:00.160)       0:00:10.068 ****** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Generate lvm_volumes structure (block only)] *****************************",
                                "Tuesday 24 February 2026  00:43:25 +0000 (0:00:00.125)       0:00:10.193 ****** ",
                                "\u001b[0;32mok: [testbed-node-3] => (item={'key': 'sdb', 'value': {'osd_lvm_uuid': '9b92f24c-dac1-5bff-a390-34d9ff9e0f4f'}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item={'key': 'sdc', 'value': {'osd_lvm_uuid': '7f6a8fe5-6adc-52b4-b256-edcdbfd9bf4e'}})\u001b[0m",
                                "",
                                "TASK [Generate lvm_volumes structure (block + db)] *****************************",
                                "Tuesday 24 February 2026  00:43:25 +0000 (0:00:00.162)       0:00:10.356 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'sdb', 'value': {'osd_lvm_uuid': '9b92f24c-dac1-5bff-a390-34d9ff9e0f4f'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'sdc', 'value': {'osd_lvm_uuid': '7f6a8fe5-6adc-52b4-b256-edcdbfd9bf4e'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Generate lvm_volumes structure (block + wal)] ****************************",
                                "Tuesday 24 February 2026  00:43:25 +0000 (0:00:00.157)       0:00:10.513 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'sdb', 'value': {'osd_lvm_uuid': '9b92f24c-dac1-5bff-a390-34d9ff9e0f4f'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'sdc', 'value': {'osd_lvm_uuid': '7f6a8fe5-6adc-52b4-b256-edcdbfd9bf4e'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Generate lvm_volumes structure (block + db + wal)] ***********************",
                                "Tuesday 24 February 2026  00:43:26 +0000 (0:00:00.386)       0:00:10.899 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'sdb', 'value': {'osd_lvm_uuid': '9b92f24c-dac1-5bff-a390-34d9ff9e0f4f'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'sdc', 'value': {'osd_lvm_uuid': '7f6a8fe5-6adc-52b4-b256-edcdbfd9bf4e'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Compile lvm_volumes] *****************************************************",
                                "Tuesday 24 February 2026  00:43:26 +0000 (0:00:00.151)       0:00:11.051 ****** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Set OSD devices config data] *********************************************",
                                "Tuesday 24 February 2026  00:43:26 +0000 (0:00:00.142)       0:00:11.194 ****** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Set DB devices config data] **********************************************",
                                "Tuesday 24 February 2026  00:43:26 +0000 (0:00:00.119)       0:00:11.313 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Set WAL devices config data] *********************************************",
                                "Tuesday 24 February 2026  00:43:26 +0000 (0:00:00.113)       0:00:11.427 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Set DB+WAL devices config data] ******************************************",
                                "Tuesday 24 February 2026  00:43:26 +0000 (0:00:00.141)       0:00:11.569 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Print ceph_osd_devices] **************************************************",
                                "Tuesday 24 February 2026  00:43:26 +0000 (0:00:00.134)       0:00:11.703 ****** ",
                                "\u001b[0;32mok: [testbed-node-3] => {\u001b[0m",
                                "\u001b[0;32m    \"ceph_osd_devices\": {\u001b[0m",
                                "\u001b[0;32m        \"sdb\": {\u001b[0m",
                                "\u001b[0;32m            \"osd_lvm_uuid\": \"9b92f24c-dac1-5bff-a390-34d9ff9e0f4f\"\u001b[0m",
                                "\u001b[0;32m        },\u001b[0m",
                                "\u001b[0;32m        \"sdc\": {\u001b[0m",
                                "\u001b[0;32m            \"osd_lvm_uuid\": \"7f6a8fe5-6adc-52b4-b256-edcdbfd9bf4e\"\u001b[0m",
                                "\u001b[0;32m        }\u001b[0m",
                                "\u001b[0;32m    }\u001b[0m",
                                "\u001b[0;32m}\u001b[0m",
                                "",
                                "TASK [Print WAL devices] *******************************************************",
                                "Tuesday 24 February 2026  00:43:27 +0000 (0:00:00.153)       0:00:11.857 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Print DB devices] ********************************************************",
                                "Tuesday 24 February 2026  00:43:27 +0000 (0:00:00.138)       0:00:11.995 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Print shared DB/WAL devices] *********************************************",
                                "Tuesday 24 February 2026  00:43:27 +0000 (0:00:00.130)       0:00:12.126 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Print configuration data] ************************************************",
                                "Tuesday 24 February 2026  00:43:27 +0000 (0:00:00.144)       0:00:12.270 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-3] => {\u001b[0m",
                                "\u001b[0;33m    \"_ceph_configure_lvm_config_data\": {\u001b[0m",
                                "\u001b[0;33m        \"ceph_osd_devices\": {\u001b[0m",
                                "\u001b[0;33m            \"sdb\": {\u001b[0m",
                                "\u001b[0;33m                \"osd_lvm_uuid\": \"9b92f24c-dac1-5bff-a390-34d9ff9e0f4f\"\u001b[0m",
                                "\u001b[0;33m            },\u001b[0m",
                                "\u001b[0;33m            \"sdc\": {\u001b[0m",
                                "\u001b[0;33m                \"osd_lvm_uuid\": \"7f6a8fe5-6adc-52b4-b256-edcdbfd9bf4e\"\u001b[0m",
                                "\u001b[0;33m            }\u001b[0m",
                                "\u001b[0;33m        },\u001b[0m",
                                "\u001b[0;33m        \"lvm_volumes\": [\u001b[0m",
                                "\u001b[0;33m            {\u001b[0m",
                                "\u001b[0;33m                \"data\": \"osd-block-9b92f24c-dac1-5bff-a390-34d9ff9e0f4f\",\u001b[0m",
                                "\u001b[0;33m                \"data_vg\": \"ceph-9b92f24c-dac1-5bff-a390-34d9ff9e0f4f\"\u001b[0m",
                                "\u001b[0;33m            },\u001b[0m",
                                "\u001b[0;33m            {\u001b[0m",
                                "\u001b[0;33m                \"data\": \"osd-block-7f6a8fe5-6adc-52b4-b256-edcdbfd9bf4e\",\u001b[0m",
                                "\u001b[0;33m                \"data_vg\": \"ceph-7f6a8fe5-6adc-52b4-b256-edcdbfd9bf4e\"\u001b[0m",
                                "\u001b[0;33m            }\u001b[0m",
                                "\u001b[0;33m        ]\u001b[0m",
                                "\u001b[0;33m    }\u001b[0m",
                                "\u001b[0;33m}\u001b[0m",
                                "",
                                "RUNNING HANDLER [Write configuration file] *************************************",
                                "Tuesday 24 February 2026  00:43:27 +0000 (0:00:00.416)       0:00:12.687 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-3 -> testbed-manager(192.168.16.5)]\u001b[0m",
                                "",
                                "PLAY [Ceph configure LVM] ******************************************************",
                                "",
                                "TASK [Get extra vars for Ceph configuration] ***********************************",
                                "Tuesday 24 February 2026  00:43:29 +0000 (0:00:01.887)       0:00:14.574 ****** ",
                                "\u001b[0;32mok: [testbed-node-4 -> testbed-manager(192.168.16.5)]\u001b[0m",
                                "",
                                "TASK [Get initial list of available block devices] *****************************",
                                "Tuesday 24 February 2026  00:43:30 +0000 (0:00:00.276)       0:00:14.851 ****** ",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Tuesday 24 February 2026  00:43:30 +0000 (0:00:00.248)       0:00:15.100 ****** ",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-4 => (item=loop0)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-4 => (item=loop1)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-4 => (item=loop2)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-4 => (item=loop3)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-4 => (item=loop4)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-4 => (item=loop5)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-4 => (item=loop6)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-4 => (item=loop7)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-4 => (item=sda)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-4 => (item=sdb)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-4 => (item=sdc)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-4 => (item=sdd)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-4 => (item=sr0)\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Tuesday 24 February 2026  00:43:30 +0000 (0:00:00.395)       0:00:15.496 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Tuesday 24 February 2026  00:43:30 +0000 (0:00:00.231)       0:00:15.727 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Tuesday 24 February 2026  00:43:31 +0000 (0:00:00.210)       0:00:15.937 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Tuesday 24 February 2026  00:43:31 +0000 (0:00:00.199)       0:00:16.136 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Tuesday 24 February 2026  00:43:31 +0000 (0:00:00.249)       0:00:16.385 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Tuesday 24 February 2026  00:43:32 +0000 (0:00:00.693)       0:00:17.079 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Tuesday 24 February 2026  00:43:32 +0000 (0:00:00.211)       0:00:17.290 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Tuesday 24 February 2026  00:43:32 +0000 (0:00:00.222)       0:00:17.513 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Tuesday 24 February 2026  00:43:33 +0000 (0:00:00.261)       0:00:17.774 ****** ",
                                "\u001b[0;32mok: [testbed-node-4] => (item=scsi-0QEMU_QEMU_HARDDISK_47e600ec-05ca-44a6-93d0-5ca3104e3639)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item=scsi-SQEMU_QEMU_HARDDISK_47e600ec-05ca-44a6-93d0-5ca3104e3639)\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Tuesday 24 February 2026  00:43:33 +0000 (0:00:00.517)       0:00:18.292 ****** ",
                                "\u001b[0;32mok: [testbed-node-4] => (item=scsi-0QEMU_QEMU_HARDDISK_f659b61f-c39f-426f-a640-6a753951661d)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item=scsi-SQEMU_QEMU_HARDDISK_f659b61f-c39f-426f-a640-6a753951661d)\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Tuesday 24 February 2026  00:43:33 +0000 (0:00:00.447)       0:00:18.739 ****** ",
                                "\u001b[0;32mok: [testbed-node-4] => (item=scsi-0QEMU_QEMU_HARDDISK_fb58bf16-a1c6-4abb-925b-29c6a6a6f100)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item=scsi-SQEMU_QEMU_HARDDISK_fb58bf16-a1c6-4abb-925b-29c6a6a6f100)\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Tuesday 24 February 2026  00:43:34 +0000 (0:00:00.459)       0:00:19.198 ****** ",
                                "\u001b[0;32mok: [testbed-node-4] => (item=scsi-0QEMU_QEMU_HARDDISK_5660bb62-e2c9-45b2-ae11-8ab0f829b194)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item=scsi-SQEMU_QEMU_HARDDISK_5660bb62-e2c9-45b2-ae11-8ab0f829b194)\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Tuesday 24 February 2026  00:43:34 +0000 (0:00:00.445)       0:00:19.643 ****** ",
                                "\u001b[0;32mok: [testbed-node-4] => (item=ata-QEMU_DVD-ROM_QM00001)\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Tuesday 24 February 2026  00:43:35 +0000 (0:00:00.382)       0:00:20.025 ****** ",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-4 => (item=loop0)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-4 => (item=loop1)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-4 => (item=loop2)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-4 => (item=loop3)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-4 => (item=loop4)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-4 => (item=loop5)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-4 => (item=loop6)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-4 => (item=loop7)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-4 => (item=sda)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-4 => (item=sdb)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-4 => (item=sdc)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-4 => (item=sdd)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-4 => (item=sr0)\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Tuesday 24 February 2026  00:43:35 +0000 (0:00:00.492)       0:00:20.518 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Tuesday 24 February 2026  00:43:36 +0000 (0:00:00.848)       0:00:21.366 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Tuesday 24 February 2026  00:43:36 +0000 (0:00:00.212)       0:00:21.579 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Tuesday 24 February 2026  00:43:37 +0000 (0:00:00.222)       0:00:21.801 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Tuesday 24 February 2026  00:43:37 +0000 (0:00:00.208)       0:00:22.009 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Tuesday 24 February 2026  00:43:37 +0000 (0:00:00.187)       0:00:22.197 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Tuesday 24 February 2026  00:43:37 +0000 (0:00:00.197)       0:00:22.394 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Tuesday 24 February 2026  00:43:37 +0000 (0:00:00.188)       0:00:22.583 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Tuesday 24 February 2026  00:43:38 +0000 (0:00:00.193)       0:00:22.776 ****** ",
                                "\u001b[0;32mok: [testbed-node-4] => (item=sda1)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item=sda14)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item=sda15)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item=sda16)\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Tuesday 24 February 2026  00:43:38 +0000 (0:00:00.902)       0:00:23.679 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Tuesday 24 February 2026  00:43:39 +0000 (0:00:00.188)       0:00:23.867 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Tuesday 24 February 2026  00:43:39 +0000 (0:00:00.203)       0:00:24.070 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Tuesday 24 February 2026  00:43:39 +0000 (0:00:00.190)       0:00:24.261 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Set UUIDs for OSD VGs/LVs] ***********************************************",
                                "Tuesday 24 February 2026  00:43:40 +0000 (0:00:00.742)       0:00:25.003 ****** ",
                                "\u001b[0;32mok: [testbed-node-4] => (item={'key': 'sdb', 'value': None})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item={'key': 'sdc', 'value': None})\u001b[0m",
                                "",
                                "TASK [Generate WAL VG names] ***************************************************",
                                "Tuesday 24 February 2026  00:43:40 +0000 (0:00:00.175)       0:00:25.179 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Generate DB VG names] ****************************************************",
                                "Tuesday 24 February 2026  00:43:40 +0000 (0:00:00.149)       0:00:25.328 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Generate shared DB/WAL VG names] *****************************************",
                                "Tuesday 24 February 2026  00:43:40 +0000 (0:00:00.137)       0:00:25.466 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Define lvm_volumes structures] *******************************************",
                                "Tuesday 24 February 2026  00:43:40 +0000 (0:00:00.134)       0:00:25.600 ****** ",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Generate lvm_volumes structure (block only)] *****************************",
                                "Tuesday 24 February 2026  00:43:40 +0000 (0:00:00.150)       0:00:25.750 ****** ",
                                "\u001b[0;32mok: [testbed-node-4] => (item={'key': 'sdb', 'value': {'osd_lvm_uuid': '294dd48f-9d6d-5079-8214-59e60af9e526'}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item={'key': 'sdc', 'value': {'osd_lvm_uuid': '7fbf4d63-7c6f-5c70-887c-2b73c978775b'}})\u001b[0m",
                                "",
                                "TASK [Generate lvm_volumes structure (block + db)] *****************************",
                                "Tuesday 24 February 2026  00:43:41 +0000 (0:00:00.145)       0:00:25.896 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'sdb', 'value': {'osd_lvm_uuid': '294dd48f-9d6d-5079-8214-59e60af9e526'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'sdc', 'value': {'osd_lvm_uuid': '7fbf4d63-7c6f-5c70-887c-2b73c978775b'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Generate lvm_volumes structure (block + wal)] ****************************",
                                "Tuesday 24 February 2026  00:43:41 +0000 (0:00:00.129)       0:00:26.026 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'sdb', 'value': {'osd_lvm_uuid': '294dd48f-9d6d-5079-8214-59e60af9e526'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'sdc', 'value': {'osd_lvm_uuid': '7fbf4d63-7c6f-5c70-887c-2b73c978775b'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Generate lvm_volumes structure (block + db + wal)] ***********************",
                                "Tuesday 24 February 2026  00:43:41 +0000 (0:00:00.153)       0:00:26.179 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'sdb', 'value': {'osd_lvm_uuid': '294dd48f-9d6d-5079-8214-59e60af9e526'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'sdc', 'value': {'osd_lvm_uuid': '7fbf4d63-7c6f-5c70-887c-2b73c978775b'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Compile lvm_volumes] *****************************************************",
                                "Tuesday 24 February 2026  00:43:41 +0000 (0:00:00.133)       0:00:26.313 ****** ",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Set OSD devices config data] *********************************************",
                                "Tuesday 24 February 2026  00:43:41 +0000 (0:00:00.114)       0:00:26.427 ****** ",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Set DB devices config data] **********************************************",
                                "Tuesday 24 February 2026  00:43:41 +0000 (0:00:00.133)       0:00:26.561 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Set WAL devices config data] *********************************************",
                                "Tuesday 24 February 2026  00:43:42 +0000 (0:00:00.238)       0:00:26.799 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Set DB+WAL devices config data] ******************************************",
                                "Tuesday 24 February 2026  00:43:42 +0000 (0:00:00.116)       0:00:26.916 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Print ceph_osd_devices] **************************************************",
                                "Tuesday 24 February 2026  00:43:42 +0000 (0:00:00.128)       0:00:27.044 ****** ",
                                "\u001b[0;32mok: [testbed-node-4] => {\u001b[0m",
                                "\u001b[0;32m    \"ceph_osd_devices\": {\u001b[0m",
                                "\u001b[0;32m        \"sdb\": {\u001b[0m",
                                "\u001b[0;32m            \"osd_lvm_uuid\": \"294dd48f-9d6d-5079-8214-59e60af9e526\"\u001b[0m",
                                "\u001b[0;32m        },\u001b[0m",
                                "\u001b[0;32m        \"sdc\": {\u001b[0m",
                                "\u001b[0;32m            \"osd_lvm_uuid\": \"7fbf4d63-7c6f-5c70-887c-2b73c978775b\"\u001b[0m",
                                "\u001b[0;32m        }\u001b[0m",
                                "\u001b[0;32m    }\u001b[0m",
                                "\u001b[0;32m}\u001b[0m",
                                "",
                                "TASK [Print WAL devices] *******************************************************",
                                "Tuesday 24 February 2026  00:43:42 +0000 (0:00:00.142)       0:00:27.187 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Print DB devices] ********************************************************",
                                "Tuesday 24 February 2026  00:43:42 +0000 (0:00:00.119)       0:00:27.307 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Print shared DB/WAL devices] *********************************************",
                                "Tuesday 24 February 2026  00:43:42 +0000 (0:00:00.122)       0:00:27.429 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Print configuration data] ************************************************",
                                "Tuesday 24 February 2026  00:43:42 +0000 (0:00:00.103)       0:00:27.532 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-4] => {\u001b[0m",
                                "\u001b[0;33m    \"_ceph_configure_lvm_config_data\": {\u001b[0m",
                                "\u001b[0;33m        \"ceph_osd_devices\": {\u001b[0m",
                                "\u001b[0;33m            \"sdb\": {\u001b[0m",
                                "\u001b[0;33m                \"osd_lvm_uuid\": \"294dd48f-9d6d-5079-8214-59e60af9e526\"\u001b[0m",
                                "\u001b[0;33m            },\u001b[0m",
                                "\u001b[0;33m            \"sdc\": {\u001b[0m",
                                "\u001b[0;33m                \"osd_lvm_uuid\": \"7fbf4d63-7c6f-5c70-887c-2b73c978775b\"\u001b[0m",
                                "\u001b[0;33m            }\u001b[0m",
                                "\u001b[0;33m        },\u001b[0m",
                                "\u001b[0;33m        \"lvm_volumes\": [\u001b[0m",
                                "\u001b[0;33m            {\u001b[0m",
                                "\u001b[0;33m                \"data\": \"osd-block-294dd48f-9d6d-5079-8214-59e60af9e526\",\u001b[0m",
                                "\u001b[0;33m                \"data_vg\": \"ceph-294dd48f-9d6d-5079-8214-59e60af9e526\"\u001b[0m",
                                "\u001b[0;33m            },\u001b[0m",
                                "\u001b[0;33m            {\u001b[0m",
                                "\u001b[0;33m                \"data\": \"osd-block-7fbf4d63-7c6f-5c70-887c-2b73c978775b\",\u001b[0m",
                                "\u001b[0;33m                \"data_vg\": \"ceph-7fbf4d63-7c6f-5c70-887c-2b73c978775b\"\u001b[0m",
                                "\u001b[0;33m            }\u001b[0m",
                                "\u001b[0;33m        ]\u001b[0m",
                                "\u001b[0;33m    }\u001b[0m",
                                "\u001b[0;33m}\u001b[0m",
                                "",
                                "RUNNING HANDLER [Write configuration file] *************************************",
                                "Tuesday 24 February 2026  00:43:42 +0000 (0:00:00.180)       0:00:27.713 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-4 -> testbed-manager(192.168.16.5)]\u001b[0m",
                                "",
                                "PLAY [Ceph configure LVM] ******************************************************",
                                "",
                                "TASK [Get extra vars for Ceph configuration] ***********************************",
                                "Tuesday 24 February 2026  00:43:43 +0000 (0:00:01.006)       0:00:28.719 ****** ",
                                "\u001b[0;32mok: [testbed-node-5 -> testbed-manager(192.168.16.5)]\u001b[0m",
                                "",
                                "TASK [Get initial list of available block devices] *****************************",
                                "Tuesday 24 February 2026  00:43:44 +0000 (0:00:00.479)       0:00:29.198 ****** ",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Tuesday 24 February 2026  00:43:44 +0000 (0:00:00.203)       0:00:29.402 ****** ",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-5 => (item=loop0)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-5 => (item=loop1)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-5 => (item=loop2)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-5 => (item=loop3)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-5 => (item=loop4)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-5 => (item=loop5)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-5 => (item=loop6)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-5 => (item=loop7)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-5 => (item=sda)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-5 => (item=sdb)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-5 => (item=sdc)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-5 => (item=sdd)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-5 => (item=sr0)\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Tuesday 24 February 2026  00:43:44 +0000 (0:00:00.334)       0:00:29.737 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Tuesday 24 February 2026  00:43:45 +0000 (0:00:00.179)       0:00:29.917 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Tuesday 24 February 2026  00:43:45 +0000 (0:00:00.172)       0:00:30.089 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Tuesday 24 February 2026  00:43:45 +0000 (0:00:00.137)       0:00:30.227 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Tuesday 24 February 2026  00:43:45 +0000 (0:00:00.131)       0:00:30.358 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Tuesday 24 February 2026  00:43:45 +0000 (0:00:00.192)       0:00:30.551 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Tuesday 24 February 2026  00:43:45 +0000 (0:00:00.191)       0:00:30.743 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Tuesday 24 February 2026  00:43:46 +0000 (0:00:00.171)       0:00:30.914 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Tuesday 24 February 2026  00:43:46 +0000 (0:00:00.188)       0:00:31.103 ****** ",
                                "\u001b[0;32mok: [testbed-node-5] => (item=scsi-0QEMU_QEMU_HARDDISK_7a9487cb-ec9e-4508-9c5a-f692700fc5a4)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item=scsi-SQEMU_QEMU_HARDDISK_7a9487cb-ec9e-4508-9c5a-f692700fc5a4)\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Tuesday 24 February 2026  00:43:47 +0000 (0:00:00.659)       0:00:31.763 ****** ",
                                "\u001b[0;32mok: [testbed-node-5] => (item=scsi-0QEMU_QEMU_HARDDISK_1558935b-4c66-4614-97e5-cda14abd2641)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item=scsi-SQEMU_QEMU_HARDDISK_1558935b-4c66-4614-97e5-cda14abd2641)\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Tuesday 24 February 2026  00:43:47 +0000 (0:00:00.453)       0:00:32.217 ****** ",
                                "\u001b[0;32mok: [testbed-node-5] => (item=scsi-0QEMU_QEMU_HARDDISK_6f4643f0-329b-4ae2-97c7-d348648fd970)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item=scsi-SQEMU_QEMU_HARDDISK_6f4643f0-329b-4ae2-97c7-d348648fd970)\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Tuesday 24 February 2026  00:43:47 +0000 (0:00:00.376)       0:00:32.593 ****** ",
                                "\u001b[0;32mok: [testbed-node-5] => (item=scsi-0QEMU_QEMU_HARDDISK_5f47789f-2a1a-442a-8a3c-644a4f92e5a8)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item=scsi-SQEMU_QEMU_HARDDISK_5f47789f-2a1a-442a-8a3c-644a4f92e5a8)\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Tuesday 24 February 2026  00:43:48 +0000 (0:00:00.403)       0:00:32.997 ****** ",
                                "\u001b[0;32mok: [testbed-node-5] => (item=ata-QEMU_DVD-ROM_QM00001)\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Tuesday 24 February 2026  00:43:48 +0000 (0:00:00.311)       0:00:33.308 ****** ",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-5 => (item=loop0)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-5 => (item=loop1)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-5 => (item=loop2)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-5 => (item=loop3)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-5 => (item=loop4)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-5 => (item=loop5)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-5 => (item=loop6)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-5 => (item=loop7)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-5 => (item=sda)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-5 => (item=sdb)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-5 => (item=sdc)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-5 => (item=sdd)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-5 => (item=sr0)\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Tuesday 24 February 2026  00:43:48 +0000 (0:00:00.354)       0:00:33.662 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Tuesday 24 February 2026  00:43:49 +0000 (0:00:00.168)       0:00:33.830 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Tuesday 24 February 2026  00:43:49 +0000 (0:00:00.169)       0:00:33.999 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Tuesday 24 February 2026  00:43:49 +0000 (0:00:00.172)       0:00:34.172 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Tuesday 24 February 2026  00:43:49 +0000 (0:00:00.148)       0:00:34.321 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Tuesday 24 February 2026  00:43:49 +0000 (0:00:00.134)       0:00:34.455 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Tuesday 24 February 2026  00:43:50 +0000 (0:00:00.402)       0:00:34.857 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Tuesday 24 February 2026  00:43:50 +0000 (0:00:00.163)       0:00:35.021 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Tuesday 24 February 2026  00:43:50 +0000 (0:00:00.178)       0:00:35.199 ****** ",
                                "\u001b[0;32mok: [testbed-node-5] => (item=sda1)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item=sda14)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item=sda15)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item=sda16)\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Tuesday 24 February 2026  00:43:51 +0000 (0:00:00.564)       0:00:35.764 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Tuesday 24 February 2026  00:43:51 +0000 (0:00:00.149)       0:00:35.914 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Tuesday 24 February 2026  00:43:51 +0000 (0:00:00.187)       0:00:36.101 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Tuesday 24 February 2026  00:43:51 +0000 (0:00:00.191)       0:00:36.292 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Set UUIDs for OSD VGs/LVs] ***********************************************",
                                "Tuesday 24 February 2026  00:43:51 +0000 (0:00:00.181)       0:00:36.473 ****** ",
                                "\u001b[0;32mok: [testbed-node-5] => (item={'key': 'sdb', 'value': None})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item={'key': 'sdc', 'value': None})\u001b[0m",
                                "",
                                "TASK [Generate WAL VG names] ***************************************************",
                                "Tuesday 24 February 2026  00:43:51 +0000 (0:00:00.153)       0:00:36.626 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Generate DB VG names] ****************************************************",
                                "Tuesday 24 February 2026  00:43:51 +0000 (0:00:00.116)       0:00:36.743 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Generate shared DB/WAL VG names] *****************************************",
                                "Tuesday 24 February 2026  00:43:52 +0000 (0:00:00.104)       0:00:36.848 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Define lvm_volumes structures] *******************************************",
                                "Tuesday 24 February 2026  00:43:52 +0000 (0:00:00.302)       0:00:37.151 ****** ",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Generate lvm_volumes structure (block only)] *****************************",
                                "Tuesday 24 February 2026  00:43:52 +0000 (0:00:00.126)       0:00:37.277 ****** ",
                                "\u001b[0;32mok: [testbed-node-5] => (item={'key': 'sdb', 'value': {'osd_lvm_uuid': '752a9ec4-4bfc-5551-82cd-92b802bdad97'}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item={'key': 'sdc', 'value': {'osd_lvm_uuid': 'ba85a0e3-2f57-51d5-a9f1-072f0e9a8deb'}})\u001b[0m",
                                "",
                                "TASK [Generate lvm_volumes structure (block + db)] *****************************",
                                "Tuesday 24 February 2026  00:43:52 +0000 (0:00:00.146)       0:00:37.424 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'sdb', 'value': {'osd_lvm_uuid': '752a9ec4-4bfc-5551-82cd-92b802bdad97'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'sdc', 'value': {'osd_lvm_uuid': 'ba85a0e3-2f57-51d5-a9f1-072f0e9a8deb'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Generate lvm_volumes structure (block + wal)] ****************************",
                                "Tuesday 24 February 2026  00:43:52 +0000 (0:00:00.134)       0:00:37.558 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'sdb', 'value': {'osd_lvm_uuid': '752a9ec4-4bfc-5551-82cd-92b802bdad97'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'sdc', 'value': {'osd_lvm_uuid': 'ba85a0e3-2f57-51d5-a9f1-072f0e9a8deb'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Generate lvm_volumes structure (block + db + wal)] ***********************",
                                "Tuesday 24 February 2026  00:43:52 +0000 (0:00:00.136)       0:00:37.695 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'sdb', 'value': {'osd_lvm_uuid': '752a9ec4-4bfc-5551-82cd-92b802bdad97'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'sdc', 'value': {'osd_lvm_uuid': 'ba85a0e3-2f57-51d5-a9f1-072f0e9a8deb'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Compile lvm_volumes] *****************************************************",
                                "Tuesday 24 February 2026  00:43:53 +0000 (0:00:00.140)       0:00:37.835 ****** ",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Set OSD devices config data] *********************************************",
                                "Tuesday 24 February 2026  00:43:53 +0000 (0:00:00.114)       0:00:37.950 ****** ",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Set DB devices config data] **********************************************",
                                "Tuesday 24 February 2026  00:43:53 +0000 (0:00:00.116)       0:00:38.066 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Set WAL devices config data] *********************************************",
                                "Tuesday 24 February 2026  00:43:53 +0000 (0:00:00.148)       0:00:38.215 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Set DB+WAL devices config data] ******************************************",
                                "Tuesday 24 February 2026  00:43:53 +0000 (0:00:00.116)       0:00:38.331 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Print ceph_osd_devices] **************************************************",
                                "Tuesday 24 February 2026  00:43:53 +0000 (0:00:00.113)       0:00:38.445 ****** ",
                                "\u001b[0;32mok: [testbed-node-5] => {\u001b[0m",
                                "\u001b[0;32m    \"ceph_osd_devices\": {\u001b[0m",
                                "\u001b[0;32m        \"sdb\": {\u001b[0m",
                                "\u001b[0;32m            \"osd_lvm_uuid\": \"752a9ec4-4bfc-5551-82cd-92b802bdad97\"\u001b[0m",
                                "\u001b[0;32m        },\u001b[0m",
                                "\u001b[0;32m        \"sdc\": {\u001b[0m",
                                "\u001b[0;32m            \"osd_lvm_uuid\": \"ba85a0e3-2f57-51d5-a9f1-072f0e9a8deb\"\u001b[0m",
                                "\u001b[0;32m        }\u001b[0m",
                                "\u001b[0;32m    }\u001b[0m",
                                "\u001b[0;32m}\u001b[0m",
                                "",
                                "TASK [Print WAL devices] *******************************************************",
                                "Tuesday 24 February 2026  00:43:53 +0000 (0:00:00.127)       0:00:38.573 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Print DB devices] ********************************************************",
                                "Tuesday 24 February 2026  00:43:54 +0000 (0:00:00.307)       0:00:38.880 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Print shared DB/WAL devices] *********************************************",
                                "Tuesday 24 February 2026  00:43:54 +0000 (0:00:00.131)       0:00:39.011 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Print configuration data] ************************************************",
                                "Tuesday 24 February 2026  00:43:54 +0000 (0:00:00.121)       0:00:39.133 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-5] => {\u001b[0m",
                                "\u001b[0;33m    \"_ceph_configure_lvm_config_data\": {\u001b[0m",
                                "\u001b[0;33m        \"ceph_osd_devices\": {\u001b[0m",
                                "\u001b[0;33m            \"sdb\": {\u001b[0m",
                                "\u001b[0;33m                \"osd_lvm_uuid\": \"752a9ec4-4bfc-5551-82cd-92b802bdad97\"\u001b[0m",
                                "\u001b[0;33m            },\u001b[0m",
                                "\u001b[0;33m            \"sdc\": {\u001b[0m",
                                "\u001b[0;33m                \"osd_lvm_uuid\": \"ba85a0e3-2f57-51d5-a9f1-072f0e9a8deb\"\u001b[0m",
                                "\u001b[0;33m            }\u001b[0m",
                                "\u001b[0;33m        },\u001b[0m",
                                "\u001b[0;33m        \"lvm_volumes\": [\u001b[0m",
                                "\u001b[0;33m            {\u001b[0m",
                                "\u001b[0;33m                \"data\": \"osd-block-752a9ec4-4bfc-5551-82cd-92b802bdad97\",\u001b[0m",
                                "\u001b[0;33m                \"data_vg\": \"ceph-752a9ec4-4bfc-5551-82cd-92b802bdad97\"\u001b[0m",
                                "\u001b[0;33m            },\u001b[0m",
                                "\u001b[0;33m            {\u001b[0m",
                                "\u001b[0;33m                \"data\": \"osd-block-ba85a0e3-2f57-51d5-a9f1-072f0e9a8deb\",\u001b[0m",
                                "\u001b[0;33m                \"data_vg\": \"ceph-ba85a0e3-2f57-51d5-a9f1-072f0e9a8deb\"\u001b[0m",
                                "\u001b[0;33m            }\u001b[0m",
                                "\u001b[0;33m        ]\u001b[0m",
                                "\u001b[0;33m    }\u001b[0m",
                                "\u001b[0;33m}\u001b[0m",
                                "",
                                "RUNNING HANDLER [Write configuration file] *************************************",
                                "Tuesday 24 February 2026  00:43:54 +0000 (0:00:00.190)       0:00:39.323 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-5 -> testbed-manager(192.168.16.5)]\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;33mtestbed-node-3\u001b[0m             : \u001b[0;32mok=42  \u001b[0m \u001b[0;33mchanged=2   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=32  \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-4\u001b[0m             : \u001b[0;32mok=42  \u001b[0m \u001b[0;33mchanged=2   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=32  \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-5\u001b[0m             : \u001b[0;32mok=42  \u001b[0m \u001b[0;33mchanged=2   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=32  \u001b[0m rescued=0    ignored=0   ",
                                "",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Tuesday 24 February 2026  00:43:55 +0000 (0:00:00.848)       0:00:40.172 ****** ",
                                "=============================================================================== ",
                                "Write configuration file ------------------------------------------------ 3.74s",
                                "Add known partitions to the list of available block devices ------------- 1.22s",
                                "Add known links to the list of available block devices ------------------ 1.13s",
                                "Add known partitions to the list of available block devices ------------- 1.09s",
                                "Get extra vars for Ceph configuration ----------------------------------- 0.96s",
                                "Add known partitions to the list of available block devices ------------- 0.90s",
                                "Add known links to the list of available block devices ------------------ 0.90s",
                                "Add known partitions to the list of available block devices ------------- 0.85s",
                                "Print configuration data ------------------------------------------------ 0.79s",
                                "Add known partitions to the list of available block devices ------------- 0.74s",
                                "Add known links to the list of available block devices ------------------ 0.69s",
                                "Generate lvm_volumes structure (block + wal) ---------------------------- 0.68s",
                                "Add known links to the list of available block devices ------------------ 0.66s",
                                "Get initial list of available block devices ----------------------------- 0.65s",
                                "Add known links to the list of available block devices ------------------ 0.63s",
                                "Generate shared DB/WAL VG names ----------------------------------------- 0.60s",
                                "Add known links to the list of available block devices ------------------ 0.58s",
                                "Print WAL devices ------------------------------------------------------- 0.57s",
                                "Add known partitions to the list of available block devices ------------- 0.56s",
                                "Set UUIDs for OSD VGs/LVs ----------------------------------------------- 0.55s",
                                "\u001b[32m2026-02-24 00:44:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 627b5eb2-2e98-4413-86a1-6581671e4d11 (sync inventory) is running in background. Output coming soon.\u001b[0m",
                                "\u001b[32m2026-02-24 00:44:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mStarting group_vars file reorganization\u001b[0m",
                                "\u001b[32m2026-02-24 00:44:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mMoved 0 file(s) to their respective directories\u001b[0m",
                                "\u001b[32m2026-02-24 00:44:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mGroup_vars file reorganization completed\u001b[0m",
                                "\u001b[32m2026-02-24 00:44:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mStarting variable preparation from inventory\u001b[0m",
                                "\u001b[32m2026-02-24 00:44:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWriting 050-kolla-ceph-rgw-hosts.yml with ceph_rgw_hosts\u001b[0m",
                                "\u001b[32m2026-02-24 00:44:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWriting 050-infrastructure-cephclient-mons.yml with cephclient_mons\u001b[0m",
                                "\u001b[32m2026-02-24 00:44:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWriting 050-ceph-cluster-fsid.yml with ceph_cluster_fsid\u001b[0m",
                                "\u001b[32m2026-02-24 00:44:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m3 file(s) written, 6 host(s) processed\u001b[0m",
                                "\u001b[32m2026-02-24 00:44:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mVariable preparation completed\u001b[0m",
                                "\u001b[32m2026-02-24 00:44:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mStarting inventory overwrite handling\u001b[0m",
                                "\u001b[32m2026-02-24 00:44:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mHandling group overwrites in 99-overwrite\u001b[0m",
                                "\u001b[32m2026-02-24 00:44:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mRemoving group frr:children from 60-generic\u001b[0m",
                                "\u001b[32m2026-02-24 00:44:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mRemoving group netbird:children from 50-infrastructure\u001b[0m",
                                "\u001b[32m2026-02-24 00:44:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mRemoving group ceph-rgw from 50-ceph\u001b[0m",
                                "\u001b[32m2026-02-24 00:44:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mRemoving group ceph-mds from 50-ceph\u001b[0m",
                                "\u001b[32m2026-02-24 00:44:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mHandling group overwrites in 20-roles\u001b[0m",
                                "\u001b[32m2026-02-24 00:44:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mRemoving group k3s_node from 50-infrastructure\u001b[0m",
                                "\u001b[32m2026-02-24 00:44:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mRemoved 5 group(s) in total\u001b[0m",
                                "\u001b[32m2026-02-24 00:44:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mInventory overwrite handling completed\u001b[0m",
                                "\u001b[32m2026-02-24 00:44:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mStarting merge of inventory files\u001b[0m",
                                "\u001b[32m2026-02-24 00:44:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mInventory files merged successfully\u001b[0m",
                                "\u001b[32m2026-02-24 00:44:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mGenerating ClusterShell configuration from Ansible inventory\u001b[0m",
                                "\u001b[32m2026-02-24 00:44:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSuccessfully wrote ClusterShell configuration\u001b[0m",
                                "[master 26aab81] 2026-02-24-00-44",
                                " 1 file changed, 30 insertions(+), 9 deletions(-)",
                                "\u001b[32m2026-02-24 00:44:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mPrepare task for execution of ceph-create-lvm-devices.\u001b[0m",
                                "\u001b[32m2026-02-24 00:44:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 1c145c4b-6a66-4ef0-a436-014fc77a1571 (ceph-create-lvm-devices) was prepared for execution.\u001b[0m",
                                "\u001b[32m2026-02-24 00:44:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task 1c145c4b-6a66-4ef0-a436-014fc77a1571 (ceph-create-lvm-devices) has been started and output is visible here.\u001b[0m",
                                "\u001b[1;35m[WARNING]: Collection community.general does not support Ansible version\u001b[0m",
                                "\u001b[1;35m2.16.14\u001b[0m",
                                "",
                                "PLAY [Ceph create LVM devices] *************************************************",
                                "",
                                "TASK [Get extra vars for Ceph configuration] ***********************************",
                                "Tuesday 24 February 2026  00:44:47 +0000 (0:00:00.223)       0:00:00.223 ****** ",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-manager(192.168.16.5)]\u001b[0m",
                                "",
                                "TASK [Get initial list of available block devices] *****************************",
                                "Tuesday 24 February 2026  00:44:47 +0000 (0:00:00.190)       0:00:00.413 ****** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Tuesday 24 February 2026  00:44:47 +0000 (0:00:00.179)       0:00:00.593 ****** ",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-3 => (item=loop0)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-3 => (item=loop1)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-3 => (item=loop2)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-3 => (item=loop3)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-3 => (item=loop4)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-3 => (item=loop5)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-3 => (item=loop6)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-3 => (item=loop7)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-3 => (item=sda)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-3 => (item=sdb)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-3 => (item=sdc)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-3 => (item=sdd)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-3 => (item=sr0)\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Tuesday 24 February 2026  00:44:48 +0000 (0:00:00.509)       0:00:01.102 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Tuesday 24 February 2026  00:44:48 +0000 (0:00:00.184)       0:00:01.287 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Tuesday 24 February 2026  00:44:48 +0000 (0:00:00.177)       0:00:01.464 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Tuesday 24 February 2026  00:44:48 +0000 (0:00:00.183)       0:00:01.648 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Tuesday 24 February 2026  00:44:48 +0000 (0:00:00.181)       0:00:01.830 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Tuesday 24 February 2026  00:44:49 +0000 (0:00:00.176)       0:00:02.006 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Tuesday 24 February 2026  00:44:49 +0000 (0:00:00.181)       0:00:02.188 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Tuesday 24 February 2026  00:44:49 +0000 (0:00:00.189)       0:00:02.377 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Tuesday 24 February 2026  00:44:49 +0000 (0:00:00.190)       0:00:02.568 ****** ",
                                "\u001b[0;32mok: [testbed-node-3] => (item=scsi-0QEMU_QEMU_HARDDISK_f90a4e80-1f36-4d14-b7da-de7b20e9fbca)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item=scsi-SQEMU_QEMU_HARDDISK_f90a4e80-1f36-4d14-b7da-de7b20e9fbca)\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Tuesday 24 February 2026  00:44:50 +0000 (0:00:00.375)       0:00:02.943 ****** ",
                                "\u001b[0;32mok: [testbed-node-3] => (item=scsi-0QEMU_QEMU_HARDDISK_9c6b6763-9704-41ec-9b6d-97eb80a5145c)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item=scsi-SQEMU_QEMU_HARDDISK_9c6b6763-9704-41ec-9b6d-97eb80a5145c)\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Tuesday 24 February 2026  00:44:50 +0000 (0:00:00.496)       0:00:03.440 ****** ",
                                "\u001b[0;32mok: [testbed-node-3] => (item=scsi-0QEMU_QEMU_HARDDISK_6644f191-69e9-4f3a-baff-cffe4152595b)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item=scsi-SQEMU_QEMU_HARDDISK_6644f191-69e9-4f3a-baff-cffe4152595b)\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Tuesday 24 February 2026  00:44:51 +0000 (0:00:00.522)       0:00:03.963 ****** ",
                                "\u001b[0;32mok: [testbed-node-3] => (item=scsi-0QEMU_QEMU_HARDDISK_4534650b-23a0-46c8-8bf1-44520caee637)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item=scsi-SQEMU_QEMU_HARDDISK_4534650b-23a0-46c8-8bf1-44520caee637)\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Tuesday 24 February 2026  00:44:51 +0000 (0:00:00.553)       0:00:04.516 ****** ",
                                "\u001b[0;32mok: [testbed-node-3] => (item=ata-QEMU_DVD-ROM_QM00001)\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Tuesday 24 February 2026  00:44:51 +0000 (0:00:00.296)       0:00:04.812 ****** ",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-3 => (item=loop0)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-3 => (item=loop1)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-3 => (item=loop2)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-3 => (item=loop3)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-3 => (item=loop4)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-3 => (item=loop5)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-3 => (item=loop6)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-3 => (item=loop7)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-3 => (item=sda)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-3 => (item=sdb)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-3 => (item=sdc)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-3 => (item=sdd)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-3 => (item=sr0)\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Tuesday 24 February 2026  00:44:52 +0000 (0:00:00.383)       0:00:05.196 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Tuesday 24 February 2026  00:44:52 +0000 (0:00:00.201)       0:00:05.398 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Tuesday 24 February 2026  00:44:52 +0000 (0:00:00.184)       0:00:05.582 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Tuesday 24 February 2026  00:44:52 +0000 (0:00:00.195)       0:00:05.778 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Tuesday 24 February 2026  00:44:53 +0000 (0:00:00.179)       0:00:05.958 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Tuesday 24 February 2026  00:44:53 +0000 (0:00:00.190)       0:00:06.149 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Tuesday 24 February 2026  00:44:53 +0000 (0:00:00.180)       0:00:06.329 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Tuesday 24 February 2026  00:44:53 +0000 (0:00:00.202)       0:00:06.532 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Tuesday 24 February 2026  00:44:53 +0000 (0:00:00.218)       0:00:06.750 ****** ",
                                "\u001b[0;32mok: [testbed-node-3] => (item=sda1)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item=sda14)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item=sda15)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item=sda16)\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Tuesday 24 February 2026  00:44:54 +0000 (0:00:01.065)       0:00:07.816 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Tuesday 24 February 2026  00:44:55 +0000 (0:00:00.306)       0:00:08.122 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Tuesday 24 February 2026  00:44:55 +0000 (0:00:00.256)       0:00:08.379 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Tuesday 24 February 2026  00:44:55 +0000 (0:00:00.233)       0:00:08.613 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Check whether ceph_db_wal_devices is used exclusively] *******************",
                                "Tuesday 24 February 2026  00:44:55 +0000 (0:00:00.190)       0:00:08.803 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Create dict of block VGs -> PVs from ceph_osd_devices] *******************",
                                "Tuesday 24 February 2026  00:44:56 +0000 (0:00:00.159)       0:00:08.963 ****** ",
                                "\u001b[0;32mok: [testbed-node-3] => (item={'key': 'sdb', 'value': {'osd_lvm_uuid': '9b92f24c-dac1-5bff-a390-34d9ff9e0f4f'}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item={'key': 'sdc', 'value': {'osd_lvm_uuid': '7f6a8fe5-6adc-52b4-b256-edcdbfd9bf4e'}})\u001b[0m",
                                "",
                                "TASK [Create block VGs] ********************************************************",
                                "Tuesday 24 February 2026  00:44:56 +0000 (0:00:00.197)       0:00:09.161 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'data': 'osd-block-9b92f24c-dac1-5bff-a390-34d9ff9e0f4f', 'data_vg': 'ceph-9b92f24c-dac1-5bff-a390-34d9ff9e0f4f'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'data': 'osd-block-7f6a8fe5-6adc-52b4-b256-edcdbfd9bf4e', 'data_vg': 'ceph-7f6a8fe5-6adc-52b4-b256-edcdbfd9bf4e'})\u001b[0m",
                                "",
                                "TASK [Print 'Create block VGs'] ************************************************",
                                "Tuesday 24 February 2026  00:44:58 +0000 (0:00:01.974)       0:00:11.135 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-9b92f24c-dac1-5bff-a390-34d9ff9e0f4f', 'data_vg': 'ceph-9b92f24c-dac1-5bff-a390-34d9ff9e0f4f'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-7f6a8fe5-6adc-52b4-b256-edcdbfd9bf4e', 'data_vg': 'ceph-7f6a8fe5-6adc-52b4-b256-edcdbfd9bf4e'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Create block LVs] ********************************************************",
                                "Tuesday 24 February 2026  00:44:58 +0000 (0:00:00.164)       0:00:11.299 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'data': 'osd-block-9b92f24c-dac1-5bff-a390-34d9ff9e0f4f', 'data_vg': 'ceph-9b92f24c-dac1-5bff-a390-34d9ff9e0f4f'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'data': 'osd-block-7f6a8fe5-6adc-52b4-b256-edcdbfd9bf4e', 'data_vg': 'ceph-7f6a8fe5-6adc-52b4-b256-edcdbfd9bf4e'})\u001b[0m",
                                "",
                                "TASK [Print 'Create block LVs'] ************************************************",
                                "Tuesday 24 February 2026  00:44:59 +0000 (0:00:01.528)       0:00:12.828 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-9b92f24c-dac1-5bff-a390-34d9ff9e0f4f', 'data_vg': 'ceph-9b92f24c-dac1-5bff-a390-34d9ff9e0f4f'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-7f6a8fe5-6adc-52b4-b256-edcdbfd9bf4e', 'data_vg': 'ceph-7f6a8fe5-6adc-52b4-b256-edcdbfd9bf4e'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Create DB VGs] ***********************************************************",
                                "Tuesday 24 February 2026  00:45:00 +0000 (0:00:00.150)       0:00:12.978 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Print 'Create DB VGs'] ***************************************************",
                                "Tuesday 24 February 2026  00:45:00 +0000 (0:00:00.128)       0:00:13.107 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-9b92f24c-dac1-5bff-a390-34d9ff9e0f4f', 'data_vg': 'ceph-9b92f24c-dac1-5bff-a390-34d9ff9e0f4f'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-7f6a8fe5-6adc-52b4-b256-edcdbfd9bf4e', 'data_vg': 'ceph-7f6a8fe5-6adc-52b4-b256-edcdbfd9bf4e'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Create WAL VGs] **********************************************************",
                                "Tuesday 24 February 2026  00:45:00 +0000 (0:00:00.357)       0:00:13.464 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Print 'Create WAL VGs'] **************************************************",
                                "Tuesday 24 February 2026  00:45:00 +0000 (0:00:00.135)       0:00:13.600 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-9b92f24c-dac1-5bff-a390-34d9ff9e0f4f', 'data_vg': 'ceph-9b92f24c-dac1-5bff-a390-34d9ff9e0f4f'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-7f6a8fe5-6adc-52b4-b256-edcdbfd9bf4e', 'data_vg': 'ceph-7f6a8fe5-6adc-52b4-b256-edcdbfd9bf4e'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Create DB+WAL VGs] *******************************************************",
                                "Tuesday 24 February 2026  00:45:00 +0000 (0:00:00.169)       0:00:13.769 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Print 'Create DB+WAL VGs'] ***********************************************",
                                "Tuesday 24 February 2026  00:45:01 +0000 (0:00:00.142)       0:00:13.912 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-9b92f24c-dac1-5bff-a390-34d9ff9e0f4f', 'data_vg': 'ceph-9b92f24c-dac1-5bff-a390-34d9ff9e0f4f'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-7f6a8fe5-6adc-52b4-b256-edcdbfd9bf4e', 'data_vg': 'ceph-7f6a8fe5-6adc-52b4-b256-edcdbfd9bf4e'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Prepare variables for OSD count check] ***********************************",
                                "Tuesday 24 February 2026  00:45:01 +0000 (0:00:00.187)       0:00:14.100 ****** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Count OSDs put on ceph_db_devices defined in lvm_volumes] ****************",
                                "Tuesday 24 February 2026  00:45:01 +0000 (0:00:00.165)       0:00:14.265 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-9b92f24c-dac1-5bff-a390-34d9ff9e0f4f', 'data_vg': 'ceph-9b92f24c-dac1-5bff-a390-34d9ff9e0f4f'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-7f6a8fe5-6adc-52b4-b256-edcdbfd9bf4e', 'data_vg': 'ceph-7f6a8fe5-6adc-52b4-b256-edcdbfd9bf4e'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Count OSDs put on ceph_wal_devices defined in lvm_volumes] ***************",
                                "Tuesday 24 February 2026  00:45:01 +0000 (0:00:00.206)       0:00:14.471 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-9b92f24c-dac1-5bff-a390-34d9ff9e0f4f', 'data_vg': 'ceph-9b92f24c-dac1-5bff-a390-34d9ff9e0f4f'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-7f6a8fe5-6adc-52b4-b256-edcdbfd9bf4e', 'data_vg': 'ceph-7f6a8fe5-6adc-52b4-b256-edcdbfd9bf4e'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Count OSDs put on ceph_db_wal_devices defined in lvm_volumes] ************",
                                "Tuesday 24 February 2026  00:45:01 +0000 (0:00:00.168)       0:00:14.639 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-9b92f24c-dac1-5bff-a390-34d9ff9e0f4f', 'data_vg': 'ceph-9b92f24c-dac1-5bff-a390-34d9ff9e0f4f'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-7f6a8fe5-6adc-52b4-b256-edcdbfd9bf4e', 'data_vg': 'ceph-7f6a8fe5-6adc-52b4-b256-edcdbfd9bf4e'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Fail if number of OSDs exceeds num_osds for a DB VG] *********************",
                                "Tuesday 24 February 2026  00:45:01 +0000 (0:00:00.179)       0:00:14.819 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Fail if number of OSDs exceeds num_osds for a WAL VG] ********************",
                                "Tuesday 24 February 2026  00:45:02 +0000 (0:00:00.142)       0:00:14.961 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Fail if number of OSDs exceeds num_osds for a DB+WAL VG] *****************",
                                "Tuesday 24 February 2026  00:45:02 +0000 (0:00:00.159)       0:00:15.121 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Print number of OSDs wanted per DB VG] ***********************************",
                                "Tuesday 24 February 2026  00:45:02 +0000 (0:00:00.131)       0:00:15.252 ****** ",
                                "\u001b[0;32mok: [testbed-node-3] => {\u001b[0m",
                                "\u001b[0;32m    \"_num_osds_wanted_per_db_vg\": {}\u001b[0m",
                                "\u001b[0;32m}\u001b[0m",
                                "",
                                "TASK [Print number of OSDs wanted per WAL VG] **********************************",
                                "Tuesday 24 February 2026  00:45:02 +0000 (0:00:00.435)       0:00:15.688 ****** ",
                                "\u001b[0;32mok: [testbed-node-3] => {\u001b[0m",
                                "\u001b[0;32m    \"_num_osds_wanted_per_wal_vg\": {}\u001b[0m",
                                "\u001b[0;32m}\u001b[0m",
                                "",
                                "TASK [Print number of OSDs wanted per DB+WAL VG] *******************************",
                                "Tuesday 24 February 2026  00:45:03 +0000 (0:00:00.164)       0:00:15.853 ****** ",
                                "\u001b[0;32mok: [testbed-node-3] => {\u001b[0m",
                                "\u001b[0;32m    \"_num_osds_wanted_per_db_wal_vg\": {}\u001b[0m",
                                "\u001b[0;32m}\u001b[0m",
                                "",
                                "TASK [Gather DB VGs with total and available size in bytes] ********************",
                                "Tuesday 24 February 2026  00:45:03 +0000 (0:00:00.157)       0:00:16.010 ****** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Gather WAL VGs with total and available size in bytes] *******************",
                                "Tuesday 24 February 2026  00:45:03 +0000 (0:00:00.682)       0:00:16.692 ****** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Gather DB+WAL VGs with total and available size in bytes] ****************",
                                "Tuesday 24 February 2026  00:45:04 +0000 (0:00:00.555)       0:00:17.248 ****** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Combine JSON from _db/wal/db_wal_vgs_cmd_output] *************************",
                                "Tuesday 24 February 2026  00:45:04 +0000 (0:00:00.552)       0:00:17.800 ****** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Calculate VG sizes (without buffer)] *************************************",
                                "Tuesday 24 February 2026  00:45:05 +0000 (0:00:00.167)       0:00:17.968 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Calculate VG sizes (with buffer)] ****************************************",
                                "Tuesday 24 February 2026  00:45:05 +0000 (0:00:00.118)       0:00:18.086 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Print LVM VGs report data] ***********************************************",
                                "Tuesday 24 February 2026  00:45:05 +0000 (0:00:00.105)       0:00:18.192 ****** ",
                                "\u001b[0;32mok: [testbed-node-3] => {\u001b[0m",
                                "\u001b[0;32m    \"vgs_report\": {\u001b[0m",
                                "\u001b[0;32m        \"vg\": []\u001b[0m",
                                "\u001b[0;32m    }\u001b[0m",
                                "\u001b[0;32m}\u001b[0m",
                                "",
                                "TASK [Print LVM VG sizes] ******************************************************",
                                "Tuesday 24 February 2026  00:45:05 +0000 (0:00:00.160)       0:00:18.353 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Calculate size needed for LVs on ceph_db_devices] ************************",
                                "Tuesday 24 February 2026  00:45:05 +0000 (0:00:00.138)       0:00:18.491 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Print size needed for LVs on ceph_db_devices] ****************************",
                                "Tuesday 24 February 2026  00:45:05 +0000 (0:00:00.157)       0:00:18.649 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Fail if size of DB LVs on ceph_db_devices > available] *******************",
                                "Tuesday 24 February 2026  00:45:06 +0000 (0:00:00.353)       0:00:19.002 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Calculate size needed for LVs on ceph_wal_devices] ***********************",
                                "Tuesday 24 February 2026  00:45:06 +0000 (0:00:00.140)       0:00:19.143 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Print size needed for LVs on ceph_wal_devices] ***************************",
                                "Tuesday 24 February 2026  00:45:06 +0000 (0:00:00.162)       0:00:19.306 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Fail if size of WAL LVs on ceph_wal_devices > available] *****************",
                                "Tuesday 24 February 2026  00:45:06 +0000 (0:00:00.187)       0:00:19.493 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Calculate size needed for WAL LVs on ceph_db_wal_devices] ****************",
                                "Tuesday 24 February 2026  00:45:06 +0000 (0:00:00.137)       0:00:19.631 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Print size needed for WAL LVs on ceph_db_wal_devices] ********************",
                                "Tuesday 24 February 2026  00:45:06 +0000 (0:00:00.138)       0:00:19.770 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Calculate size needed for DB LVs on ceph_db_wal_devices] *****************",
                                "Tuesday 24 February 2026  00:45:07 +0000 (0:00:00.138)       0:00:19.908 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Print size needed for DB LVs on ceph_db_wal_devices] *********************",
                                "Tuesday 24 February 2026  00:45:07 +0000 (0:00:00.146)       0:00:20.055 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Fail if size of DB+WAL LVs on ceph_db_wal_devices > available] ***********",
                                "Tuesday 24 February 2026  00:45:07 +0000 (0:00:00.135)       0:00:20.191 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Fail if DB LV size < 30 GiB for ceph_db_devices] *************************",
                                "Tuesday 24 February 2026  00:45:07 +0000 (0:00:00.126)       0:00:20.318 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Fail if DB LV size < 30 GiB for ceph_db_wal_devices] *********************",
                                "Tuesday 24 February 2026  00:45:07 +0000 (0:00:00.131)       0:00:20.449 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Create DB LVs for ceph_db_devices] ***************************************",
                                "Tuesday 24 February 2026  00:45:07 +0000 (0:00:00.136)       0:00:20.586 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-9b92f24c-dac1-5bff-a390-34d9ff9e0f4f', 'data_vg': 'ceph-9b92f24c-dac1-5bff-a390-34d9ff9e0f4f'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-7f6a8fe5-6adc-52b4-b256-edcdbfd9bf4e', 'data_vg': 'ceph-7f6a8fe5-6adc-52b4-b256-edcdbfd9bf4e'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Print 'Create DB LVs for ceph_db_devices'] *******************************",
                                "Tuesday 24 February 2026  00:45:08 +0000 (0:00:00.346)       0:00:20.932 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-9b92f24c-dac1-5bff-a390-34d9ff9e0f4f', 'data_vg': 'ceph-9b92f24c-dac1-5bff-a390-34d9ff9e0f4f'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-7f6a8fe5-6adc-52b4-b256-edcdbfd9bf4e', 'data_vg': 'ceph-7f6a8fe5-6adc-52b4-b256-edcdbfd9bf4e'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Create WAL LVs for ceph_wal_devices] *************************************",
                                "Tuesday 24 February 2026  00:45:08 +0000 (0:00:00.158)       0:00:21.091 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-9b92f24c-dac1-5bff-a390-34d9ff9e0f4f', 'data_vg': 'ceph-9b92f24c-dac1-5bff-a390-34d9ff9e0f4f'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-7f6a8fe5-6adc-52b4-b256-edcdbfd9bf4e', 'data_vg': 'ceph-7f6a8fe5-6adc-52b4-b256-edcdbfd9bf4e'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Print 'Create WAL LVs for ceph_wal_devices'] *****************************",
                                "Tuesday 24 February 2026  00:45:08 +0000 (0:00:00.152)       0:00:21.243 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-9b92f24c-dac1-5bff-a390-34d9ff9e0f4f', 'data_vg': 'ceph-9b92f24c-dac1-5bff-a390-34d9ff9e0f4f'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-7f6a8fe5-6adc-52b4-b256-edcdbfd9bf4e', 'data_vg': 'ceph-7f6a8fe5-6adc-52b4-b256-edcdbfd9bf4e'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Create WAL LVs for ceph_db_wal_devices] **********************************",
                                "Tuesday 24 February 2026  00:45:08 +0000 (0:00:00.161)       0:00:21.404 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-9b92f24c-dac1-5bff-a390-34d9ff9e0f4f', 'data_vg': 'ceph-9b92f24c-dac1-5bff-a390-34d9ff9e0f4f'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-7f6a8fe5-6adc-52b4-b256-edcdbfd9bf4e', 'data_vg': 'ceph-7f6a8fe5-6adc-52b4-b256-edcdbfd9bf4e'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Print 'Create WAL LVs for ceph_db_wal_devices'] **************************",
                                "Tuesday 24 February 2026  00:45:08 +0000 (0:00:00.165)       0:00:21.570 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-9b92f24c-dac1-5bff-a390-34d9ff9e0f4f', 'data_vg': 'ceph-9b92f24c-dac1-5bff-a390-34d9ff9e0f4f'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-7f6a8fe5-6adc-52b4-b256-edcdbfd9bf4e', 'data_vg': 'ceph-7f6a8fe5-6adc-52b4-b256-edcdbfd9bf4e'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Create DB LVs for ceph_db_wal_devices] ***********************************",
                                "Tuesday 24 February 2026  00:45:08 +0000 (0:00:00.146)       0:00:21.716 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-9b92f24c-dac1-5bff-a390-34d9ff9e0f4f', 'data_vg': 'ceph-9b92f24c-dac1-5bff-a390-34d9ff9e0f4f'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-7f6a8fe5-6adc-52b4-b256-edcdbfd9bf4e', 'data_vg': 'ceph-7f6a8fe5-6adc-52b4-b256-edcdbfd9bf4e'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Print 'Create DB LVs for ceph_db_wal_devices'] ***************************",
                                "Tuesday 24 February 2026  00:45:09 +0000 (0:00:00.138)       0:00:21.855 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-9b92f24c-dac1-5bff-a390-34d9ff9e0f4f', 'data_vg': 'ceph-9b92f24c-dac1-5bff-a390-34d9ff9e0f4f'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-7f6a8fe5-6adc-52b4-b256-edcdbfd9bf4e', 'data_vg': 'ceph-7f6a8fe5-6adc-52b4-b256-edcdbfd9bf4e'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Get list of Ceph LVs with associated VGs] ********************************",
                                "Tuesday 24 February 2026  00:45:09 +0000 (0:00:00.133)       0:00:21.989 ****** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Get list of Ceph PVs with associated VGs] ********************************",
                                "Tuesday 24 February 2026  00:45:09 +0000 (0:00:00.496)       0:00:22.485 ****** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Combine JSON from _lvs_cmd_output/_pvs_cmd_output] ***********************",
                                "Tuesday 24 February 2026  00:45:10 +0000 (0:00:00.502)       0:00:22.988 ****** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Create list of VG/LV names] **********************************************",
                                "Tuesday 24 February 2026  00:45:10 +0000 (0:00:00.137)       0:00:23.126 ****** ",
                                "\u001b[0;32mok: [testbed-node-3] => (item={'lv_name': 'osd-block-7f6a8fe5-6adc-52b4-b256-edcdbfd9bf4e', 'vg_name': 'ceph-7f6a8fe5-6adc-52b4-b256-edcdbfd9bf4e'})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item={'lv_name': 'osd-block-9b92f24c-dac1-5bff-a390-34d9ff9e0f4f', 'vg_name': 'ceph-9b92f24c-dac1-5bff-a390-34d9ff9e0f4f'})\u001b[0m",
                                "",
                                "TASK [Fail if block LV defined in lvm_volumes is missing] **********************",
                                "Tuesday 24 February 2026  00:45:10 +0000 (0:00:00.146)       0:00:23.273 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-9b92f24c-dac1-5bff-a390-34d9ff9e0f4f', 'data_vg': 'ceph-9b92f24c-dac1-5bff-a390-34d9ff9e0f4f'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-7f6a8fe5-6adc-52b4-b256-edcdbfd9bf4e', 'data_vg': 'ceph-7f6a8fe5-6adc-52b4-b256-edcdbfd9bf4e'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Fail if DB LV defined in lvm_volumes is missing] *************************",
                                "Tuesday 24 February 2026  00:45:10 +0000 (0:00:00.256)       0:00:23.529 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-9b92f24c-dac1-5bff-a390-34d9ff9e0f4f', 'data_vg': 'ceph-9b92f24c-dac1-5bff-a390-34d9ff9e0f4f'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-7f6a8fe5-6adc-52b4-b256-edcdbfd9bf4e', 'data_vg': 'ceph-7f6a8fe5-6adc-52b4-b256-edcdbfd9bf4e'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Fail if WAL LV defined in lvm_volumes is missing] ************************",
                                "Tuesday 24 February 2026  00:45:10 +0000 (0:00:00.128)       0:00:23.658 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-9b92f24c-dac1-5bff-a390-34d9ff9e0f4f', 'data_vg': 'ceph-9b92f24c-dac1-5bff-a390-34d9ff9e0f4f'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'data': 'osd-block-7f6a8fe5-6adc-52b4-b256-edcdbfd9bf4e', 'data_vg': 'ceph-7f6a8fe5-6adc-52b4-b256-edcdbfd9bf4e'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Print LVM report data] ***************************************************",
                                "Tuesday 24 February 2026  00:45:10 +0000 (0:00:00.134)       0:00:23.793 ****** ",
                                "\u001b[0;32mok: [testbed-node-3] => {\u001b[0m",
                                "\u001b[0;32m    \"lvm_report\": {\u001b[0m",
                                "\u001b[0;32m        \"lv\": [\u001b[0m",
                                "\u001b[0;32m            {\u001b[0m",
                                "\u001b[0;32m                \"lv_name\": \"osd-block-7f6a8fe5-6adc-52b4-b256-edcdbfd9bf4e\",\u001b[0m",
                                "\u001b[0;32m                \"vg_name\": \"ceph-7f6a8fe5-6adc-52b4-b256-edcdbfd9bf4e\"\u001b[0m",
                                "\u001b[0;32m            },\u001b[0m",
                                "\u001b[0;32m            {\u001b[0m",
                                "\u001b[0;32m                \"lv_name\": \"osd-block-9b92f24c-dac1-5bff-a390-34d9ff9e0f4f\",\u001b[0m",
                                "\u001b[0;32m                \"vg_name\": \"ceph-9b92f24c-dac1-5bff-a390-34d9ff9e0f4f\"\u001b[0m",
                                "\u001b[0;32m            }\u001b[0m",
                                "\u001b[0;32m        ],\u001b[0m",
                                "\u001b[0;32m        \"pv\": [\u001b[0m",
                                "\u001b[0;32m            {\u001b[0m",
                                "\u001b[0;32m                \"pv_name\": \"/dev/sdb\",\u001b[0m",
                                "\u001b[0;32m                \"vg_name\": \"ceph-9b92f24c-dac1-5bff-a390-34d9ff9e0f4f\"\u001b[0m",
                                "\u001b[0;32m            },\u001b[0m",
                                "\u001b[0;32m            {\u001b[0m",
                                "\u001b[0;32m                \"pv_name\": \"/dev/sdc\",\u001b[0m",
                                "\u001b[0;32m                \"vg_name\": \"ceph-7f6a8fe5-6adc-52b4-b256-edcdbfd9bf4e\"\u001b[0m",
                                "\u001b[0;32m            }\u001b[0m",
                                "\u001b[0;32m        ]\u001b[0m",
                                "\u001b[0;32m    }\u001b[0m",
                                "\u001b[0;32m}\u001b[0m",
                                "",
                                "PLAY [Ceph create LVM devices] *************************************************",
                                "",
                                "TASK [Get extra vars for Ceph configuration] ***********************************",
                                "Tuesday 24 February 2026  00:45:11 +0000 (0:00:00.253)       0:00:24.047 ****** ",
                                "\u001b[0;32mok: [testbed-node-4 -> testbed-manager(192.168.16.5)]\u001b[0m",
                                "",
                                "TASK [Get initial list of available block devices] *****************************",
                                "Tuesday 24 February 2026  00:45:11 +0000 (0:00:00.281)       0:00:24.328 ****** ",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Tuesday 24 February 2026  00:45:11 +0000 (0:00:00.229)       0:00:24.558 ****** ",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-4 => (item=loop0)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-4 => (item=loop1)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-4 => (item=loop2)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-4 => (item=loop3)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-4 => (item=loop4)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-4 => (item=loop5)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-4 => (item=loop6)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-4 => (item=loop7)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-4 => (item=sda)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-4 => (item=sdb)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-4 => (item=sdc)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-4 => (item=sdd)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-4 => (item=sr0)\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Tuesday 24 February 2026  00:45:12 +0000 (0:00:00.378)       0:00:24.936 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Tuesday 24 February 2026  00:45:12 +0000 (0:00:00.203)       0:00:25.140 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Tuesday 24 February 2026  00:45:12 +0000 (0:00:00.186)       0:00:25.326 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Tuesday 24 February 2026  00:45:12 +0000 (0:00:00.481)       0:00:25.808 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Tuesday 24 February 2026  00:45:13 +0000 (0:00:00.188)       0:00:25.996 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Tuesday 24 February 2026  00:45:13 +0000 (0:00:00.204)       0:00:26.201 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Tuesday 24 February 2026  00:45:13 +0000 (0:00:00.175)       0:00:26.377 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Tuesday 24 February 2026  00:45:13 +0000 (0:00:00.182)       0:00:26.559 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Tuesday 24 February 2026  00:45:13 +0000 (0:00:00.174)       0:00:26.734 ****** ",
                                "\u001b[0;32mok: [testbed-node-4] => (item=scsi-0QEMU_QEMU_HARDDISK_47e600ec-05ca-44a6-93d0-5ca3104e3639)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item=scsi-SQEMU_QEMU_HARDDISK_47e600ec-05ca-44a6-93d0-5ca3104e3639)\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Tuesday 24 February 2026  00:45:14 +0000 (0:00:00.393)       0:00:27.127 ****** ",
                                "\u001b[0;32mok: [testbed-node-4] => (item=scsi-0QEMU_QEMU_HARDDISK_f659b61f-c39f-426f-a640-6a753951661d)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item=scsi-SQEMU_QEMU_HARDDISK_f659b61f-c39f-426f-a640-6a753951661d)\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Tuesday 24 February 2026  00:45:14 +0000 (0:00:00.378)       0:00:27.506 ****** ",
                                "\u001b[0;32mok: [testbed-node-4] => (item=scsi-0QEMU_QEMU_HARDDISK_fb58bf16-a1c6-4abb-925b-29c6a6a6f100)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item=scsi-SQEMU_QEMU_HARDDISK_fb58bf16-a1c6-4abb-925b-29c6a6a6f100)\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Tuesday 24 February 2026  00:45:15 +0000 (0:00:00.401)       0:00:27.907 ****** ",
                                "\u001b[0;32mok: [testbed-node-4] => (item=scsi-0QEMU_QEMU_HARDDISK_5660bb62-e2c9-45b2-ae11-8ab0f829b194)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item=scsi-SQEMU_QEMU_HARDDISK_5660bb62-e2c9-45b2-ae11-8ab0f829b194)\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Tuesday 24 February 2026  00:45:15 +0000 (0:00:00.517)       0:00:28.424 ****** ",
                                "\u001b[0;32mok: [testbed-node-4] => (item=ata-QEMU_DVD-ROM_QM00001)\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Tuesday 24 February 2026  00:45:16 +0000 (0:00:00.447)       0:00:28.872 ****** ",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-4 => (item=loop0)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-4 => (item=loop1)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-4 => (item=loop2)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-4 => (item=loop3)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-4 => (item=loop4)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-4 => (item=loop5)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-4 => (item=loop6)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-4 => (item=loop7)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-4 => (item=sda)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-4 => (item=sdb)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-4 => (item=sdc)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-4 => (item=sdd)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-4 => (item=sr0)\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Tuesday 24 February 2026  00:45:16 +0000 (0:00:00.679)       0:00:29.551 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Tuesday 24 February 2026  00:45:16 +0000 (0:00:00.175)       0:00:29.727 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Tuesday 24 February 2026  00:45:17 +0000 (0:00:00.188)       0:00:29.915 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Tuesday 24 February 2026  00:45:17 +0000 (0:00:00.183)       0:00:30.099 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Tuesday 24 February 2026  00:45:17 +0000 (0:00:00.187)       0:00:30.286 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Tuesday 24 February 2026  00:45:17 +0000 (0:00:00.185)       0:00:30.472 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Tuesday 24 February 2026  00:45:17 +0000 (0:00:00.178)       0:00:30.650 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Tuesday 24 February 2026  00:45:18 +0000 (0:00:00.192)       0:00:30.843 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Tuesday 24 February 2026  00:45:18 +0000 (0:00:00.270)       0:00:31.114 ****** ",
                                "\u001b[0;32mok: [testbed-node-4] => (item=sda1)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item=sda14)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item=sda15)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item=sda16)\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Tuesday 24 February 2026  00:45:19 +0000 (0:00:00.737)       0:00:31.851 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Tuesday 24 February 2026  00:45:19 +0000 (0:00:00.181)       0:00:32.033 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Tuesday 24 February 2026  00:45:19 +0000 (0:00:00.475)       0:00:32.509 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Tuesday 24 February 2026  00:45:19 +0000 (0:00:00.196)       0:00:32.705 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Check whether ceph_db_wal_devices is used exclusively] *******************",
                                "Tuesday 24 February 2026  00:45:20 +0000 (0:00:00.174)       0:00:32.880 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Create dict of block VGs -> PVs from ceph_osd_devices] *******************",
                                "Tuesday 24 February 2026  00:45:20 +0000 (0:00:00.112)       0:00:32.992 ****** ",
                                "\u001b[0;32mok: [testbed-node-4] => (item={'key': 'sdb', 'value': {'osd_lvm_uuid': '294dd48f-9d6d-5079-8214-59e60af9e526'}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item={'key': 'sdc', 'value': {'osd_lvm_uuid': '7fbf4d63-7c6f-5c70-887c-2b73c978775b'}})\u001b[0m",
                                "",
                                "TASK [Create block VGs] ********************************************************",
                                "Tuesday 24 February 2026  00:45:20 +0000 (0:00:00.176)       0:00:33.169 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'data': 'osd-block-294dd48f-9d6d-5079-8214-59e60af9e526', 'data_vg': 'ceph-294dd48f-9d6d-5079-8214-59e60af9e526'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'data': 'osd-block-7fbf4d63-7c6f-5c70-887c-2b73c978775b', 'data_vg': 'ceph-7fbf4d63-7c6f-5c70-887c-2b73c978775b'})\u001b[0m",
                                "",
                                "TASK [Print 'Create block VGs'] ************************************************",
                                "Tuesday 24 February 2026  00:45:22 +0000 (0:00:01.839)       0:00:35.009 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-294dd48f-9d6d-5079-8214-59e60af9e526', 'data_vg': 'ceph-294dd48f-9d6d-5079-8214-59e60af9e526'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-7fbf4d63-7c6f-5c70-887c-2b73c978775b', 'data_vg': 'ceph-7fbf4d63-7c6f-5c70-887c-2b73c978775b'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Create block LVs] ********************************************************",
                                "Tuesday 24 February 2026  00:45:22 +0000 (0:00:00.144)       0:00:35.154 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'data': 'osd-block-294dd48f-9d6d-5079-8214-59e60af9e526', 'data_vg': 'ceph-294dd48f-9d6d-5079-8214-59e60af9e526'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'data': 'osd-block-7fbf4d63-7c6f-5c70-887c-2b73c978775b', 'data_vg': 'ceph-7fbf4d63-7c6f-5c70-887c-2b73c978775b'})\u001b[0m",
                                "",
                                "TASK [Print 'Create block LVs'] ************************************************",
                                "Tuesday 24 February 2026  00:45:23 +0000 (0:00:01.290)       0:00:36.444 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-294dd48f-9d6d-5079-8214-59e60af9e526', 'data_vg': 'ceph-294dd48f-9d6d-5079-8214-59e60af9e526'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-7fbf4d63-7c6f-5c70-887c-2b73c978775b', 'data_vg': 'ceph-7fbf4d63-7c6f-5c70-887c-2b73c978775b'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Create DB VGs] ***********************************************************",
                                "Tuesday 24 February 2026  00:45:23 +0000 (0:00:00.137)       0:00:36.582 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Print 'Create DB VGs'] ***************************************************",
                                "Tuesday 24 February 2026  00:45:23 +0000 (0:00:00.125)       0:00:36.707 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-294dd48f-9d6d-5079-8214-59e60af9e526', 'data_vg': 'ceph-294dd48f-9d6d-5079-8214-59e60af9e526'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-7fbf4d63-7c6f-5c70-887c-2b73c978775b', 'data_vg': 'ceph-7fbf4d63-7c6f-5c70-887c-2b73c978775b'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Create WAL VGs] **********************************************************",
                                "Tuesday 24 February 2026  00:45:24 +0000 (0:00:00.133)       0:00:36.841 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Print 'Create WAL VGs'] **************************************************",
                                "Tuesday 24 February 2026  00:45:24 +0000 (0:00:00.100)       0:00:36.941 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-294dd48f-9d6d-5079-8214-59e60af9e526', 'data_vg': 'ceph-294dd48f-9d6d-5079-8214-59e60af9e526'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-7fbf4d63-7c6f-5c70-887c-2b73c978775b', 'data_vg': 'ceph-7fbf4d63-7c6f-5c70-887c-2b73c978775b'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Create DB+WAL VGs] *******************************************************",
                                "Tuesday 24 February 2026  00:45:24 +0000 (0:00:00.252)       0:00:37.194 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Print 'Create DB+WAL VGs'] ***********************************************",
                                "Tuesday 24 February 2026  00:45:24 +0000 (0:00:00.121)       0:00:37.316 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-294dd48f-9d6d-5079-8214-59e60af9e526', 'data_vg': 'ceph-294dd48f-9d6d-5079-8214-59e60af9e526'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-7fbf4d63-7c6f-5c70-887c-2b73c978775b', 'data_vg': 'ceph-7fbf4d63-7c6f-5c70-887c-2b73c978775b'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Prepare variables for OSD count check] ***********************************",
                                "Tuesday 24 February 2026  00:45:24 +0000 (0:00:00.142)       0:00:37.459 ****** ",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Count OSDs put on ceph_db_devices defined in lvm_volumes] ****************",
                                "Tuesday 24 February 2026  00:45:24 +0000 (0:00:00.124)       0:00:37.584 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-294dd48f-9d6d-5079-8214-59e60af9e526', 'data_vg': 'ceph-294dd48f-9d6d-5079-8214-59e60af9e526'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-7fbf4d63-7c6f-5c70-887c-2b73c978775b', 'data_vg': 'ceph-7fbf4d63-7c6f-5c70-887c-2b73c978775b'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Count OSDs put on ceph_wal_devices defined in lvm_volumes] ***************",
                                "Tuesday 24 February 2026  00:45:24 +0000 (0:00:00.135)       0:00:37.719 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-294dd48f-9d6d-5079-8214-59e60af9e526', 'data_vg': 'ceph-294dd48f-9d6d-5079-8214-59e60af9e526'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-7fbf4d63-7c6f-5c70-887c-2b73c978775b', 'data_vg': 'ceph-7fbf4d63-7c6f-5c70-887c-2b73c978775b'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Count OSDs put on ceph_db_wal_devices defined in lvm_volumes] ************",
                                "Tuesday 24 February 2026  00:45:25 +0000 (0:00:00.138)       0:00:37.858 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-294dd48f-9d6d-5079-8214-59e60af9e526', 'data_vg': 'ceph-294dd48f-9d6d-5079-8214-59e60af9e526'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-7fbf4d63-7c6f-5c70-887c-2b73c978775b', 'data_vg': 'ceph-7fbf4d63-7c6f-5c70-887c-2b73c978775b'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Fail if number of OSDs exceeds num_osds for a DB VG] *********************",
                                "Tuesday 24 February 2026  00:45:25 +0000 (0:00:00.142)       0:00:38.000 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Fail if number of OSDs exceeds num_osds for a WAL VG] ********************",
                                "Tuesday 24 February 2026  00:45:25 +0000 (0:00:00.132)       0:00:38.133 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Fail if number of OSDs exceeds num_osds for a DB+WAL VG] *****************",
                                "Tuesday 24 February 2026  00:45:25 +0000 (0:00:00.125)       0:00:38.258 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Print number of OSDs wanted per DB VG] ***********************************",
                                "Tuesday 24 February 2026  00:45:25 +0000 (0:00:00.121)       0:00:38.380 ****** ",
                                "\u001b[0;32mok: [testbed-node-4] => {\u001b[0m",
                                "\u001b[0;32m    \"_num_osds_wanted_per_db_vg\": {}\u001b[0m",
                                "\u001b[0;32m}\u001b[0m",
                                "",
                                "TASK [Print number of OSDs wanted per WAL VG] **********************************",
                                "Tuesday 24 February 2026  00:45:25 +0000 (0:00:00.137)       0:00:38.518 ****** ",
                                "\u001b[0;32mok: [testbed-node-4] => {\u001b[0m",
                                "\u001b[0;32m    \"_num_osds_wanted_per_wal_vg\": {}\u001b[0m",
                                "\u001b[0;32m}\u001b[0m",
                                "",
                                "TASK [Print number of OSDs wanted per DB+WAL VG] *******************************",
                                "Tuesday 24 February 2026  00:45:25 +0000 (0:00:00.108)       0:00:38.626 ****** ",
                                "\u001b[0;32mok: [testbed-node-4] => {\u001b[0m",
                                "\u001b[0;32m    \"_num_osds_wanted_per_db_wal_vg\": {}\u001b[0m",
                                "\u001b[0;32m}\u001b[0m",
                                "",
                                "TASK [Gather DB VGs with total and available size in bytes] ********************",
                                "Tuesday 24 February 2026  00:45:26 +0000 (0:00:00.230)       0:00:38.856 ****** ",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Gather WAL VGs with total and available size in bytes] *******************",
                                "Tuesday 24 February 2026  00:45:26 +0000 (0:00:00.502)       0:00:39.359 ****** ",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Gather DB+WAL VGs with total and available size in bytes] ****************",
                                "Tuesday 24 February 2026  00:45:26 +0000 (0:00:00.468)       0:00:39.827 ****** ",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Combine JSON from _db/wal/db_wal_vgs_cmd_output] *************************",
                                "Tuesday 24 February 2026  00:45:27 +0000 (0:00:00.480)       0:00:40.308 ****** ",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Calculate VG sizes (without buffer)] *************************************",
                                "Tuesday 24 February 2026  00:45:27 +0000 (0:00:00.140)       0:00:40.448 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Calculate VG sizes (with buffer)] ****************************************",
                                "Tuesday 24 February 2026  00:45:27 +0000 (0:00:00.104)       0:00:40.552 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Print LVM VGs report data] ***********************************************",
                                "Tuesday 24 February 2026  00:45:27 +0000 (0:00:00.089)       0:00:40.642 ****** ",
                                "\u001b[0;32mok: [testbed-node-4] => {\u001b[0m",
                                "\u001b[0;32m    \"vgs_report\": {\u001b[0m",
                                "\u001b[0;32m        \"vg\": []\u001b[0m",
                                "\u001b[0;32m    }\u001b[0m",
                                "\u001b[0;32m}\u001b[0m",
                                "",
                                "TASK [Print LVM VG sizes] ******************************************************",
                                "Tuesday 24 February 2026  00:45:27 +0000 (0:00:00.113)       0:00:40.755 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Calculate size needed for LVs on ceph_db_devices] ************************",
                                "Tuesday 24 February 2026  00:45:28 +0000 (0:00:00.120)       0:00:40.876 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Print size needed for LVs on ceph_db_devices] ****************************",
                                "Tuesday 24 February 2026  00:45:28 +0000 (0:00:00.125)       0:00:41.001 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Fail if size of DB LVs on ceph_db_devices > available] *******************",
                                "Tuesday 24 February 2026  00:45:28 +0000 (0:00:00.115)       0:00:41.117 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Calculate size needed for LVs on ceph_wal_devices] ***********************",
                                "Tuesday 24 February 2026  00:45:28 +0000 (0:00:00.114)       0:00:41.231 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Print size needed for LVs on ceph_wal_devices] ***************************",
                                "Tuesday 24 February 2026  00:45:28 +0000 (0:00:00.235)       0:00:41.467 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Fail if size of WAL LVs on ceph_wal_devices > available] *****************",
                                "Tuesday 24 February 2026  00:45:28 +0000 (0:00:00.114)       0:00:41.581 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Calculate size needed for WAL LVs on ceph_db_wal_devices] ****************",
                                "Tuesday 24 February 2026  00:45:28 +0000 (0:00:00.116)       0:00:41.697 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Print size needed for WAL LVs on ceph_db_wal_devices] ********************",
                                "Tuesday 24 February 2026  00:45:28 +0000 (0:00:00.126)       0:00:41.824 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Calculate size needed for DB LVs on ceph_db_wal_devices] *****************",
                                "Tuesday 24 February 2026  00:45:29 +0000 (0:00:00.117)       0:00:41.942 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Print size needed for DB LVs on ceph_db_wal_devices] *********************",
                                "Tuesday 24 February 2026  00:45:29 +0000 (0:00:00.117)       0:00:42.060 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Fail if size of DB+WAL LVs on ceph_db_wal_devices > available] ***********",
                                "Tuesday 24 February 2026  00:45:29 +0000 (0:00:00.126)       0:00:42.186 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Fail if DB LV size < 30 GiB for ceph_db_devices] *************************",
                                "Tuesday 24 February 2026  00:45:29 +0000 (0:00:00.126)       0:00:42.313 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Fail if DB LV size < 30 GiB for ceph_db_wal_devices] *********************",
                                "Tuesday 24 February 2026  00:45:29 +0000 (0:00:00.111)       0:00:42.424 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Create DB LVs for ceph_db_devices] ***************************************",
                                "Tuesday 24 February 2026  00:45:29 +0000 (0:00:00.130)       0:00:42.554 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-294dd48f-9d6d-5079-8214-59e60af9e526', 'data_vg': 'ceph-294dd48f-9d6d-5079-8214-59e60af9e526'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-7fbf4d63-7c6f-5c70-887c-2b73c978775b', 'data_vg': 'ceph-7fbf4d63-7c6f-5c70-887c-2b73c978775b'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Print 'Create DB LVs for ceph_db_devices'] *******************************",
                                "Tuesday 24 February 2026  00:45:29 +0000 (0:00:00.140)       0:00:42.694 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-294dd48f-9d6d-5079-8214-59e60af9e526', 'data_vg': 'ceph-294dd48f-9d6d-5079-8214-59e60af9e526'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-7fbf4d63-7c6f-5c70-887c-2b73c978775b', 'data_vg': 'ceph-7fbf4d63-7c6f-5c70-887c-2b73c978775b'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Create WAL LVs for ceph_wal_devices] *************************************",
                                "Tuesday 24 February 2026  00:45:29 +0000 (0:00:00.135)       0:00:42.829 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-294dd48f-9d6d-5079-8214-59e60af9e526', 'data_vg': 'ceph-294dd48f-9d6d-5079-8214-59e60af9e526'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-7fbf4d63-7c6f-5c70-887c-2b73c978775b', 'data_vg': 'ceph-7fbf4d63-7c6f-5c70-887c-2b73c978775b'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Print 'Create WAL LVs for ceph_wal_devices'] *****************************",
                                "Tuesday 24 February 2026  00:45:30 +0000 (0:00:00.257)       0:00:43.087 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-294dd48f-9d6d-5079-8214-59e60af9e526', 'data_vg': 'ceph-294dd48f-9d6d-5079-8214-59e60af9e526'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-7fbf4d63-7c6f-5c70-887c-2b73c978775b', 'data_vg': 'ceph-7fbf4d63-7c6f-5c70-887c-2b73c978775b'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Create WAL LVs for ceph_db_wal_devices] **********************************",
                                "Tuesday 24 February 2026  00:45:30 +0000 (0:00:00.116)       0:00:43.204 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-294dd48f-9d6d-5079-8214-59e60af9e526', 'data_vg': 'ceph-294dd48f-9d6d-5079-8214-59e60af9e526'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-7fbf4d63-7c6f-5c70-887c-2b73c978775b', 'data_vg': 'ceph-7fbf4d63-7c6f-5c70-887c-2b73c978775b'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Print 'Create WAL LVs for ceph_db_wal_devices'] **************************",
                                "Tuesday 24 February 2026  00:45:30 +0000 (0:00:00.106)       0:00:43.310 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-294dd48f-9d6d-5079-8214-59e60af9e526', 'data_vg': 'ceph-294dd48f-9d6d-5079-8214-59e60af9e526'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-7fbf4d63-7c6f-5c70-887c-2b73c978775b', 'data_vg': 'ceph-7fbf4d63-7c6f-5c70-887c-2b73c978775b'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Create DB LVs for ceph_db_wal_devices] ***********************************",
                                "Tuesday 24 February 2026  00:45:30 +0000 (0:00:00.136)       0:00:43.447 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-294dd48f-9d6d-5079-8214-59e60af9e526', 'data_vg': 'ceph-294dd48f-9d6d-5079-8214-59e60af9e526'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-7fbf4d63-7c6f-5c70-887c-2b73c978775b', 'data_vg': 'ceph-7fbf4d63-7c6f-5c70-887c-2b73c978775b'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Print 'Create DB LVs for ceph_db_wal_devices'] ***************************",
                                "Tuesday 24 February 2026  00:45:30 +0000 (0:00:00.150)       0:00:43.597 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-294dd48f-9d6d-5079-8214-59e60af9e526', 'data_vg': 'ceph-294dd48f-9d6d-5079-8214-59e60af9e526'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-7fbf4d63-7c6f-5c70-887c-2b73c978775b', 'data_vg': 'ceph-7fbf4d63-7c6f-5c70-887c-2b73c978775b'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Get list of Ceph LVs with associated VGs] ********************************",
                                "Tuesday 24 February 2026  00:45:30 +0000 (0:00:00.133)       0:00:43.731 ****** ",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Get list of Ceph PVs with associated VGs] ********************************",
                                "Tuesday 24 February 2026  00:45:31 +0000 (0:00:00.485)       0:00:44.216 ****** ",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Combine JSON from _lvs_cmd_output/_pvs_cmd_output] ***********************",
                                "Tuesday 24 February 2026  00:45:31 +0000 (0:00:00.475)       0:00:44.692 ****** ",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Create list of VG/LV names] **********************************************",
                                "Tuesday 24 February 2026  00:45:32 +0000 (0:00:00.142)       0:00:44.834 ****** ",
                                "\u001b[0;32mok: [testbed-node-4] => (item={'lv_name': 'osd-block-294dd48f-9d6d-5079-8214-59e60af9e526', 'vg_name': 'ceph-294dd48f-9d6d-5079-8214-59e60af9e526'})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item={'lv_name': 'osd-block-7fbf4d63-7c6f-5c70-887c-2b73c978775b', 'vg_name': 'ceph-7fbf4d63-7c6f-5c70-887c-2b73c978775b'})\u001b[0m",
                                "",
                                "TASK [Fail if block LV defined in lvm_volumes is missing] **********************",
                                "Tuesday 24 February 2026  00:45:32 +0000 (0:00:00.162)       0:00:44.996 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-294dd48f-9d6d-5079-8214-59e60af9e526', 'data_vg': 'ceph-294dd48f-9d6d-5079-8214-59e60af9e526'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-7fbf4d63-7c6f-5c70-887c-2b73c978775b', 'data_vg': 'ceph-7fbf4d63-7c6f-5c70-887c-2b73c978775b'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Fail if DB LV defined in lvm_volumes is missing] *************************",
                                "Tuesday 24 February 2026  00:45:32 +0000 (0:00:00.138)       0:00:45.135 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-294dd48f-9d6d-5079-8214-59e60af9e526', 'data_vg': 'ceph-294dd48f-9d6d-5079-8214-59e60af9e526'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-7fbf4d63-7c6f-5c70-887c-2b73c978775b', 'data_vg': 'ceph-7fbf4d63-7c6f-5c70-887c-2b73c978775b'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Fail if WAL LV defined in lvm_volumes is missing] ************************",
                                "Tuesday 24 February 2026  00:45:32 +0000 (0:00:00.147)       0:00:45.282 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-294dd48f-9d6d-5079-8214-59e60af9e526', 'data_vg': 'ceph-294dd48f-9d6d-5079-8214-59e60af9e526'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'data': 'osd-block-7fbf4d63-7c6f-5c70-887c-2b73c978775b', 'data_vg': 'ceph-7fbf4d63-7c6f-5c70-887c-2b73c978775b'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [Print LVM report data] ***************************************************",
                                "Tuesday 24 February 2026  00:45:32 +0000 (0:00:00.169)       0:00:45.452 ****** ",
                                "\u001b[0;32mok: [testbed-node-4] => {\u001b[0m",
                                "\u001b[0;32m    \"lvm_report\": {\u001b[0m",
                                "\u001b[0;32m        \"lv\": [\u001b[0m",
                                "\u001b[0;32m            {\u001b[0m",
                                "\u001b[0;32m                \"lv_name\": \"osd-block-294dd48f-9d6d-5079-8214-59e60af9e526\",\u001b[0m",
                                "\u001b[0;32m                \"vg_name\": \"ceph-294dd48f-9d6d-5079-8214-59e60af9e526\"\u001b[0m",
                                "\u001b[0;32m            },\u001b[0m",
                                "\u001b[0;32m            {\u001b[0m",
                                "\u001b[0;32m                \"lv_name\": \"osd-block-7fbf4d63-7c6f-5c70-887c-2b73c978775b\",\u001b[0m",
                                "\u001b[0;32m                \"vg_name\": \"ceph-7fbf4d63-7c6f-5c70-887c-2b73c978775b\"\u001b[0m",
                                "\u001b[0;32m            }\u001b[0m",
                                "\u001b[0;32m        ],\u001b[0m",
                                "\u001b[0;32m        \"pv\": [\u001b[0m",
                                "\u001b[0;32m            {\u001b[0m",
                                "\u001b[0;32m                \"pv_name\": \"/dev/sdb\",\u001b[0m",
                                "\u001b[0;32m                \"vg_name\": \"ceph-294dd48f-9d6d-5079-8214-59e60af9e526\"\u001b[0m",
                                "\u001b[0;32m            },\u001b[0m",
                                "\u001b[0;32m            {\u001b[0m",
                                "\u001b[0;32m                \"pv_name\": \"/dev/sdc\",\u001b[0m",
                                "\u001b[0;32m                \"vg_name\": \"ceph-7fbf4d63-7c6f-5c70-887c-2b73c978775b\"\u001b[0m",
                                "\u001b[0;32m            }\u001b[0m",
                                "\u001b[0;32m        ]\u001b[0m",
                                "\u001b[0;32m    }\u001b[0m",
                                "\u001b[0;32m}\u001b[0m",
                                "",
                                "PLAY [Ceph create LVM devices] *************************************************",
                                "",
                                "TASK [Get extra vars for Ceph configuration] ***********************************",
                                "Tuesday 24 February 2026  00:45:33 +0000 (0:00:00.394)       0:00:45.847 ****** ",
                                "\u001b[0;32mok: [testbed-node-5 -> testbed-manager(192.168.16.5)]\u001b[0m",
                                "",
                                "TASK [Get initial list of available block devices] *****************************",
                                "Tuesday 24 February 2026  00:45:33 +0000 (0:00:00.287)       0:00:46.134 ****** ",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Tuesday 24 February 2026  00:45:33 +0000 (0:00:00.236)       0:00:46.371 ****** ",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-5 => (item=loop0)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-5 => (item=loop1)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-5 => (item=loop2)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-5 => (item=loop3)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-5 => (item=loop4)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-5 => (item=loop5)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-5 => (item=loop6)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-5 => (item=loop7)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-5 => (item=sda)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-5 => (item=sdb)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-5 => (item=sdc)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-5 => (item=sdd)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-links.yml for testbed-node-5 => (item=sr0)\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Tuesday 24 February 2026  00:45:33 +0000 (0:00:00.401)       0:00:46.773 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Tuesday 24 February 2026  00:45:34 +0000 (0:00:00.196)       0:00:46.969 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Tuesday 24 February 2026  00:45:34 +0000 (0:00:00.203)       0:00:47.173 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Tuesday 24 February 2026  00:45:34 +0000 (0:00:00.224)       0:00:47.398 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Tuesday 24 February 2026  00:45:34 +0000 (0:00:00.191)       0:00:47.590 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Tuesday 24 February 2026  00:45:35 +0000 (0:00:00.607)       0:00:48.197 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Tuesday 24 February 2026  00:45:35 +0000 (0:00:00.194)       0:00:48.391 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Tuesday 24 February 2026  00:45:35 +0000 (0:00:00.208)       0:00:48.600 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Tuesday 24 February 2026  00:45:35 +0000 (0:00:00.198)       0:00:48.798 ****** ",
                                "\u001b[0;32mok: [testbed-node-5] => (item=scsi-0QEMU_QEMU_HARDDISK_7a9487cb-ec9e-4508-9c5a-f692700fc5a4)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item=scsi-SQEMU_QEMU_HARDDISK_7a9487cb-ec9e-4508-9c5a-f692700fc5a4)\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Tuesday 24 February 2026  00:45:36 +0000 (0:00:00.411)       0:00:49.210 ****** ",
                                "\u001b[0;32mok: [testbed-node-5] => (item=scsi-0QEMU_QEMU_HARDDISK_1558935b-4c66-4614-97e5-cda14abd2641)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item=scsi-SQEMU_QEMU_HARDDISK_1558935b-4c66-4614-97e5-cda14abd2641)\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Tuesday 24 February 2026  00:45:36 +0000 (0:00:00.424)       0:00:49.634 ****** ",
                                "\u001b[0;32mok: [testbed-node-5] => (item=scsi-0QEMU_QEMU_HARDDISK_6f4643f0-329b-4ae2-97c7-d348648fd970)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item=scsi-SQEMU_QEMU_HARDDISK_6f4643f0-329b-4ae2-97c7-d348648fd970)\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Tuesday 24 February 2026  00:45:37 +0000 (0:00:00.418)       0:00:50.053 ****** ",
                                "\u001b[0;32mok: [testbed-node-5] => (item=scsi-0QEMU_QEMU_HARDDISK_5f47789f-2a1a-442a-8a3c-644a4f92e5a8)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item=scsi-SQEMU_QEMU_HARDDISK_5f47789f-2a1a-442a-8a3c-644a4f92e5a8)\u001b[0m",
                                "",
                                "TASK [Add known links to the list of available block devices] ******************",
                                "Tuesday 24 February 2026  00:45:37 +0000 (0:00:00.414)       0:00:50.468 ****** ",
                                "\u001b[0;32mok: [testbed-node-5] => (item=ata-QEMU_DVD-ROM_QM00001)\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Tuesday 24 February 2026  00:45:37 +0000 (0:00:00.318)       0:00:50.786 ****** ",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-5 => (item=loop0)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-5 => (item=loop1)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-5 => (item=loop2)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-5 => (item=loop3)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-5 => (item=loop4)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-5 => (item=loop5)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-5 => (item=loop6)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-5 => (item=loop7)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-5 => (item=sda)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-5 => (item=sdb)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-5 => (item=sdc)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-5 => (item=sdd)\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/_add-device-partitions.yml for testbed-node-5 => (item=sr0)\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Tuesday 24 February 2026  00:45:38 +0000 (0:00:00.421)       0:00:51.207 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Tuesday 24 February 2026  00:45:38 +0000 (0:00:00.206)       0:00:51.413 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Tuesday 24 February 2026  00:45:39 +0000 (0:00:00.595)       0:00:52.009 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Tuesday 24 February 2026  00:45:39 +0000 (0:00:00.195)       0:00:52.204 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Tuesday 24 February 2026  00:45:39 +0000 (0:00:00.190)       0:00:52.394 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Tuesday 24 February 2026  00:45:39 +0000 (0:00:00.205)       0:00:52.600 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Tuesday 24 February 2026  00:45:39 +0000 (0:00:00.193)       0:00:52.794 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Tuesday 24 February 2026  00:45:40 +0000 (0:00:00.210)       0:00:53.004 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Tuesday 24 February 2026  00:45:40 +0000 (0:00:00.192)       0:00:53.196 ****** ",
                                "\u001b[0;32mok: [testbed-node-5] => (item=sda1)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item=sda14)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item=sda15)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item=sda16)\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Tuesday 24 February 2026  00:45:40 +0000 (0:00:00.631)       0:00:53.828 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Tuesday 24 February 2026  00:45:41 +0000 (0:00:00.196)       0:00:54.024 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Tuesday 24 February 2026  00:45:41 +0000 (0:00:00.200)       0:00:54.225 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Add known partitions to the list of available block devices] *************",
                                "Tuesday 24 February 2026  00:45:41 +0000 (0:00:00.180)       0:00:54.406 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Check whether ceph_db_wal_devices is used exclusively] *******************",
                                "Tuesday 24 February 2026  00:45:41 +0000 (0:00:00.194)       0:00:54.600 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Create dict of block VGs -> PVs from ceph_osd_devices] *******************",
                                "Tuesday 24 February 2026  00:45:42 +0000 (0:00:00.297)       0:00:54.898 ****** ",
                                "\u001b[0;32mok: [testbed-node-5] => (item={'key': 'sdb', 'value': {'osd_lvm_uuid': '752a9ec4-4bfc-5551-82cd-92b802bdad97'}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item={'key': 'sdc', 'value': {'osd_lvm_uuid': 'ba85a0e3-2f57-51d5-a9f1-072f0e9a8deb'}})\u001b[0m",
                                "",
                                "TASK [Create block VGs] ********************************************************",
                                "Tuesday 24 February 2026  00:45:42 +0000 (0:00:00.200)       0:00:55.099 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'data': 'osd-block-752a9ec4-4bfc-5551-82cd-92b802bdad97', 'data_vg': 'ceph-752a9ec4-4bfc-5551-82cd-92b802bdad97'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'data': 'osd-block-ba85a0e3-2f57-51d5-a9f1-072f0e9a8deb', 'data_vg': 'ceph-ba85a0e3-2f57-51d5-a9f1-072f0e9a8deb'})\u001b[0m",
                                "",
                                "TASK [Print 'Create block VGs'] ************************************************",
                                "Tuesday 24 February 2026  00:45:44 +0000 (0:00:01.937)       0:00:57.036 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-752a9ec4-4bfc-5551-82cd-92b802bdad97', 'data_vg': 'ceph-752a9ec4-4bfc-5551-82cd-92b802bdad97'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-ba85a0e3-2f57-51d5-a9f1-072f0e9a8deb', 'data_vg': 'ceph-ba85a0e3-2f57-51d5-a9f1-072f0e9a8deb'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Create block LVs] ********************************************************",
                                "Tuesday 24 February 2026  00:45:44 +0000 (0:00:00.153)       0:00:57.190 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'data': 'osd-block-752a9ec4-4bfc-5551-82cd-92b802bdad97', 'data_vg': 'ceph-752a9ec4-4bfc-5551-82cd-92b802bdad97'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'data': 'osd-block-ba85a0e3-2f57-51d5-a9f1-072f0e9a8deb', 'data_vg': 'ceph-ba85a0e3-2f57-51d5-a9f1-072f0e9a8deb'})\u001b[0m",
                                "",
                                "TASK [Print 'Create block LVs'] ************************************************",
                                "Tuesday 24 February 2026  00:45:45 +0000 (0:00:01.349)       0:00:58.539 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-752a9ec4-4bfc-5551-82cd-92b802bdad97', 'data_vg': 'ceph-752a9ec4-4bfc-5551-82cd-92b802bdad97'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-ba85a0e3-2f57-51d5-a9f1-072f0e9a8deb', 'data_vg': 'ceph-ba85a0e3-2f57-51d5-a9f1-072f0e9a8deb'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Create DB VGs] ***********************************************************",
                                "Tuesday 24 February 2026  00:45:45 +0000 (0:00:00.146)       0:00:58.686 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Print 'Create DB VGs'] ***************************************************",
                                "Tuesday 24 February 2026  00:45:45 +0000 (0:00:00.132)       0:00:58.819 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-752a9ec4-4bfc-5551-82cd-92b802bdad97', 'data_vg': 'ceph-752a9ec4-4bfc-5551-82cd-92b802bdad97'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-ba85a0e3-2f57-51d5-a9f1-072f0e9a8deb', 'data_vg': 'ceph-ba85a0e3-2f57-51d5-a9f1-072f0e9a8deb'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Create WAL VGs] **********************************************************",
                                "Tuesday 24 February 2026  00:45:46 +0000 (0:00:00.149)       0:00:58.968 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Print 'Create WAL VGs'] **************************************************",
                                "Tuesday 24 February 2026  00:45:46 +0000 (0:00:00.145)       0:00:59.113 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-752a9ec4-4bfc-5551-82cd-92b802bdad97', 'data_vg': 'ceph-752a9ec4-4bfc-5551-82cd-92b802bdad97'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-ba85a0e3-2f57-51d5-a9f1-072f0e9a8deb', 'data_vg': 'ceph-ba85a0e3-2f57-51d5-a9f1-072f0e9a8deb'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Create DB+WAL VGs] *******************************************************",
                                "Tuesday 24 February 2026  00:45:46 +0000 (0:00:00.148)       0:00:59.262 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Print 'Create DB+WAL VGs'] ***********************************************",
                                "Tuesday 24 February 2026  00:45:46 +0000 (0:00:00.136)       0:00:59.398 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-752a9ec4-4bfc-5551-82cd-92b802bdad97', 'data_vg': 'ceph-752a9ec4-4bfc-5551-82cd-92b802bdad97'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-ba85a0e3-2f57-51d5-a9f1-072f0e9a8deb', 'data_vg': 'ceph-ba85a0e3-2f57-51d5-a9f1-072f0e9a8deb'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Prepare variables for OSD count check] ***********************************",
                                "Tuesday 24 February 2026  00:45:46 +0000 (0:00:00.147)       0:00:59.546 ****** ",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Count OSDs put on ceph_db_devices defined in lvm_volumes] ****************",
                                "Tuesday 24 February 2026  00:45:47 +0000 (0:00:00.305)       0:00:59.851 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-752a9ec4-4bfc-5551-82cd-92b802bdad97', 'data_vg': 'ceph-752a9ec4-4bfc-5551-82cd-92b802bdad97'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-ba85a0e3-2f57-51d5-a9f1-072f0e9a8deb', 'data_vg': 'ceph-ba85a0e3-2f57-51d5-a9f1-072f0e9a8deb'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Count OSDs put on ceph_wal_devices defined in lvm_volumes] ***************",
                                "Tuesday 24 February 2026  00:45:47 +0000 (0:00:00.153)       0:01:00.004 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-752a9ec4-4bfc-5551-82cd-92b802bdad97', 'data_vg': 'ceph-752a9ec4-4bfc-5551-82cd-92b802bdad97'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-ba85a0e3-2f57-51d5-a9f1-072f0e9a8deb', 'data_vg': 'ceph-ba85a0e3-2f57-51d5-a9f1-072f0e9a8deb'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Count OSDs put on ceph_db_wal_devices defined in lvm_volumes] ************",
                                "Tuesday 24 February 2026  00:45:47 +0000 (0:00:00.151)       0:01:00.156 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-752a9ec4-4bfc-5551-82cd-92b802bdad97', 'data_vg': 'ceph-752a9ec4-4bfc-5551-82cd-92b802bdad97'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-ba85a0e3-2f57-51d5-a9f1-072f0e9a8deb', 'data_vg': 'ceph-ba85a0e3-2f57-51d5-a9f1-072f0e9a8deb'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Fail if number of OSDs exceeds num_osds for a DB VG] *********************",
                                "Tuesday 24 February 2026  00:45:47 +0000 (0:00:00.168)       0:01:00.324 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Fail if number of OSDs exceeds num_osds for a WAL VG] ********************",
                                "Tuesday 24 February 2026  00:45:47 +0000 (0:00:00.137)       0:01:00.461 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Fail if number of OSDs exceeds num_osds for a DB+WAL VG] *****************",
                                "Tuesday 24 February 2026  00:45:47 +0000 (0:00:00.123)       0:01:00.585 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Print number of OSDs wanted per DB VG] ***********************************",
                                "Tuesday 24 February 2026  00:45:47 +0000 (0:00:00.134)       0:01:00.719 ****** ",
                                "\u001b[0;32mok: [testbed-node-5] => {\u001b[0m",
                                "\u001b[0;32m    \"_num_osds_wanted_per_db_vg\": {}\u001b[0m",
                                "\u001b[0;32m}\u001b[0m",
                                "",
                                "TASK [Print number of OSDs wanted per WAL VG] **********************************",
                                "Tuesday 24 February 2026  00:45:48 +0000 (0:00:00.138)       0:01:00.858 ****** ",
                                "\u001b[0;32mok: [testbed-node-5] => {\u001b[0m",
                                "\u001b[0;32m    \"_num_osds_wanted_per_wal_vg\": {}\u001b[0m",
                                "\u001b[0;32m}\u001b[0m",
                                "",
                                "TASK [Print number of OSDs wanted per DB+WAL VG] *******************************",
                                "Tuesday 24 February 2026  00:45:48 +0000 (0:00:00.141)       0:01:00.999 ****** ",
                                "\u001b[0;32mok: [testbed-node-5] => {\u001b[0m",
                                "\u001b[0;32m    \"_num_osds_wanted_per_db_wal_vg\": {}\u001b[0m",
                                "\u001b[0;32m}\u001b[0m",
                                "",
                                "TASK [Gather DB VGs with total and available size in bytes] ********************",
                                "Tuesday 24 February 2026  00:45:48 +0000 (0:00:00.132)       0:01:01.132 ****** ",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Gather WAL VGs with total and available size in bytes] *******************",
                                "Tuesday 24 February 2026  00:45:48 +0000 (0:00:00.526)       0:01:01.658 ****** ",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Gather DB+WAL VGs with total and available size in bytes] ****************",
                                "Tuesday 24 February 2026  00:45:49 +0000 (0:00:00.540)       0:01:02.198 ****** ",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Combine JSON from _db/wal/db_wal_vgs_cmd_output] *************************",
                                "Tuesday 24 February 2026  00:45:50 +0000 (0:00:00.706)       0:01:02.904 ****** ",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Calculate VG sizes (without buffer)] *************************************",
                                "Tuesday 24 February 2026  00:45:50 +0000 (0:00:00.136)       0:01:03.041 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Calculate VG sizes (with buffer)] ****************************************",
                                "Tuesday 24 February 2026  00:45:50 +0000 (0:00:00.091)       0:01:03.133 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Print LVM VGs report data] ***********************************************",
                                "Tuesday 24 February 2026  00:45:50 +0000 (0:00:00.097)       0:01:03.231 ****** ",
                                "\u001b[0;32mok: [testbed-node-5] => {\u001b[0m",
                                "\u001b[0;32m    \"vgs_report\": {\u001b[0m",
                                "\u001b[0;32m        \"vg\": []\u001b[0m",
                                "\u001b[0;32m    }\u001b[0m",
                                "\u001b[0;32m}\u001b[0m",
                                "",
                                "TASK [Print LVM VG sizes] ******************************************************",
                                "Tuesday 24 February 2026  00:45:50 +0000 (0:00:00.127)       0:01:03.359 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Calculate size needed for LVs on ceph_db_devices] ************************",
                                "Tuesday 24 February 2026  00:45:50 +0000 (0:00:00.121)       0:01:03.480 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Print size needed for LVs on ceph_db_devices] ****************************",
                                "Tuesday 24 February 2026  00:45:50 +0000 (0:00:00.115)       0:01:03.596 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Fail if size of DB LVs on ceph_db_devices > available] *******************",
                                "Tuesday 24 February 2026  00:45:50 +0000 (0:00:00.117)       0:01:03.714 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Calculate size needed for LVs on ceph_wal_devices] ***********************",
                                "Tuesday 24 February 2026  00:45:51 +0000 (0:00:00.129)       0:01:03.843 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Print size needed for LVs on ceph_wal_devices] ***************************",
                                "Tuesday 24 February 2026  00:45:51 +0000 (0:00:00.116)       0:01:03.959 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Fail if size of WAL LVs on ceph_wal_devices > available] *****************",
                                "Tuesday 24 February 2026  00:45:51 +0000 (0:00:00.112)       0:01:04.072 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Calculate size needed for WAL LVs on ceph_db_wal_devices] ****************",
                                "Tuesday 24 February 2026  00:45:51 +0000 (0:00:00.099)       0:01:04.172 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Print size needed for WAL LVs on ceph_db_wal_devices] ********************",
                                "Tuesday 24 February 2026  00:45:51 +0000 (0:00:00.228)       0:01:04.401 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Calculate size needed for DB LVs on ceph_db_wal_devices] *****************",
                                "Tuesday 24 February 2026  00:45:51 +0000 (0:00:00.123)       0:01:04.524 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Print size needed for DB LVs on ceph_db_wal_devices] *********************",
                                "Tuesday 24 February 2026  00:45:51 +0000 (0:00:00.119)       0:01:04.644 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Fail if size of DB+WAL LVs on ceph_db_wal_devices > available] ***********",
                                "Tuesday 24 February 2026  00:45:51 +0000 (0:00:00.123)       0:01:04.767 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Fail if DB LV size < 30 GiB for ceph_db_devices] *************************",
                                "Tuesday 24 February 2026  00:45:52 +0000 (0:00:00.107)       0:01:04.875 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Fail if DB LV size < 30 GiB for ceph_db_wal_devices] *********************",
                                "Tuesday 24 February 2026  00:45:52 +0000 (0:00:00.109)       0:01:04.985 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Create DB LVs for ceph_db_devices] ***************************************",
                                "Tuesday 24 February 2026  00:45:52 +0000 (0:00:00.118)       0:01:05.104 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-752a9ec4-4bfc-5551-82cd-92b802bdad97', 'data_vg': 'ceph-752a9ec4-4bfc-5551-82cd-92b802bdad97'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-ba85a0e3-2f57-51d5-a9f1-072f0e9a8deb', 'data_vg': 'ceph-ba85a0e3-2f57-51d5-a9f1-072f0e9a8deb'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Print 'Create DB LVs for ceph_db_devices'] *******************************",
                                "Tuesday 24 February 2026  00:45:52 +0000 (0:00:00.128)       0:01:05.232 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-752a9ec4-4bfc-5551-82cd-92b802bdad97', 'data_vg': 'ceph-752a9ec4-4bfc-5551-82cd-92b802bdad97'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-ba85a0e3-2f57-51d5-a9f1-072f0e9a8deb', 'data_vg': 'ceph-ba85a0e3-2f57-51d5-a9f1-072f0e9a8deb'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Create WAL LVs for ceph_wal_devices] *************************************",
                                "Tuesday 24 February 2026  00:45:52 +0000 (0:00:00.123)       0:01:05.355 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-752a9ec4-4bfc-5551-82cd-92b802bdad97', 'data_vg': 'ceph-752a9ec4-4bfc-5551-82cd-92b802bdad97'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-ba85a0e3-2f57-51d5-a9f1-072f0e9a8deb', 'data_vg': 'ceph-ba85a0e3-2f57-51d5-a9f1-072f0e9a8deb'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Print 'Create WAL LVs for ceph_wal_devices'] *****************************",
                                "Tuesday 24 February 2026  00:45:52 +0000 (0:00:00.142)       0:01:05.497 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-752a9ec4-4bfc-5551-82cd-92b802bdad97', 'data_vg': 'ceph-752a9ec4-4bfc-5551-82cd-92b802bdad97'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-ba85a0e3-2f57-51d5-a9f1-072f0e9a8deb', 'data_vg': 'ceph-ba85a0e3-2f57-51d5-a9f1-072f0e9a8deb'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Create WAL LVs for ceph_db_wal_devices] **********************************",
                                "Tuesday 24 February 2026  00:45:52 +0000 (0:00:00.135)       0:01:05.632 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-752a9ec4-4bfc-5551-82cd-92b802bdad97', 'data_vg': 'ceph-752a9ec4-4bfc-5551-82cd-92b802bdad97'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-ba85a0e3-2f57-51d5-a9f1-072f0e9a8deb', 'data_vg': 'ceph-ba85a0e3-2f57-51d5-a9f1-072f0e9a8deb'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Print 'Create WAL LVs for ceph_db_wal_devices'] **************************",
                                "Tuesday 24 February 2026  00:45:52 +0000 (0:00:00.138)       0:01:05.771 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-752a9ec4-4bfc-5551-82cd-92b802bdad97', 'data_vg': 'ceph-752a9ec4-4bfc-5551-82cd-92b802bdad97'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-ba85a0e3-2f57-51d5-a9f1-072f0e9a8deb', 'data_vg': 'ceph-ba85a0e3-2f57-51d5-a9f1-072f0e9a8deb'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Create DB LVs for ceph_db_wal_devices] ***********************************",
                                "Tuesday 24 February 2026  00:45:53 +0000 (0:00:00.261)       0:01:06.033 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-752a9ec4-4bfc-5551-82cd-92b802bdad97', 'data_vg': 'ceph-752a9ec4-4bfc-5551-82cd-92b802bdad97'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-ba85a0e3-2f57-51d5-a9f1-072f0e9a8deb', 'data_vg': 'ceph-ba85a0e3-2f57-51d5-a9f1-072f0e9a8deb'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Print 'Create DB LVs for ceph_db_wal_devices'] ***************************",
                                "Tuesday 24 February 2026  00:45:53 +0000 (0:00:00.143)       0:01:06.176 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-752a9ec4-4bfc-5551-82cd-92b802bdad97', 'data_vg': 'ceph-752a9ec4-4bfc-5551-82cd-92b802bdad97'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-ba85a0e3-2f57-51d5-a9f1-072f0e9a8deb', 'data_vg': 'ceph-ba85a0e3-2f57-51d5-a9f1-072f0e9a8deb'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Get list of Ceph LVs with associated VGs] ********************************",
                                "Tuesday 24 February 2026  00:45:53 +0000 (0:00:00.153)       0:01:06.330 ****** ",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Get list of Ceph PVs with associated VGs] ********************************",
                                "Tuesday 24 February 2026  00:45:54 +0000 (0:00:00.531)       0:01:06.862 ****** ",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Combine JSON from _lvs_cmd_output/_pvs_cmd_output] ***********************",
                                "Tuesday 24 February 2026  00:45:54 +0000 (0:00:00.484)       0:01:07.347 ****** ",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Create list of VG/LV names] **********************************************",
                                "Tuesday 24 February 2026  00:45:54 +0000 (0:00:00.120)       0:01:07.468 ****** ",
                                "\u001b[0;32mok: [testbed-node-5] => (item={'lv_name': 'osd-block-752a9ec4-4bfc-5551-82cd-92b802bdad97', 'vg_name': 'ceph-752a9ec4-4bfc-5551-82cd-92b802bdad97'})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item={'lv_name': 'osd-block-ba85a0e3-2f57-51d5-a9f1-072f0e9a8deb', 'vg_name': 'ceph-ba85a0e3-2f57-51d5-a9f1-072f0e9a8deb'})\u001b[0m",
                                "",
                                "TASK [Fail if block LV defined in lvm_volumes is missing] **********************",
                                "Tuesday 24 February 2026  00:45:54 +0000 (0:00:00.145)       0:01:07.613 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-752a9ec4-4bfc-5551-82cd-92b802bdad97', 'data_vg': 'ceph-752a9ec4-4bfc-5551-82cd-92b802bdad97'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-ba85a0e3-2f57-51d5-a9f1-072f0e9a8deb', 'data_vg': 'ceph-ba85a0e3-2f57-51d5-a9f1-072f0e9a8deb'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Fail if DB LV defined in lvm_volumes is missing] *************************",
                                "Tuesday 24 February 2026  00:45:54 +0000 (0:00:00.138)       0:01:07.752 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-752a9ec4-4bfc-5551-82cd-92b802bdad97', 'data_vg': 'ceph-752a9ec4-4bfc-5551-82cd-92b802bdad97'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-ba85a0e3-2f57-51d5-a9f1-072f0e9a8deb', 'data_vg': 'ceph-ba85a0e3-2f57-51d5-a9f1-072f0e9a8deb'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Fail if WAL LV defined in lvm_volumes is missing] ************************",
                                "Tuesday 24 February 2026  00:45:55 +0000 (0:00:00.137)       0:01:07.890 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-752a9ec4-4bfc-5551-82cd-92b802bdad97', 'data_vg': 'ceph-752a9ec4-4bfc-5551-82cd-92b802bdad97'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'data': 'osd-block-ba85a0e3-2f57-51d5-a9f1-072f0e9a8deb', 'data_vg': 'ceph-ba85a0e3-2f57-51d5-a9f1-072f0e9a8deb'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Print LVM report data] ***************************************************",
                                "Tuesday 24 February 2026  00:45:55 +0000 (0:00:00.153)       0:01:08.044 ****** ",
                                "\u001b[0;32mok: [testbed-node-5] => {\u001b[0m",
                                "\u001b[0;32m    \"lvm_report\": {\u001b[0m",
                                "\u001b[0;32m        \"lv\": [\u001b[0m",
                                "\u001b[0;32m            {\u001b[0m",
                                "\u001b[0;32m                \"lv_name\": \"osd-block-752a9ec4-4bfc-5551-82cd-92b802bdad97\",\u001b[0m",
                                "\u001b[0;32m                \"vg_name\": \"ceph-752a9ec4-4bfc-5551-82cd-92b802bdad97\"\u001b[0m",
                                "\u001b[0;32m            },\u001b[0m",
                                "\u001b[0;32m            {\u001b[0m",
                                "\u001b[0;32m                \"lv_name\": \"osd-block-ba85a0e3-2f57-51d5-a9f1-072f0e9a8deb\",\u001b[0m",
                                "\u001b[0;32m                \"vg_name\": \"ceph-ba85a0e3-2f57-51d5-a9f1-072f0e9a8deb\"\u001b[0m",
                                "\u001b[0;32m            }\u001b[0m",
                                "\u001b[0;32m        ],\u001b[0m",
                                "\u001b[0;32m        \"pv\": [\u001b[0m",
                                "\u001b[0;32m            {\u001b[0m",
                                "\u001b[0;32m                \"pv_name\": \"/dev/sdb\",\u001b[0m",
                                "\u001b[0;32m                \"vg_name\": \"ceph-752a9ec4-4bfc-5551-82cd-92b802bdad97\"\u001b[0m",
                                "\u001b[0;32m            },\u001b[0m",
                                "\u001b[0;32m            {\u001b[0m",
                                "\u001b[0;32m                \"pv_name\": \"/dev/sdc\",\u001b[0m",
                                "\u001b[0;32m                \"vg_name\": \"ceph-ba85a0e3-2f57-51d5-a9f1-072f0e9a8deb\"\u001b[0m",
                                "\u001b[0;32m            }\u001b[0m",
                                "\u001b[0;32m        ]\u001b[0m",
                                "\u001b[0;32m    }\u001b[0m",
                                "\u001b[0;32m}\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;33mtestbed-node-3\u001b[0m             : \u001b[0;32mok=51  \u001b[0m \u001b[0;33mchanged=2   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=62  \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-4\u001b[0m             : \u001b[0;32mok=51  \u001b[0m \u001b[0;33mchanged=2   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=62  \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-5\u001b[0m             : \u001b[0;32mok=51  \u001b[0m \u001b[0;33mchanged=2   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=62  \u001b[0m rescued=0    ignored=0   ",
                                "",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Tuesday 24 February 2026  00:45:55 +0000 (0:00:00.135)       0:01:08.179 ****** ",
                                "=============================================================================== ",
                                "Create block VGs -------------------------------------------------------- 5.75s",
                                "Create block LVs -------------------------------------------------------- 4.17s",
                                "Gather DB+WAL VGs with total and available size in bytes ---------------- 1.74s",
                                "Gather DB VGs with total and available size in bytes -------------------- 1.71s",
                                "Gather WAL VGs with total and available size in bytes ------------------- 1.56s",
                                "Get list of Ceph LVs with associated VGs -------------------------------- 1.51s",
                                "Add known partitions to the list of available block devices ------------- 1.48s",
                                "Get list of Ceph PVs with associated VGs -------------------------------- 1.46s",
                                "Add known links to the list of available block devices ------------------ 1.29s",
                                "Add known partitions to the list of available block devices ------------- 1.07s",
                                "Print LVM report data --------------------------------------------------- 0.78s",
                                "Get extra vars for Ceph configuration ----------------------------------- 0.76s",
                                "Add known partitions to the list of available block devices ------------- 0.74s",
                                "Print number of OSDs wanted per DB VG ----------------------------------- 0.71s",
                                "Get initial list of available block devices ----------------------------- 0.65s",
                                "Print 'Create DB VGs' --------------------------------------------------- 0.64s",
                                "Add known partitions to the list of available block devices ------------- 0.63s",
                                "Create DB LVs for ceph_db_devices --------------------------------------- 0.62s",
                                "Add known links to the list of available block devices ------------------ 0.61s",
                                "Add known partitions to the list of available block devices ------------- 0.60s",
                                "\u001b[32m2026-02-24 00:46:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mPrepare task for execution of facts.\u001b[0m",
                                "\u001b[32m2026-02-24 00:46:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask eb8126f1-e933-4fbf-9464-6ddb35899e54 (facts) was prepared for execution.\u001b[0m",
                                "\u001b[32m2026-02-24 00:46:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task eb8126f1-e933-4fbf-9464-6ddb35899e54 (facts) has been started and output is visible here.\u001b[0m",
                                "",
                                "PLAY [Apply role facts] ********************************************************",
                                "",
                                "TASK [osism.commons.facts : Create custom facts directory] *********************",
                                "Tuesday 24 February 2026  00:46:12 +0000 (0:00:00.240)       0:00:00.240 ****** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.commons.facts : Copy fact files] ***********************************",
                                "Tuesday 24 February 2026  00:46:13 +0000 (0:00:01.011)       0:00:01.252 ****** ",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "PLAY [Gather facts for all hosts] **********************************************",
                                "",
                                "TASK [Gathers facts about hosts] ***********************************************",
                                "Tuesday 24 February 2026  00:46:14 +0000 (0:00:01.065)       0:00:02.318 ****** ",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "PLAY [Gather facts for all hosts if using --limit] *****************************",
                                "",
                                "TASK [Gather facts for all hosts] **********************************************",
                                "Tuesday 24 February 2026  00:46:19 +0000 (0:00:04.998)       0:00:07.316 ****** ",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;32mtestbed-manager\u001b[0m            : \u001b[0;32mok=2   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;32mtestbed-node-0\u001b[0m             : \u001b[0;32mok=2   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;32mtestbed-node-1\u001b[0m             : \u001b[0;32mok=2   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;32mtestbed-node-2\u001b[0m             : \u001b[0;32mok=2   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;32mtestbed-node-3\u001b[0m             : \u001b[0;32mok=2   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;32mtestbed-node-4\u001b[0m             : \u001b[0;32mok=2   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;32mtestbed-node-5\u001b[0m             : \u001b[0;32mok=2   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Tuesday 24 February 2026  00:46:19 +0000 (0:00:00.481)       0:00:07.797 ****** ",
                                "=============================================================================== ",
                                "Gathers facts about hosts ----------------------------------------------- 5.00s",
                                "osism.commons.facts : Copy fact files ----------------------------------- 1.07s",
                                "osism.commons.facts : Create custom facts directory --------------------- 1.01s",
                                "Gather facts for all hosts ---------------------------------------------- 0.48s",
                                "\u001b[32m2026-02-24 00:46:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mPrepare task for execution of frr.\u001b[0m",
                                "\u001b[32m2026-02-24 00:46:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 7e87fc91-2dc6-4330-b4c9-6b6256ab6f40 (frr) was prepared for execution.\u001b[0m",
                                "\u001b[32m2026-02-24 00:46:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task 7e87fc91-2dc6-4330-b4c9-6b6256ab6f40 (frr) has been started and output is visible here.\u001b[0m",
                                "",
                                "PLAY [Apply role frr] **********************************************************",
                                "",
                                "TASK [osism.services.frr : Include distribution specific install tasks] ********",
                                "Tuesday 24 February 2026  00:46:36 +0000 (0:00:00.232)       0:00:00.232 ****** ",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/services/roles/frr/tasks/install-Debian-family.yml for testbed-manager\u001b[0m",
                                "",
                                "TASK [osism.services.frr : Pin frr package version] ****************************",
                                "Tuesday 24 February 2026  00:46:37 +0000 (0:00:00.239)       0:00:00.472 ****** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.services.frr : Install frr package] ********************************",
                                "Tuesday 24 February 2026  00:46:38 +0000 (0:00:01.048)       0:00:01.520 ****** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.services.frr : Copy file: /etc/frr/vtysh.conf] *********************",
                                "Tuesday 24 February 2026  00:46:46 +0000 (0:00:08.738)       0:00:10.258 ****** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.services.frr : Copy file: /etc/frr/daemons] ************************",
                                "Tuesday 24 February 2026  00:46:47 +0000 (0:00:01.082)       0:00:11.341 ****** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.services.frr : Set _frr_uplinks fact] ******************************",
                                "Tuesday 24 February 2026  00:46:48 +0000 (0:00:00.968)       0:00:12.309 ****** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.services.frr : Write frr_config_template to temporary file] ********",
                                "Tuesday 24 February 2026  00:46:50 +0000 (0:00:01.179)       0:00:13.489 ****** ",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.services.frr : Render frr.conf from frr_config_template variable] ***",
                                "Tuesday 24 February 2026  00:46:50 +0000 (0:00:00.148)       0:00:13.638 ****** ",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.services.frr : Remove temporary frr_config_template file] **********",
                                "Tuesday 24 February 2026  00:46:50 +0000 (0:00:00.153)       0:00:13.792 ****** ",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.services.frr : Check for frr.conf file in the configuration repository] ***",
                                "Tuesday 24 February 2026  00:46:50 +0000 (0:00:00.155)       0:00:13.948 ****** ",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.services.frr : Copy frr.conf file from the configuration repository] ***",
                                "Tuesday 24 February 2026  00:46:50 +0000 (0:00:00.144)       0:00:14.092 ****** ",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.services.frr : Copy default frr.conf file of type k3s_cilium] ******",
                                "Tuesday 24 February 2026  00:46:50 +0000 (0:00:00.152)       0:00:14.244 ****** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.services.frr : Set sysctl parameters] ******************************",
                                "Tuesday 24 February 2026  00:46:51 +0000 (0:00:01.113)       0:00:15.357 ****** ",
                                "\u001b[0;33mchanged: [testbed-manager] => (item={'name': 'net.ipv4.ip_forward', 'value': 1})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item={'name': 'net.ipv4.conf.all.send_redirects', 'value': 0})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item={'name': 'net.ipv4.conf.all.accept_redirects', 'value': 0})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item={'name': 'net.ipv4.fib_multipath_hash_policy', 'value': 1})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item={'name': 'net.ipv4.conf.default.ignore_routes_with_linkdown', 'value': 1})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item={'name': 'net.ipv4.conf.all.rp_filter', 'value': 2})\u001b[0m",
                                "",
                                "TASK [osism.services.frr : Manage frr service] *********************************",
                                "Tuesday 24 February 2026  00:46:54 +0000 (0:00:02.262)       0:00:17.620 ****** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "",
                                "RUNNING HANDLER [osism.services.frr : Restart frr service] *********************",
                                "Tuesday 24 February 2026  00:46:55 +0000 (0:00:01.205)       0:00:18.825 ****** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;33mtestbed-manager\u001b[0m            : \u001b[0;32mok=10  \u001b[0m \u001b[0;33mchanged=6   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=5   \u001b[0m rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Tuesday 24 February 2026  00:46:56 +0000 (0:00:01.295)       0:00:20.121 ****** ",
                                "=============================================================================== ",
                                "osism.services.frr : Install frr package -------------------------------- 8.74s",
                                "osism.services.frr : Set sysctl parameters ------------------------------ 2.26s",
                                "osism.services.frr : Restart frr service -------------------------------- 1.30s",
                                "osism.services.frr : Manage frr service --------------------------------- 1.21s",
                                "osism.services.frr : Set _frr_uplinks fact ------------------------------ 1.18s",
                                "osism.services.frr : Copy default frr.conf file of type k3s_cilium ------ 1.11s",
                                "osism.services.frr : Copy file: /etc/frr/vtysh.conf --------------------- 1.08s",
                                "osism.services.frr : Pin frr package version ---------------------------- 1.05s",
                                "osism.services.frr : Copy file: /etc/frr/daemons ------------------------ 0.97s",
                                "osism.services.frr : Include distribution specific install tasks -------- 0.24s",
                                "osism.services.frr : Remove temporary frr_config_template file ---------- 0.16s",
                                "osism.services.frr : Render frr.conf from frr_config_template variable --- 0.15s",
                                "osism.services.frr : Copy frr.conf file from the configuration repository --- 0.15s",
                                "osism.services.frr : Write frr_config_template to temporary file -------- 0.15s",
                                "osism.services.frr : Check for frr.conf file in the configuration repository --- 0.14s",
                                "",
                                "--> DEPLOY IN A NUTSHELL -- START -- Tue Feb 24 00:46:57 UTC 2026",
                                "",
                                "\u001b[32m2026-02-24 00:46:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mCollection nutshell is prepared for execution\u001b[0m",
                                "\u001b[32m2026-02-24 00:46:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mA [0] - dotfiles\u001b[0m",
                                "\u001b[32m2026-02-24 00:47:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mA [0] - homer\u001b[0m",
                                "\u001b[32m2026-02-24 00:47:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mA [0] - netdata\u001b[0m",
                                "\u001b[32m2026-02-24 00:47:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mA [0] - openstackclient\u001b[0m",
                                "\u001b[32m2026-02-24 00:47:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mA [0] - phpmyadmin\u001b[0m",
                                "\u001b[32m2026-02-24 00:47:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mA [0] - common\u001b[0m",
                                "\u001b[32m2026-02-24 00:47:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mA [1] -- loadbalancer\u001b[0m",
                                "\u001b[32m2026-02-24 00:47:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mA [2] --- opensearch\u001b[0m",
                                "\u001b[32m2026-02-24 00:47:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mA [2] --- mariadb-ng\u001b[0m",
                                "\u001b[32m2026-02-24 00:47:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mA [3] ---- horizon\u001b[0m",
                                "\u001b[32m2026-02-24 00:47:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mA [3] ---- keystone\u001b[0m",
                                "\u001b[32m2026-02-24 00:47:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mA [4] ----- neutron\u001b[0m",
                                "\u001b[32m2026-02-24 00:47:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mA [5] ------ wait-for-nova\u001b[0m",
                                "\u001b[32m2026-02-24 00:47:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mA [6] ------- octavia\u001b[0m",
                                "\u001b[32m2026-02-24 00:47:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mA [4] ----- barbican\u001b[0m",
                                "\u001b[32m2026-02-24 00:47:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mA [4] ----- designate\u001b[0m",
                                "\u001b[32m2026-02-24 00:47:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mA [4] ----- ironic\u001b[0m",
                                "\u001b[32m2026-02-24 00:47:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mA [4] ----- placement\u001b[0m",
                                "\u001b[32m2026-02-24 00:47:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mA [4] ----- magnum\u001b[0m",
                                "\u001b[32m2026-02-24 00:47:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mA [1] -- openvswitch\u001b[0m",
                                "\u001b[32m2026-02-24 00:47:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mA [2] --- ovn\u001b[0m",
                                "\u001b[32m2026-02-24 00:47:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mA [1] -- memcached\u001b[0m",
                                "\u001b[32m2026-02-24 00:47:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mA [1] -- redis\u001b[0m",
                                "\u001b[32m2026-02-24 00:47:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mA [1] -- rabbitmq-ng\u001b[0m",
                                "\u001b[32m2026-02-24 00:47:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mA [0] - kubernetes\u001b[0m",
                                "\u001b[32m2026-02-24 00:47:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mA [1] -- kubeconfig\u001b[0m",
                                "\u001b[32m2026-02-24 00:47:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mA [1] -- copy-kubeconfig\u001b[0m",
                                "\u001b[32m2026-02-24 00:47:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mA [0] - ceph\u001b[0m",
                                "\u001b[32m2026-02-24 00:47:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mA [1] -- ceph-pools\u001b[0m",
                                "\u001b[32m2026-02-24 00:47:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mA [2] --- copy-ceph-keys\u001b[0m",
                                "\u001b[32m2026-02-24 00:47:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mA [3] ---- cephclient\u001b[0m",
                                "\u001b[32m2026-02-24 00:47:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mA [4] ----- ceph-bootstrap-dashboard\u001b[0m",
                                "\u001b[32m2026-02-24 00:47:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mA [4] ----- wait-for-keystone\u001b[0m",
                                "\u001b[32m2026-02-24 00:47:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mA [5] ------ kolla-ceph-rgw\u001b[0m",
                                "\u001b[32m2026-02-24 00:47:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mA [5] ------ glance\u001b[0m",
                                "\u001b[32m2026-02-24 00:47:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mA [5] ------ cinder\u001b[0m",
                                "\u001b[32m2026-02-24 00:47:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mA [5] ------ nova\u001b[0m",
                                "\u001b[32m2026-02-24 00:47:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mA [4] ----- prometheus\u001b[0m",
                                "\u001b[32m2026-02-24 00:47:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mA [5] ------ grafana\u001b[0m",
                                "\u001b[32m2026-02-24 00:47:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mAll tasks of the collection nutshell are prepared for execution\u001b[0m",
                                "\u001b[32m2026-02-24 00:47:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTasks are running in the background\u001b[0m",
                                "\u001b[32m2026-02-24 00:47:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mNo task IDs specified, wait for all currently running tasks\u001b[0m",
                                "\u001b[32m2026-02-24 00:47:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ff4e45b7-3128-4ce1-95d9-909f33d92301 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:47:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f826982c-1ebc-4d5d-843f-4287d6696757 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:47:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f60b9a97-c893-4bee-b88b-e953f42f36db is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:47:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c72fddd9-9c8a-4ed1-9b6f-116f1787f9e8 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:47:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 991890e5-75ae-4ceb-aafe-9a43e3178763 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:47:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:47:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3803c886-08de-4d66-ab8f-1da8b65e33ca is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:47:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 00:47:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ff4e45b7-3128-4ce1-95d9-909f33d92301 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:47:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f826982c-1ebc-4d5d-843f-4287d6696757 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:47:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f60b9a97-c893-4bee-b88b-e953f42f36db is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:47:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c72fddd9-9c8a-4ed1-9b6f-116f1787f9e8 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:47:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 991890e5-75ae-4ceb-aafe-9a43e3178763 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:47:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:47:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3803c886-08de-4d66-ab8f-1da8b65e33ca is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:47:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 00:47:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ff4e45b7-3128-4ce1-95d9-909f33d92301 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:47:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f826982c-1ebc-4d5d-843f-4287d6696757 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:47:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f60b9a97-c893-4bee-b88b-e953f42f36db is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:47:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c72fddd9-9c8a-4ed1-9b6f-116f1787f9e8 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:47:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 991890e5-75ae-4ceb-aafe-9a43e3178763 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:47:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:47:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3803c886-08de-4d66-ab8f-1da8b65e33ca is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:47:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 00:47:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ff4e45b7-3128-4ce1-95d9-909f33d92301 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:47:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f826982c-1ebc-4d5d-843f-4287d6696757 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:47:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f60b9a97-c893-4bee-b88b-e953f42f36db is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:47:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c72fddd9-9c8a-4ed1-9b6f-116f1787f9e8 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:47:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 991890e5-75ae-4ceb-aafe-9a43e3178763 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:47:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:47:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3803c886-08de-4d66-ab8f-1da8b65e33ca is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:47:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 00:47:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ff4e45b7-3128-4ce1-95d9-909f33d92301 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:47:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f826982c-1ebc-4d5d-843f-4287d6696757 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:47:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f60b9a97-c893-4bee-b88b-e953f42f36db is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:47:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c72fddd9-9c8a-4ed1-9b6f-116f1787f9e8 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:47:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 991890e5-75ae-4ceb-aafe-9a43e3178763 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:47:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:47:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3803c886-08de-4d66-ab8f-1da8b65e33ca is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:47:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 00:47:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ff4e45b7-3128-4ce1-95d9-909f33d92301 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:47:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f826982c-1ebc-4d5d-843f-4287d6696757 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:47:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f60b9a97-c893-4bee-b88b-e953f42f36db is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:47:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c72fddd9-9c8a-4ed1-9b6f-116f1787f9e8 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:47:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 991890e5-75ae-4ceb-aafe-9a43e3178763 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:47:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:47:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3803c886-08de-4d66-ab8f-1da8b65e33ca is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:47:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 00:47:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ff4e45b7-3128-4ce1-95d9-909f33d92301 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:47:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f826982c-1ebc-4d5d-843f-4287d6696757 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:47:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f60b9a97-c893-4bee-b88b-e953f42f36db is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:47:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c72fddd9-9c8a-4ed1-9b6f-116f1787f9e8 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:47:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 991890e5-75ae-4ceb-aafe-9a43e3178763 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:47:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:47:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3803c886-08de-4d66-ab8f-1da8b65e33ca is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:47:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 00:47:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ff4e45b7-3128-4ce1-95d9-909f33d92301 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:47:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f826982c-1ebc-4d5d-843f-4287d6696757 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:47:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f60b9a97-c893-4bee-b88b-e953f42f36db is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:47:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c72fddd9-9c8a-4ed1-9b6f-116f1787f9e8 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:47:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 991890e5-75ae-4ceb-aafe-9a43e3178763 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:47:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:47:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3803c886-08de-4d66-ab8f-1da8b65e33ca is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:47:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "",
                                "PLAY [Apply role geerlingguy.dotfiles] *****************************************",
                                "",
                                "TASK [geerlingguy.dotfiles : Ensure dotfiles repository is cloned locally.] ****",
                                "Tuesday 24 February 2026  00:47:23 +0000 (0:00:00.524)       0:00:00.524 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [geerlingguy.dotfiles : Ensure all configured dotfiles are links.] ********",
                                "Tuesday 24 February 2026  00:47:27 +0000 (0:00:03.858)       0:00:04.382 ****** ",
                                "\u001b[0;32mok: [testbed-manager] => (item=.tmux.conf)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0] => (item=.tmux.conf)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item=.tmux.conf)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item=.tmux.conf)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item=.tmux.conf)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item=.tmux.conf)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item=.tmux.conf)\u001b[0m",
                                "",
                                "TASK [geerlingguy.dotfiles : Remove existing dotfiles file if a replacement is being linked.] ***",
                                "Tuesday 24 February 2026  00:47:29 +0000 (0:00:01.831)       0:00:06.214 ****** ",
                                "\u001b[0;32mok: [testbed-manager] => (item=[0, {'changed': False, 'stdout': '', 'stderr': \"ls: cannot access '/home/dragon/.tmux.conf': No such file or directory\", 'rc': 2, 'cmd': ['ls', '-F', '~/.tmux.conf'], 'start': '2026-02-24 00:47:28.113823', 'end': '2026-02-24 00:47:28.117993', 'delta': '0:00:00.004170', 'failed': False, 'msg': 'non-zero return code', 'invocation': {'module_args': {'_raw_params': 'ls -F ~/.tmux.conf', '_uses_shell': False, 'expand_argument_vars': True, 'stdin_add_newline': True, 'strip_empty_ends': True, 'argv': None, 'chdir': None, 'executable': None, 'creates': None, 'removes': None, 'stdin': None}}, 'stdout_lines': [], 'stderr_lines': [\"ls: cannot access '/home/dragon/.tmux.conf': No such file or directory\"], 'failed_when_result': False, 'item': '.tmux.conf', 'ansible_loop_var': 'item'}])\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0] => (item=[0, {'changed': False, 'stdout': '', 'stderr': \"ls: cannot access '/home/dragon/.tmux.conf': No such file or directory\", 'rc': 2, 'cmd': ['ls', '-F', '~/.tmux.conf'], 'start': '2026-02-24 00:47:28.319974', 'end': '2026-02-24 00:47:28.328652', 'delta': '0:00:00.008678', 'failed': False, 'msg': 'non-zero return code', 'invocation': {'module_args': {'_raw_params': 'ls -F ~/.tmux.conf', '_uses_shell': False, 'expand_argument_vars': True, 'stdin_add_newline': True, 'strip_empty_ends': True, 'argv': None, 'chdir': None, 'executable': None, 'creates': None, 'removes': None, 'stdin': None}}, 'stdout_lines': [], 'stderr_lines': [\"ls: cannot access '/home/dragon/.tmux.conf': No such file or directory\"], 'failed_when_result': False, 'item': '.tmux.conf', 'ansible_loop_var': 'item'}])\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item=[0, {'changed': False, 'stdout': '', 'stderr': \"ls: cannot access '/home/dragon/.tmux.conf': No such file or directory\", 'rc': 2, 'cmd': ['ls', '-F', '~/.tmux.conf'], 'start': '2026-02-24 00:47:28.608254', 'end': '2026-02-24 00:47:28.617681', 'delta': '0:00:00.009427', 'failed': False, 'msg': 'non-zero return code', 'invocation': {'module_args': {'_raw_params': 'ls -F ~/.tmux.conf', '_uses_shell': False, 'expand_argument_vars': True, 'stdin_add_newline': True, 'strip_empty_ends': True, 'argv': None, 'chdir': None, 'executable': None, 'creates': None, 'removes': None, 'stdin': None}}, 'stdout_lines': [], 'stderr_lines': [\"ls: cannot access '/home/dragon/.tmux.conf': No such file or directory\"], 'failed_when_result': False, 'item': '.tmux.conf', 'ansible_loop_var': 'item'}])\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item=[0, {'changed': False, 'stdout': '', 'stderr': \"ls: cannot access '/home/dragon/.tmux.conf': No such file or directory\", 'rc': 2, 'cmd': ['ls', '-F', '~/.tmux.conf'], 'start': '2026-02-24 00:47:28.831013', 'end': '2026-02-24 00:47:28.838516', 'delta': '0:00:00.007503', 'failed': False, 'msg': 'non-zero return code', 'invocation': {'module_args': {'_raw_params': 'ls -F ~/.tmux.conf', '_uses_shell': False, 'expand_argument_vars': True, 'stdin_add_newline': True, 'strip_empty_ends': True, 'argv': None, 'chdir': None, 'executable': None, 'creates': None, 'removes': None, 'stdin': None}}, 'stdout_lines': [], 'stderr_lines': [\"ls: cannot access '/home/dragon/.tmux.conf': No such file or directory\"], 'failed_when_result': False, 'item': '.tmux.conf', 'ansible_loop_var': 'item'}])\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item=[0, {'changed': False, 'stdout': '', 'stderr': \"ls: cannot access '/home/dragon/.tmux.conf': No such file or directory\", 'rc': 2, 'cmd': ['ls', '-F', '~/.tmux.conf'], 'start': '2026-02-24 00:47:29.019007', 'end': '2026-02-24 00:47:29.030269', 'delta': '0:00:00.011262', 'failed': False, 'msg': 'non-zero return code', 'invocation': {'module_args': {'_raw_params': 'ls -F ~/.tmux.conf', '_uses_shell': False, 'expand_argument_vars': True, 'stdin_add_newline': True, 'strip_empty_ends': True, 'argv': None, 'chdir': None, 'executable': None, 'creates': None, 'removes': None, 'stdin': None}}, 'stdout_lines': [], 'stderr_lines': [\"ls: cannot access '/home/dragon/.tmux.conf': No such file or directory\"], 'failed_when_result': False, 'item': '.tmux.conf', 'ansible_loop_var': 'item'}])\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item=[0, {'changed': False, 'stdout': '', 'stderr': \"ls: cannot access '/home/dragon/.tmux.conf': No such file or directory\", 'rc': 2, 'cmd': ['ls', '-F', '~/.tmux.conf'], 'start': '2026-02-24 00:47:29.029611', 'end': '2026-02-24 00:47:29.037746', 'delta': '0:00:00.008135', 'failed': False, 'msg': 'non-zero return code', 'invocation': {'module_args': {'_raw_params': 'ls -F ~/.tmux.conf', '_uses_shell': False, 'expand_argument_vars': True, 'stdin_add_newline': True, 'strip_empty_ends': True, 'argv': None, 'chdir': None, 'executable': None, 'creates': None, 'removes': None, 'stdin': None}}, 'stdout_lines': [], 'stderr_lines': [\"ls: cannot access '/home/dragon/.tmux.conf': No such file or directory\"], 'failed_when_result': False, 'item': '.tmux.conf', 'ansible_loop_var': 'item'}])\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item=[0, {'changed': False, 'stdout': '', 'stderr': \"ls: cannot access '/home/dragon/.tmux.conf': No such file or directory\", 'rc': 2, 'cmd': ['ls', '-F', '~/.tmux.conf'], 'start': '2026-02-24 00:47:29.083881', 'end': '2026-02-24 00:47:29.090232', 'delta': '0:00:00.006351', 'failed': False, 'msg': 'non-zero return code', 'invocation': {'module_args': {'_raw_params': 'ls -F ~/.tmux.conf', '_uses_shell': False, 'expand_argument_vars': True, 'stdin_add_newline': True, 'strip_empty_ends': True, 'argv': None, 'chdir': None, 'executable': None, 'creates': None, 'removes': None, 'stdin': None}}, 'stdout_lines': [], 'stderr_lines': [\"ls: cannot access '/home/dragon/.tmux.conf': No such file or directory\"], 'failed_when_result': False, 'item': '.tmux.conf', 'ansible_loop_var': 'item'}])\u001b[0m",
                                "",
                                "TASK [geerlingguy.dotfiles : Ensure parent folders of link dotfiles exist.] ****",
                                "Tuesday 24 February 2026  00:47:32 +0000 (0:00:02.917)       0:00:09.131 ****** ",
                                "\u001b[0;32mok: [testbed-node-0] => (item=.tmux.conf)\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager] => (item=.tmux.conf)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item=.tmux.conf)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item=.tmux.conf)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item=.tmux.conf)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item=.tmux.conf)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item=.tmux.conf)\u001b[0m",
                                "",
                                "TASK [geerlingguy.dotfiles : Link dotfiles into home folder.] ******************",
                                "Tuesday 24 February 2026  00:47:34 +0000 (0:00:01.821)       0:00:10.953 ****** ",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=.tmux.conf)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=.tmux.conf)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=.tmux.conf)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=.tmux.conf)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=.tmux.conf)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=.tmux.conf)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=.tmux.conf)\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;33mtestbed-manager\u001b[0m            : \u001b[0;32mok=5   \u001b[0m \u001b[0;33mchanged=2   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=5   \u001b[0m \u001b[0;33mchanged=2   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=5   \u001b[0m \u001b[0;33mchanged=2   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=5   \u001b[0m \u001b[0;33mchanged=2   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-3\u001b[0m             : \u001b[0;32mok=5   \u001b[0m \u001b[0;33mchanged=2   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-4\u001b[0m             : \u001b[0;32mok=5   \u001b[0m \u001b[0;33mchanged=2   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-5\u001b[0m             : \u001b[0;32mok=5   \u001b[0m \u001b[0;33mchanged=2   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Tuesday 24 February 2026  00:47:35 +0000 (0:00:01.862)       0:00:12.816 ****** ",
                                "=============================================================================== ",
                                "geerlingguy.dotfiles : Ensure dotfiles repository is cloned locally. ---- 3.86s",
                                "geerlingguy.dotfiles : Remove existing dotfiles file if a replacement is being linked. --- 2.92s",
                                "geerlingguy.dotfiles : Link dotfiles into home folder. ------------------ 1.86s",
                                "geerlingguy.dotfiles : Ensure all configured dotfiles are links. -------- 1.83s",
                                "geerlingguy.dotfiles : Ensure parent folders of link dotfiles exist. ---- 1.82s",
                                "\u001b[32m2026-02-24 00:47:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ff4e45b7-3128-4ce1-95d9-909f33d92301 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:47:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f826982c-1ebc-4d5d-843f-4287d6696757 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:47:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f60b9a97-c893-4bee-b88b-e953f42f36db is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:47:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask da7eb094-0467-4cde-920c-d33e5527beb2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:47:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c72fddd9-9c8a-4ed1-9b6f-116f1787f9e8 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:47:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 991890e5-75ae-4ceb-aafe-9a43e3178763 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:47:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:47:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3803c886-08de-4d66-ab8f-1da8b65e33ca is in state SUCCESS\u001b[0m",
                                "\u001b[32m2026-02-24 00:47:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 00:47:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ff4e45b7-3128-4ce1-95d9-909f33d92301 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:47:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f826982c-1ebc-4d5d-843f-4287d6696757 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:47:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f60b9a97-c893-4bee-b88b-e953f42f36db is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:47:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask da7eb094-0467-4cde-920c-d33e5527beb2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:47:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c72fddd9-9c8a-4ed1-9b6f-116f1787f9e8 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:47:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 991890e5-75ae-4ceb-aafe-9a43e3178763 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:47:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:47:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 00:47:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ff4e45b7-3128-4ce1-95d9-909f33d92301 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:47:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f826982c-1ebc-4d5d-843f-4287d6696757 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:47:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f60b9a97-c893-4bee-b88b-e953f42f36db is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:47:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask da7eb094-0467-4cde-920c-d33e5527beb2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:47:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c72fddd9-9c8a-4ed1-9b6f-116f1787f9e8 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:47:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 991890e5-75ae-4ceb-aafe-9a43e3178763 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:47:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:47:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 00:47:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ff4e45b7-3128-4ce1-95d9-909f33d92301 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:47:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f826982c-1ebc-4d5d-843f-4287d6696757 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:47:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f60b9a97-c893-4bee-b88b-e953f42f36db is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:47:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask da7eb094-0467-4cde-920c-d33e5527beb2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:47:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c72fddd9-9c8a-4ed1-9b6f-116f1787f9e8 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:47:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 991890e5-75ae-4ceb-aafe-9a43e3178763 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:47:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:47:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 00:47:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ff4e45b7-3128-4ce1-95d9-909f33d92301 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:47:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f826982c-1ebc-4d5d-843f-4287d6696757 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:47:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f60b9a97-c893-4bee-b88b-e953f42f36db is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:47:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask da7eb094-0467-4cde-920c-d33e5527beb2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:47:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c72fddd9-9c8a-4ed1-9b6f-116f1787f9e8 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:47:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 991890e5-75ae-4ceb-aafe-9a43e3178763 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:47:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:47:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 00:47:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ff4e45b7-3128-4ce1-95d9-909f33d92301 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:47:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f826982c-1ebc-4d5d-843f-4287d6696757 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:47:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f60b9a97-c893-4bee-b88b-e953f42f36db is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:47:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask da7eb094-0467-4cde-920c-d33e5527beb2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:47:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c72fddd9-9c8a-4ed1-9b6f-116f1787f9e8 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:47:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 991890e5-75ae-4ceb-aafe-9a43e3178763 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:47:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:47:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 00:47:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ff4e45b7-3128-4ce1-95d9-909f33d92301 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:47:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f826982c-1ebc-4d5d-843f-4287d6696757 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:47:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f60b9a97-c893-4bee-b88b-e953f42f36db is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:47:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask da7eb094-0467-4cde-920c-d33e5527beb2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:47:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c72fddd9-9c8a-4ed1-9b6f-116f1787f9e8 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:47:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 991890e5-75ae-4ceb-aafe-9a43e3178763 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:47:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:47:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 00:48:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ff4e45b7-3128-4ce1-95d9-909f33d92301 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:48:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f826982c-1ebc-4d5d-843f-4287d6696757 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:48:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f60b9a97-c893-4bee-b88b-e953f42f36db is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:48:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask da7eb094-0467-4cde-920c-d33e5527beb2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:48:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c72fddd9-9c8a-4ed1-9b6f-116f1787f9e8 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:48:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 991890e5-75ae-4ceb-aafe-9a43e3178763 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:48:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:48:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 00:48:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ff4e45b7-3128-4ce1-95d9-909f33d92301 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:48:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f826982c-1ebc-4d5d-843f-4287d6696757 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:48:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f60b9a97-c893-4bee-b88b-e953f42f36db is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:48:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask da7eb094-0467-4cde-920c-d33e5527beb2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:48:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c72fddd9-9c8a-4ed1-9b6f-116f1787f9e8 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:48:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 991890e5-75ae-4ceb-aafe-9a43e3178763 is in state SUCCESS\u001b[0m",
                                "\u001b[32m2026-02-24 00:48:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:48:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 00:48:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ff4e45b7-3128-4ce1-95d9-909f33d92301 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:48:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f826982c-1ebc-4d5d-843f-4287d6696757 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:48:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f60b9a97-c893-4bee-b88b-e953f42f36db is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:48:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask da7eb094-0467-4cde-920c-d33e5527beb2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:48:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c72fddd9-9c8a-4ed1-9b6f-116f1787f9e8 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:48:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:48:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 00:48:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ff4e45b7-3128-4ce1-95d9-909f33d92301 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:48:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f826982c-1ebc-4d5d-843f-4287d6696757 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:48:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f60b9a97-c893-4bee-b88b-e953f42f36db is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:48:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask da7eb094-0467-4cde-920c-d33e5527beb2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:48:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c72fddd9-9c8a-4ed1-9b6f-116f1787f9e8 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:48:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:48:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 00:48:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ff4e45b7-3128-4ce1-95d9-909f33d92301 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:48:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f826982c-1ebc-4d5d-843f-4287d6696757 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:48:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f60b9a97-c893-4bee-b88b-e953f42f36db is in state SUCCESS\u001b[0m",
                                "\u001b[32m2026-02-24 00:48:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask da7eb094-0467-4cde-920c-d33e5527beb2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:48:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c72fddd9-9c8a-4ed1-9b6f-116f1787f9e8 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:48:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:48:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 00:48:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ff4e45b7-3128-4ce1-95d9-909f33d92301 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:48:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f826982c-1ebc-4d5d-843f-4287d6696757 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:48:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask da7eb094-0467-4cde-920c-d33e5527beb2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:48:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c72fddd9-9c8a-4ed1-9b6f-116f1787f9e8 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:48:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:48:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 00:48:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ff4e45b7-3128-4ce1-95d9-909f33d92301 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:48:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f826982c-1ebc-4d5d-843f-4287d6696757 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:48:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask da7eb094-0467-4cde-920c-d33e5527beb2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:48:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c72fddd9-9c8a-4ed1-9b6f-116f1787f9e8 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:48:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:48:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 00:48:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ff4e45b7-3128-4ce1-95d9-909f33d92301 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:48:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f826982c-1ebc-4d5d-843f-4287d6696757 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:48:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask da7eb094-0467-4cde-920c-d33e5527beb2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:48:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c72fddd9-9c8a-4ed1-9b6f-116f1787f9e8 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:48:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:48:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 00:48:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ff4e45b7-3128-4ce1-95d9-909f33d92301 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:48:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f826982c-1ebc-4d5d-843f-4287d6696757 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:48:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask da7eb094-0467-4cde-920c-d33e5527beb2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:48:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c72fddd9-9c8a-4ed1-9b6f-116f1787f9e8 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:48:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:48:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 00:48:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ff4e45b7-3128-4ce1-95d9-909f33d92301 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:48:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f826982c-1ebc-4d5d-843f-4287d6696757 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:48:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask da7eb094-0467-4cde-920c-d33e5527beb2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:48:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c72fddd9-9c8a-4ed1-9b6f-116f1787f9e8 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:48:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:48:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 00:48:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ff4e45b7-3128-4ce1-95d9-909f33d92301 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:48:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f826982c-1ebc-4d5d-843f-4287d6696757 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:48:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask da7eb094-0467-4cde-920c-d33e5527beb2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:48:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c72fddd9-9c8a-4ed1-9b6f-116f1787f9e8 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:48:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:48:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 00:48:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ff4e45b7-3128-4ce1-95d9-909f33d92301 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:48:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f826982c-1ebc-4d5d-843f-4287d6696757 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:48:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask da7eb094-0467-4cde-920c-d33e5527beb2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:48:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c72fddd9-9c8a-4ed1-9b6f-116f1787f9e8 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:48:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:48:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 00:48:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ff4e45b7-3128-4ce1-95d9-909f33d92301 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:48:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f826982c-1ebc-4d5d-843f-4287d6696757 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:48:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask da7eb094-0467-4cde-920c-d33e5527beb2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:48:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c72fddd9-9c8a-4ed1-9b6f-116f1787f9e8 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:48:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:48:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 00:48:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ff4e45b7-3128-4ce1-95d9-909f33d92301 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:48:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f826982c-1ebc-4d5d-843f-4287d6696757 is in state STARTED\u001b[0m",
                                "",
                                "",
                                "PLAY [Apply role homer] ********************************************************",
                                "",
                                "TASK [osism.services.homer : Inform about new parameter homer_url_opensearch_dashboards] ***",
                                "Tuesday 24 February 2026  00:47:23 +0000 (0:00:00.536)       0:00:00.536 ****** ",
                                "\u001b[0;32mok: [testbed-manager] => {\u001b[0m",
                                "\u001b[0;32m    \"msg\": \"The support for the homer_url_kibana has been removed. Please use the homer_url_opensearch_dashboards parameter.\"\u001b[0m",
                                "\u001b[0;32m}\u001b[0m",
                                "",
                                "TASK [osism.services.homer : Create traefik external network] ******************",
                                "Tuesday 24 February 2026  00:47:23 +0000 (0:00:00.145)       0:00:00.681 ****** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.services.homer : Create required directories] **********************",
                                "Tuesday 24 February 2026  00:47:26 +0000 (0:00:02.987)       0:00:03.669 ****** ",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=/opt/homer/configuration)\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager] => (item=/opt/homer)\u001b[0m",
                                "",
                                "TASK [osism.services.homer : Copy config.yml configuration file] ***************",
                                "Tuesday 24 February 2026  00:47:28 +0000 (0:00:02.054)       0:00:05.723 ****** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.services.homer : Copy docker-compose.yml file] *********************",
                                "Tuesday 24 February 2026  00:47:31 +0000 (0:00:03.272)       0:00:08.995 ****** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.services.homer : Manage homer service] *****************************",
                                "Tuesday 24 February 2026  00:47:33 +0000 (0:00:01.655)       0:00:10.651 ****** ",
                                "\u001b[1;30mFAILED - RETRYING: [testbed-manager]: Manage homer service (10 retries left).\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "",
                                "RUNNING HANDLER [osism.services.homer : Restart homer service] *****************",
                                "Tuesday 24 February 2026  00:48:00 +0000 (0:00:27.336)       0:00:37.987 ****** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;33mtestbed-manager\u001b[0m            : \u001b[0;32mok=7   \u001b[0m \u001b[0;33mchanged=4   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Tuesday 24 February 2026  00:48:03 +0000 (0:00:02.318)       0:00:40.306 ****** ",
                                "=============================================================================== ",
                                "osism.services.homer : Manage homer service ---------------------------- 27.34s",
                                "osism.services.homer : Copy config.yml configuration file --------------- 3.27s",
                                "osism.services.homer : Create traefik external network ------------------ 2.99s",
                                "osism.services.homer : Restart homer service ---------------------------- 2.32s",
                                "osism.services.homer : Create required directories ---------------------- 2.05s",
                                "osism.services.homer : Copy docker-compose.yml file --------------------- 1.66s",
                                "osism.services.homer : Inform about new parameter homer_url_opensearch_dashboards --- 0.15s",
                                "",
                                "",
                                "PLAY [Apply role openstackclient] **********************************************",
                                "",
                                "TASK [osism.services.openstackclient : Include tasks] **************************",
                                "Tuesday 24 February 2026  00:47:24 +0000 (0:00:00.850)       0:00:00.850 ****** ",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/services/roles/openstackclient/tasks/container-Debian-family.yml for testbed-manager\u001b[0m",
                                "",
                                "TASK [osism.services.openstackclient : Create required directories] ************",
                                "Tuesday 24 February 2026  00:47:25 +0000 (0:00:00.868)       0:00:01.718 ****** ",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=/opt/configuration/environments/openstack)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=/opt/openstackclient/data)\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager] => (item=/opt/openstackclient)\u001b[0m",
                                "",
                                "TASK [osism.services.openstackclient : Copy docker-compose.yml file] ***********",
                                "Tuesday 24 February 2026  00:47:27 +0000 (0:00:02.235)       0:00:03.954 ****** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.services.openstackclient : Manage openstackclient service] *********",
                                "Tuesday 24 February 2026  00:47:30 +0000 (0:00:03.090)       0:00:07.044 ****** ",
                                "\u001b[1;30mFAILED - RETRYING: [testbed-manager]: Manage openstackclient service (10 retries left).\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.services.openstackclient : Copy openstack wrapper script] **********",
                                "Tuesday 24 February 2026  00:48:03 +0000 (0:00:33.482)       0:00:40.527 ****** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.services.openstackclient : Remove ospurge wrapper script] **********",
                                "Tuesday 24 February 2026  00:48:05 +0000 (0:00:02.038)       0:00:42.566 ****** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "",
                                "RUNNING HANDLER [osism.services.openstackclient : Restart openstackclient service] ***",
                                "Tuesday 24 February 2026  00:48:07 +0000 (0:00:01.616)       0:00:44.182 ****** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "RUNNING HANDLER [osism.services.openstackclient : Ensure that all containers are up] ***",
                                "Tuesday 24 February 2026  00:48:11 +0000 (0:00:03.591)       0:00:47.774 ****** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "RUNNING HANDLER [osism.services.openstackclient : Wait for an healthy service] ***",
                                "Tuesday 24 February 2026  00:48:12 +0000 (0:00:01.063)       0:00:48.837 ****** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "RUNNING HANDLER [osism.services.openstackclient : Copy bash completion script] ***",
                                "Tuesday 24 February 2026  00:48:12 +0000 (0:00:00.486)       0:00:49.323 ****** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;33mtestbed-manager\u001b[0m            : \u001b[0;32mok=10  \u001b[0m \u001b[0;33mchanged=6   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Tuesday 24 February 2026  00:48:13 +0000 (0:00:00.434)       0:00:49.758 ****** ",
                                "=============================================================================== ",
                                "osism.services.openstackclient : Manage openstackclient service -------- 33.48s",
                                "osism.services.openstackclient : Restart openstackclient service -------- 3.59s",
                                "osism.services.openstackclient : Copy docker-compose.yml file ----------- 3.09s",
                                "osism.services.openstackclient : Create required directories ------------ 2.24s",
                                "osism.services.openstackclient : Copy openstack wrapper script ---------- 2.04s",
                                "osism.services.openstackclient : Remove ospurge wrapper script ---------- 1.62s",
                                "osism.services.openstackclient : Ensure that all containers are up ------ 1.06s",
                                "osism.services.openstackclient : Include tasks -------------------------- 0.87s",
                                "osism.services.openstackclient : Wait for an healthy service ------------ 0.49s",
                                "osism.services.openstackclient : Copy bash completion script ------------ 0.43s",
                                "",
                                "",
                                "PLAY [Apply role phpmyadmin] ***************************************************",
                                "",
                                "TASK [osism.services.phpmyadmin : Create traefik external network] *************",
                                "Tuesday 24 February 2026  00:47:42 +0000 (0:00:00.401)       0:00:00.401 ****** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.services.phpmyadmin : Create required directories] *****************",
                                "Tuesday 24 February 2026  00:47:43 +0000 (0:00:01.010)       0:00:01.411 ****** ",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=/opt/phpmyadmin)\u001b[0m",
                                "",
                                "TASK [osism.services.phpmyadmin : Copy docker-compose.yml file] ****************",
                                "Tuesday 24 February 2026  00:47:44 +0000 (0:00:00.538)       0:00:01.950 ****** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.services.phpmyadmin : Manage phpmyadmin service] *******************",
                                "Tuesday 24 February 2026  00:47:45 +0000 (0:00:01.092)       0:00:03.042 ****** ",
                                "\u001b[1;30mFAILED - RETRYING: [testbed-manager]: Manage phpmyadmin service (10 retries left).\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "",
                                "RUNNING HANDLER [osism.services.phpmyadmin : Restart phpmyadmin service] *******",
                                "Tuesday 24 February 2026  00:48:36 +0000 (0:00:50.758)       0:00:53.800 ****** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;33mtestbed-manager\u001b[0m            : \u001b[0;32mok=5   \u001b[0m \u001b[0;33mchanged=3   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Tuesday 24 February 2026  00:48:39 +0000 (0:00:03.634)       0:00:57.435 ****** ",
                                "=============================================================================== ",
                                "osism.services.phpmyadmin : Manage phpmyadmin service ------------------ 50.76s",
                                "osism.services.phpmyadmin : Restart phpmyadmin service ------------------ 3.63s",
                                "osism.services.phpmyadmin : Copy docker-compose.yml file ---------------- 1.09s",
                                "osism.services.phpmyadmin : Create traefik external network ------------- 1.01s",
                                "osism.services.phpmyadmin : Create required directories ----------------- 0.54s",
                                "\u001b[32m2026-02-24 00:48:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask da7eb094-0467-4cde-920c-d33e5527beb2 is in state SUCCESS\u001b[0m",
                                "\u001b[32m2026-02-24 00:48:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c72fddd9-9c8a-4ed1-9b6f-116f1787f9e8 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:48:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:48:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 00:48:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ff4e45b7-3128-4ce1-95d9-909f33d92301 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:48:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f826982c-1ebc-4d5d-843f-4287d6696757 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:48:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c72fddd9-9c8a-4ed1-9b6f-116f1787f9e8 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:48:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:48:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 00:48:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ff4e45b7-3128-4ce1-95d9-909f33d92301 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:48:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f826982c-1ebc-4d5d-843f-4287d6696757 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:48:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c72fddd9-9c8a-4ed1-9b6f-116f1787f9e8 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:48:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:48:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 00:48:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ff4e45b7-3128-4ce1-95d9-909f33d92301 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:48:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f826982c-1ebc-4d5d-843f-4287d6696757 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:48:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c72fddd9-9c8a-4ed1-9b6f-116f1787f9e8 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:48:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:48:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 00:48:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ff4e45b7-3128-4ce1-95d9-909f33d92301 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:48:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f826982c-1ebc-4d5d-843f-4287d6696757 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:48:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c72fddd9-9c8a-4ed1-9b6f-116f1787f9e8 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:48:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:48:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 00:48:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ff4e45b7-3128-4ce1-95d9-909f33d92301 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:48:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f826982c-1ebc-4d5d-843f-4287d6696757 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:48:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c72fddd9-9c8a-4ed1-9b6f-116f1787f9e8 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:48:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:48:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "",
                                "",
                                "PLAY [Group hosts based on configuration] **************************************",
                                "",
                                "TASK [Group hosts based on enabled services] ***********************************",
                                "Tuesday 24 February 2026  00:47:23 +0000 (0:00:00.461)       0:00:00.461 ****** ",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=enable_netdata_True)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=enable_netdata_True)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=enable_netdata_True)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=enable_netdata_True)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=enable_netdata_True)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=enable_netdata_True)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=enable_netdata_True)\u001b[0m",
                                "",
                                "PLAY [Apply role netdata] ******************************************************",
                                "",
                                "TASK [osism.services.netdata : Include distribution specific install tasks] ****",
                                "Tuesday 24 February 2026  00:47:25 +0000 (0:00:01.694)       0:00:02.156 ****** ",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/services/roles/netdata/tasks/install-Debian-family.yml for testbed-manager, testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m",
                                "",
                                "TASK [osism.services.netdata : Remove old architecture-dependent repository] ***",
                                "Tuesday 24 February 2026  00:47:26 +0000 (0:00:01.552)       0:00:03.709 ****** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.services.netdata : Install apt-transport-https package] ************",
                                "Tuesday 24 February 2026  00:47:28 +0000 (0:00:02.134)       0:00:05.843 ****** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.services.netdata : Add repository gpg key] *************************",
                                "Tuesday 24 February 2026  00:47:32 +0000 (0:00:03.965)       0:00:09.809 ****** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [osism.services.netdata : Add repository] *********************************",
                                "Tuesday 24 February 2026  00:47:35 +0000 (0:00:02.722)       0:00:12.531 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.services.netdata : Install package netdata] ************************",
                                "Tuesday 24 February 2026  00:47:47 +0000 (0:00:11.477)       0:00:24.009 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.services.netdata : Include config tasks] ***************************",
                                "Tuesday 24 February 2026  00:48:29 +0000 (0:00:42.307)       0:01:06.316 ****** ",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/services/roles/netdata/tasks/config.yml for testbed-manager, testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m",
                                "",
                                "TASK [osism.services.netdata : Copy configuration files] ***********************",
                                "Tuesday 24 February 2026  00:48:30 +0000 (0:00:01.326)       0:01:07.643 ****** ",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=netdata.conf)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=netdata.conf)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=netdata.conf)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=netdata.conf)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=netdata.conf)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=netdata.conf)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=netdata.conf)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=stream.conf)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=stream.conf)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=stream.conf)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=stream.conf)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=stream.conf)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=stream.conf)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=stream.conf)\u001b[0m",
                                "",
                                "TASK [osism.services.netdata : Retrieve /etc/netdata/.opt-out-from-anonymous-statistics status] ***",
                                "Tuesday 24 February 2026  00:48:36 +0000 (0:00:05.356)       0:01:12.999 ****** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.services.netdata : Opt out from anonymous statistics] **************",
                                "Tuesday 24 February 2026  00:48:37 +0000 (0:00:01.346)       0:01:14.346 ****** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.services.netdata : Add netdata user to docker group] ***************",
                                "Tuesday 24 February 2026  00:48:39 +0000 (0:00:02.002)       0:01:16.349 ****** ",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.services.netdata : Manage service netdata] *************************",
                                "Tuesday 24 February 2026  00:48:41 +0000 (0:00:01.752)       0:01:18.101 ****** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [osism.services.netdata : Include host type specific tasks] ***************",
                                "Tuesday 24 February 2026  00:48:43 +0000 (0:00:01.935)       0:01:20.036 ****** ",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/services/roles/netdata/tasks/server.yml for testbed-manager\u001b[0m",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/services/roles/netdata/tasks/client.yml for testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m",
                                "",
                                "TASK [osism.services.netdata : Set sysctl vm.max_map_count parameter] **********",
                                "Tuesday 24 February 2026  00:48:44 +0000 (0:00:01.167)       0:01:21.204 ****** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "RUNNING HANDLER [osism.services.netdata : Restart service netdata] *************",
                                "Tuesday 24 February 2026  00:48:45 +0000 (0:00:01.736)       0:01:22.941 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;33mtestbed-manager\u001b[0m            : \u001b[0;32mok=16  \u001b[0m \u001b[0;33mchanged=8   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=15  \u001b[0m \u001b[0;33mchanged=7   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=15  \u001b[0m \u001b[0;33mchanged=7   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=15  \u001b[0m \u001b[0;33mchanged=7   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-3\u001b[0m             : \u001b[0;32mok=15  \u001b[0m \u001b[0;33mchanged=7   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-4\u001b[0m             : \u001b[0;32mok=15  \u001b[0m \u001b[0;33mchanged=7   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-5\u001b[0m             : \u001b[0;32mok=15  \u001b[0m \u001b[0;33mchanged=7   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Tuesday 24 February 2026  00:48:57 +0000 (0:00:11.256)       0:01:34.197 ****** ",
                                "=============================================================================== ",
                                "osism.services.netdata : Install package netdata ----------------------- 42.31s",
                                "osism.services.netdata : Add repository -------------------------------- 11.48s",
                                "osism.services.netdata : Restart service netdata ----------------------- 11.26s",
                                "osism.services.netdata : Copy configuration files ----------------------- 5.36s",
                                "osism.services.netdata : Install apt-transport-https package ------------ 3.97s",
                                "osism.services.netdata : Add repository gpg key ------------------------- 2.72s",
                                "osism.services.netdata : Remove old architecture-dependent repository --- 2.13s",
                                "osism.services.netdata : Opt out from anonymous statistics -------------- 2.00s",
                                "osism.services.netdata : Manage service netdata ------------------------- 1.94s",
                                "osism.services.netdata : Add netdata user to docker group --------------- 1.75s",
                                "osism.services.netdata : Set sysctl vm.max_map_count parameter ---------- 1.74s",
                                "Group hosts based on enabled services ----------------------------------- 1.69s",
                                "osism.services.netdata : Include distribution specific install tasks ---- 1.55s",
                                "osism.services.netdata : Retrieve /etc/netdata/.opt-out-from-anonymous-statistics status --- 1.35s",
                                "osism.services.netdata : Include config tasks --------------------------- 1.33s",
                                "osism.services.netdata : Include host type specific tasks --------------- 1.17s",
                                "\u001b[32m2026-02-24 00:48:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ff4e45b7-3128-4ce1-95d9-909f33d92301 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:48:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f826982c-1ebc-4d5d-843f-4287d6696757 is in state SUCCESS\u001b[0m",
                                "\u001b[32m2026-02-24 00:48:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c72fddd9-9c8a-4ed1-9b6f-116f1787f9e8 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:48:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:48:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 00:49:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ff4e45b7-3128-4ce1-95d9-909f33d92301 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:49:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c72fddd9-9c8a-4ed1-9b6f-116f1787f9e8 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:49:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:49:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 00:49:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ff4e45b7-3128-4ce1-95d9-909f33d92301 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:49:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c72fddd9-9c8a-4ed1-9b6f-116f1787f9e8 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:49:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:49:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 00:49:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ff4e45b7-3128-4ce1-95d9-909f33d92301 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:49:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c72fddd9-9c8a-4ed1-9b6f-116f1787f9e8 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:49:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:49:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 00:49:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ff4e45b7-3128-4ce1-95d9-909f33d92301 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:49:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c72fddd9-9c8a-4ed1-9b6f-116f1787f9e8 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:49:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:49:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 00:49:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ff4e45b7-3128-4ce1-95d9-909f33d92301 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:49:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c72fddd9-9c8a-4ed1-9b6f-116f1787f9e8 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:49:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:49:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 00:49:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ff4e45b7-3128-4ce1-95d9-909f33d92301 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:49:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c72fddd9-9c8a-4ed1-9b6f-116f1787f9e8 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:49:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:49:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 00:49:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ff4e45b7-3128-4ce1-95d9-909f33d92301 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:49:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c72fddd9-9c8a-4ed1-9b6f-116f1787f9e8 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:49:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:49:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 00:49:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ff4e45b7-3128-4ce1-95d9-909f33d92301 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:49:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c72fddd9-9c8a-4ed1-9b6f-116f1787f9e8 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:49:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:49:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 00:49:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ff4e45b7-3128-4ce1-95d9-909f33d92301 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:49:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c72fddd9-9c8a-4ed1-9b6f-116f1787f9e8 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:49:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:49:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 00:49:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ff4e45b7-3128-4ce1-95d9-909f33d92301 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:49:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c72fddd9-9c8a-4ed1-9b6f-116f1787f9e8 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:49:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:49:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 00:49:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ff4e45b7-3128-4ce1-95d9-909f33d92301 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:49:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f8694a2a-a9eb-4540-948b-ced7e58054f8 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:49:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c72fddd9-9c8a-4ed1-9b6f-116f1787f9e8 is in state SUCCESS\u001b[0m",
                                "",
                                "",
                                "PLAY [Apply role common] *******************************************************",
                                "",
                                "TASK [common : include_tasks] **************************************************",
                                "Tuesday 24 February 2026  00:47:14 +0000 (0:00:00.237)       0:00:00.238 ****** ",
                                "\u001b[0;36mincluded: /ansible/roles/common/tasks/deploy.yml for testbed-manager, testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m",
                                "",
                                "TASK [common : Ensuring config directories exist] ******************************",
                                "Tuesday 24 February 2026  00:47:15 +0000 (0:00:01.442)       0:00:01.680 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=[{'service_name': 'cron'}, 'cron'])\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=[{'service_name': 'cron'}, 'cron'])\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=[{'service_name': 'cron'}, 'cron'])\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=[{'service_name': 'fluentd'}, 'fluentd'])\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=[{'service_name': 'cron'}, 'cron'])\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=[{'service_name': 'cron'}, 'cron'])\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=[{'service_name': 'cron'}, 'cron'])\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=[{'service_name': 'cron'}, 'cron'])\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=[{'service_name': 'fluentd'}, 'fluentd'])\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=[{'service_name': 'kolla-toolbox'}, 'kolla-toolbox'])\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=[{'service_name': 'fluentd'}, 'fluentd'])\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=[{'service_name': 'fluentd'}, 'fluentd'])\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=[{'service_name': 'fluentd'}, 'fluentd'])\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=[{'service_name': 'fluentd'}, 'fluentd'])\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=[{'service_name': 'fluentd'}, 'fluentd'])\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=[{'service_name': 'kolla-toolbox'}, 'kolla-toolbox'])\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=[{'service_name': 'kolla-toolbox'}, 'kolla-toolbox'])\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=[{'service_name': 'kolla-toolbox'}, 'kolla-toolbox'])\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=[{'service_name': 'kolla-toolbox'}, 'kolla-toolbox'])\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=[{'service_name': 'kolla-toolbox'}, 'kolla-toolbox'])\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=[{'service_name': 'kolla-toolbox'}, 'kolla-toolbox'])\u001b[0m",
                                "",
                                "TASK [common : include_tasks] **************************************************",
                                "Tuesday 24 February 2026  00:47:19 +0000 (0:00:04.142)       0:00:05.822 ****** ",
                                "\u001b[0;36mincluded: /ansible/roles/common/tasks/copy-certs.yml for testbed-manager, testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m",
                                "",
                                "TASK [service-cert-copy : common | Copying over extra CA certificates] *********",
                                "Tuesday 24 February 2026  00:47:21 +0000 (0:00:01.559)       0:00:07.382 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2024.2', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2024.2', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2024.2', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2024.2', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2024.2', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2024.2', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2024.2', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2024.2', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2024.2', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2024.2', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2024.2', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2024.2', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2024.2', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2024.2', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2024.2', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2024.2', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2024.2', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2024.2', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2024.2', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2024.2', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2024.2', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "",
                                "TASK [service-cert-copy : common | Copying over backend internal TLS certificate] ***",
                                "Tuesday 24 February 2026  00:47:26 +0000 (0:00:05.066)       0:00:12.448 ****** ",
                                "\u001b[0;36mskipping: [testbed-manager] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2024.2', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2024.2', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2024.2', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2024.2', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2024.2', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2024.2', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2024.2', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2024.2', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2024.2', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2024.2', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2024.2', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2024.2', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2024.2', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2024.2', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2024.2', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2024.2', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2024.2', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2024.2', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2024.2', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2024.2', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2024.2', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [service-cert-copy : common | Copying over backend internal TLS key] ******",
                                "Tuesday 24 February 2026  00:47:28 +0000 (0:00:02.141)       0:00:14.590 ****** ",
                                "\u001b[0;36mskipping: [testbed-manager] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2024.2', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2024.2', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2024.2', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2024.2', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2024.2', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2024.2', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2024.2', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2024.2', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2024.2', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2024.2', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2024.2', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2024.2', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2024.2', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2024.2', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2024.2', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2024.2', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2024.2', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2024.2', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2024.2', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2024.2', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2024.2', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [common : Copying over /run subdirectories conf] **************************",
                                "Tuesday 24 February 2026  00:47:30 +0000 (0:00:02.241)       0:00:16.832 ****** ",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [common : Restart systemd-tmpfiles] ***************************************",
                                "Tuesday 24 February 2026  00:47:31 +0000 (0:00:00.783)       0:00:17.616 ****** ",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [common : Copying over config.json files for services] ********************",
                                "Tuesday 24 February 2026  00:47:32 +0000 (0:00:01.089)       0:00:18.705 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2024.2', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2024.2', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2024.2', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2024.2', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2024.2', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2024.2', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2024.2', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2024.2', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2024.2', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2024.2', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2024.2', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2024.2', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2024.2', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2024.2', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2024.2', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2024.2', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2024.2', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2024.2', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2024.2', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2024.2', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2024.2', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "",
                                "TASK [common : Find custom fluentd input config files] *************************",
                                "Tuesday 24 February 2026  00:47:41 +0000 (0:00:08.714)       0:00:27.419 ****** ",
                                "\u001b[1;35m[WARNING]: Skipped\u001b[0m",
                                "\u001b[1;35m'/opt/configuration/environments/kolla/files/overlays/fluentd/input' path due\u001b[0m",
                                "\u001b[1;35mto this access issue:\u001b[0m",
                                "\u001b[1;35m'/opt/configuration/environments/kolla/files/overlays/fluentd/input' is not a\u001b[0m",
                                "\u001b[1;35mdirectory\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager -> localhost]\u001b[0m",
                                "",
                                "TASK [common : Find custom fluentd filter config files] ************************",
                                "Tuesday 24 February 2026  00:47:43 +0000 (0:00:01.752)       0:00:29.172 ****** ",
                                "\u001b[1;35m[WARNING]: Skipped\u001b[0m",
                                "\u001b[1;35m'/opt/configuration/environments/kolla/files/overlays/fluentd/filter' path due\u001b[0m",
                                "\u001b[1;35mto this access issue:\u001b[0m",
                                "\u001b[1;35m'/opt/configuration/environments/kolla/files/overlays/fluentd/filter' is not a\u001b[0m",
                                "\u001b[1;35mdirectory\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager -> localhost]\u001b[0m",
                                "",
                                "TASK [common : Find custom fluentd format config files] ************************",
                                "Tuesday 24 February 2026  00:47:44 +0000 (0:00:01.199)       0:00:30.371 ****** ",
                                "\u001b[1;35m[WARNING]: Skipped\u001b[0m",
                                "\u001b[1;35m'/opt/configuration/environments/kolla/files/overlays/fluentd/format' path due\u001b[0m",
                                "\u001b[1;35mto this access issue:\u001b[0m",
                                "\u001b[1;35m'/opt/configuration/environments/kolla/files/overlays/fluentd/format' is not a\u001b[0m",
                                "\u001b[1;35mdirectory\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager -> localhost]\u001b[0m",
                                "",
                                "TASK [common : Find custom fluentd output config files] ************************",
                                "Tuesday 24 February 2026  00:47:45 +0000 (0:00:00.791)       0:00:31.163 ****** ",
                                "\u001b[1;35m[WARNING]: Skipped\u001b[0m",
                                "\u001b[1;35m'/opt/configuration/environments/kolla/files/overlays/fluentd/output' path due\u001b[0m",
                                "\u001b[1;35mto this access issue:\u001b[0m",
                                "\u001b[1;35m'/opt/configuration/environments/kolla/files/overlays/fluentd/output' is not a\u001b[0m",
                                "\u001b[1;35mdirectory\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager -> localhost]\u001b[0m",
                                "",
                                "TASK [common : Copying over fluentd.conf] **************************************",
                                "Tuesday 24 February 2026  00:47:45 +0000 (0:00:00.679)       0:00:31.842 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [common : Copying over cron logrotate config file] ************************",
                                "Tuesday 24 February 2026  00:47:50 +0000 (0:00:04.267)       0:00:36.110 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=/ansible/roles/common/templates/cron-logrotate-global.conf.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=/ansible/roles/common/templates/cron-logrotate-global.conf.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=/ansible/roles/common/templates/cron-logrotate-global.conf.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=/ansible/roles/common/templates/cron-logrotate-global.conf.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=/ansible/roles/common/templates/cron-logrotate-global.conf.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=/ansible/roles/common/templates/cron-logrotate-global.conf.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=/ansible/roles/common/templates/cron-logrotate-global.conf.j2)\u001b[0m",
                                "",
                                "TASK [common : Ensure RabbitMQ Erlang cookie exists] ***************************",
                                "Tuesday 24 February 2026  00:47:54 +0000 (0:00:04.599)       0:00:40.709 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [common : Ensuring config directories have correct owner and permission] ***",
                                "Tuesday 24 February 2026  00:47:57 +0000 (0:00:02.928)       0:00:43.638 ****** ",
                                "\u001b[0;32mok: [testbed-node-0] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2024.2', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2024.2', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;32mok: [testbed-manager] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2024.2', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2024.2', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2024.2', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2024.2', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2024.2', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2024.2', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2024.2', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2024.2', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2024.2', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2024.2', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2024.2', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2024.2', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2024.2', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2024.2', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2024.2', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2024.2', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2024.2', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2024.2', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2024.2', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "",
                                "TASK [common : Copy rabbitmq-env.conf to kolla toolbox] ************************",
                                "Tuesday 24 February 2026  00:48:01 +0000 (0:00:03.343)       0:00:46.981 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=/ansible/roles/common/templates/rabbitmq-env.conf.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=/ansible/roles/common/templates/rabbitmq-env.conf.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=/ansible/roles/common/templates/rabbitmq-env.conf.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=/ansible/roles/common/templates/rabbitmq-env.conf.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=/ansible/roles/common/templates/rabbitmq-env.conf.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=/ansible/roles/common/templates/rabbitmq-env.conf.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=/ansible/roles/common/templates/rabbitmq-env.conf.j2)\u001b[0m",
                                "",
                                "TASK [common : Copy rabbitmq erl_inetrc to kolla toolbox] **********************",
                                "Tuesday 24 February 2026  00:48:03 +0000 (0:00:02.665)       0:00:49.647 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=/ansible/roles/common/templates/erl_inetrc.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=/ansible/roles/common/templates/erl_inetrc.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=/ansible/roles/common/templates/erl_inetrc.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=/ansible/roles/common/templates/erl_inetrc.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=/ansible/roles/common/templates/erl_inetrc.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=/ansible/roles/common/templates/erl_inetrc.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=/ansible/roles/common/templates/erl_inetrc.j2)\u001b[0m",
                                "",
                                "TASK [common : Check common containers] ****************************************",
                                "Tuesday 24 February 2026  00:48:06 +0000 (0:00:02.582)       0:00:52.229 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2024.2', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2024.2', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2024.2', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2024.2', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2024.2', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2024.2', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'fluentd', 'value': {'container_name': 'fluentd', 'group': 'fluentd', 'enabled': True, 'image': 'registry.osism.tech/kolla/fluentd:2024.2', 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS'}, 'volumes': ['/etc/kolla/fluentd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'fluentd_data:/var/lib/fluentd/data/', '/var/log/journal:/var/log/journal:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2024.2', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2024.2', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2024.2', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2024.2', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2024.2', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2024.2', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'kolla-toolbox', 'value': {'container_name': 'kolla_toolbox', 'group': 'kolla-toolbox', 'enabled': True, 'image': 'registry.osism.tech/kolla/kolla-toolbox:2024.2', 'environment': {'ANSIBLE_NOCOLOR': '1', 'ANSIBLE_LIBRARY': '/usr/share/ansible', 'REQUESTS_CA_BUNDLE': '/etc/ssl/certs/ca-certificates.crt'}, 'privileged': True, 'volumes': ['/etc/kolla/kolla-toolbox/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run/:/run/:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2024.2', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2024.2', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2024.2', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2024.2', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2024.2', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2024.2', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'cron', 'value': {'container_name': 'cron', 'group': 'cron', 'enabled': True, 'image': 'registry.osism.tech/kolla/cron:2024.2', 'environment': {'KOLLA_LOGROTATE_SCHEDULE': 'daily'}, 'volumes': ['/etc/kolla/cron/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "",
                                "TASK [common : Creating log volume] ********************************************",
                                "Tuesday 24 February 2026  00:48:10 +0000 (0:00:04.623)       0:00:56.854 ****** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [common : Link kolla_logs volume to /var/log/kolla] ***********************",
                                "Tuesday 24 February 2026  00:48:13 +0000 (0:00:02.092)       0:00:58.946 ****** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [common : Flush handlers] *************************************************",
                                "Tuesday 24 February 2026  00:48:14 +0000 (0:00:01.316)       0:01:00.263 ****** ",
                                "",
                                "TASK [common : Flush handlers] *************************************************",
                                "Tuesday 24 February 2026  00:48:14 +0000 (0:00:00.064)       0:01:00.327 ****** ",
                                "",
                                "TASK [common : Flush handlers] *************************************************",
                                "Tuesday 24 February 2026  00:48:14 +0000 (0:00:00.061)       0:01:00.388 ****** ",
                                "",
                                "TASK [common : Flush handlers] *************************************************",
                                "Tuesday 24 February 2026  00:48:14 +0000 (0:00:00.229)       0:01:00.618 ****** ",
                                "",
                                "TASK [common : Flush handlers] *************************************************",
                                "Tuesday 24 February 2026  00:48:14 +0000 (0:00:00.075)       0:01:00.693 ****** ",
                                "",
                                "TASK [common : Flush handlers] *************************************************",
                                "Tuesday 24 February 2026  00:48:14 +0000 (0:00:00.066)       0:01:00.760 ****** ",
                                "",
                                "TASK [common : Flush handlers] *************************************************",
                                "Tuesday 24 February 2026  00:48:14 +0000 (0:00:00.059)       0:01:00.819 ****** ",
                                "",
                                "RUNNING HANDLER [common : Restart fluentd container] ***************************",
                                "Tuesday 24 February 2026  00:48:14 +0000 (0:00:00.081)       0:01:00.901 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [common : Restart kolla-toolbox container] *********************",
                                "Tuesday 24 February 2026  00:48:46 +0000 (0:00:31.962)       0:01:32.863 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "",
                                "RUNNING HANDLER [common : Initializing toolbox container using normal user] ****",
                                "Tuesday 24 February 2026  00:49:17 +0000 (0:00:30.979)       0:02:03.843 ****** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "",
                                "RUNNING HANDLER [common : Restart cron container] ******************************",
                                "Tuesday 24 February 2026  00:49:20 +0000 (0:00:02.533)       0:02:06.376 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;33mtestbed-manager\u001b[0m            : \u001b[0;32mok=22  \u001b[0m \u001b[0;33mchanged=14  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=4   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=18  \u001b[0m \u001b[0;33mchanged=14  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=4   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=18  \u001b[0m \u001b[0;33mchanged=14  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=4   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=18  \u001b[0m \u001b[0;33mchanged=14  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=4   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-3\u001b[0m             : \u001b[0;32mok=18  \u001b[0m \u001b[0;33mchanged=14  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=4   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-4\u001b[0m             : \u001b[0;32mok=18  \u001b[0m \u001b[0;33mchanged=14  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=4   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-5\u001b[0m             : \u001b[0;32mok=18  \u001b[0m \u001b[0;33mchanged=14  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=4   \u001b[0m rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Tuesday 24 February 2026  00:49:29 +0000 (0:00:09.227)       0:02:15.603 ****** ",
                                "=============================================================================== ",
                                "common : Restart fluentd container ------------------------------------- 31.96s",
                                "common : Restart kolla-toolbox container ------------------------------- 30.98s",
                                "common : Restart cron container ----------------------------------------- 9.23s",
                                "common : Copying over config.json files for services -------------------- 8.71s",
                                "service-cert-copy : common | Copying over extra CA certificates --------- 5.07s",
                                "common : Check common containers ---------------------------------------- 4.62s",
                                "common : Copying over cron logrotate config file ------------------------ 4.60s",
                                "common : Copying over fluentd.conf -------------------------------------- 4.27s",
                                "common : Ensuring config directories exist ------------------------------ 4.14s",
                                "common : Ensuring config directories have correct owner and permission --- 3.34s",
                                "common : Ensure RabbitMQ Erlang cookie exists --------------------------- 2.93s",
                                "common : Copy rabbitmq-env.conf to kolla toolbox ------------------------ 2.67s",
                                "common : Copy rabbitmq erl_inetrc to kolla toolbox ---------------------- 2.58s",
                                "common : Initializing toolbox container using normal user --------------- 2.53s",
                                "service-cert-copy : common | Copying over backend internal TLS key ------ 2.24s",
                                "service-cert-copy : common | Copying over backend internal TLS certificate --- 2.14s",
                                "common : Creating log volume -------------------------------------------- 2.09s",
                                "common : Find custom fluentd input config files ------------------------- 1.75s",
                                "common : include_tasks -------------------------------------------------- 1.56s",
                                "common : include_tasks -------------------------------------------------- 1.44s",
                                "\u001b[32m2026-02-24 00:49:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:49:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 404270fa-50db-4a78-8c44-21df48acba39 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:49:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 343b9f6a-0e5e-4af4-8877-c8c90685dcbb is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:49:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 0f568c96-7c6a-46f2-9e12-c74cfe06cd40 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:49:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 00:49:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ff4e45b7-3128-4ce1-95d9-909f33d92301 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:49:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f8694a2a-a9eb-4540-948b-ced7e58054f8 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:49:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:49:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 404270fa-50db-4a78-8c44-21df48acba39 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:49:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 343b9f6a-0e5e-4af4-8877-c8c90685dcbb is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:49:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 0f568c96-7c6a-46f2-9e12-c74cfe06cd40 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:49:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 00:49:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ff4e45b7-3128-4ce1-95d9-909f33d92301 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:49:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f8694a2a-a9eb-4540-948b-ced7e58054f8 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:49:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:49:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 404270fa-50db-4a78-8c44-21df48acba39 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:49:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 343b9f6a-0e5e-4af4-8877-c8c90685dcbb is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:49:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 0f568c96-7c6a-46f2-9e12-c74cfe06cd40 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:49:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 00:49:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ff4e45b7-3128-4ce1-95d9-909f33d92301 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:49:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f8694a2a-a9eb-4540-948b-ced7e58054f8 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:49:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:49:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 404270fa-50db-4a78-8c44-21df48acba39 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:49:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 343b9f6a-0e5e-4af4-8877-c8c90685dcbb is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:49:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 0f568c96-7c6a-46f2-9e12-c74cfe06cd40 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:49:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 00:49:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ff4e45b7-3128-4ce1-95d9-909f33d92301 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:49:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f8694a2a-a9eb-4540-948b-ced7e58054f8 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:49:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:49:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 404270fa-50db-4a78-8c44-21df48acba39 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:49:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 343b9f6a-0e5e-4af4-8877-c8c90685dcbb is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:49:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 0f568c96-7c6a-46f2-9e12-c74cfe06cd40 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:49:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 00:49:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ff4e45b7-3128-4ce1-95d9-909f33d92301 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:49:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f8694a2a-a9eb-4540-948b-ced7e58054f8 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:49:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:49:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 404270fa-50db-4a78-8c44-21df48acba39 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:49:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 343b9f6a-0e5e-4af4-8877-c8c90685dcbb is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:49:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 0f568c96-7c6a-46f2-9e12-c74cfe06cd40 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:49:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 00:49:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ff4e45b7-3128-4ce1-95d9-909f33d92301 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:49:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f8694a2a-a9eb-4540-948b-ced7e58054f8 is in state SUCCESS\u001b[0m",
                                "\u001b[32m2026-02-24 00:49:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:49:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 404270fa-50db-4a78-8c44-21df48acba39 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:49:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 343b9f6a-0e5e-4af4-8877-c8c90685dcbb is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:49:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 0f568c96-7c6a-46f2-9e12-c74cfe06cd40 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:49:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 00:49:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ff4e45b7-3128-4ce1-95d9-909f33d92301 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:49:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e7b4207c-f95e-49ed-815c-ab81afbfeb3f is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:49:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:49:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 404270fa-50db-4a78-8c44-21df48acba39 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:49:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 343b9f6a-0e5e-4af4-8877-c8c90685dcbb is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:49:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 0f568c96-7c6a-46f2-9e12-c74cfe06cd40 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:49:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 00:49:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ff4e45b7-3128-4ce1-95d9-909f33d92301 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:49:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e7b4207c-f95e-49ed-815c-ab81afbfeb3f is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:49:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:49:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 404270fa-50db-4a78-8c44-21df48acba39 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:49:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 343b9f6a-0e5e-4af4-8877-c8c90685dcbb is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:49:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 0f568c96-7c6a-46f2-9e12-c74cfe06cd40 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:49:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 00:50:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ff4e45b7-3128-4ce1-95d9-909f33d92301 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:50:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e7b4207c-f95e-49ed-815c-ab81afbfeb3f is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:50:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:50:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 404270fa-50db-4a78-8c44-21df48acba39 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:50:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 343b9f6a-0e5e-4af4-8877-c8c90685dcbb is in state SUCCESS\u001b[0m",
                                "",
                                "",
                                "PLAY [Group hosts based on configuration] **************************************",
                                "",
                                "TASK [Group hosts based on Kolla action] ***************************************",
                                "Tuesday 24 February 2026  00:49:36 +0000 (0:00:00.473)       0:00:00.473 ****** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [Group hosts based on enabled services] ***********************************",
                                "Tuesday 24 February 2026  00:49:37 +0000 (0:00:00.621)       0:00:01.094 ****** ",
                                "\u001b[0;32mok: [testbed-node-0] => (item=enable_memcached_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item=enable_memcached_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item=enable_memcached_True)\u001b[0m",
                                "",
                                "PLAY [Apply role memcached] ****************************************************",
                                "",
                                "TASK [memcached : include_tasks] ***********************************************",
                                "Tuesday 24 February 2026  00:49:37 +0000 (0:00:00.796)       0:00:01.891 ****** ",
                                "\u001b[0;36mincluded: /ansible/roles/memcached/tasks/deploy.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [memcached : Ensuring config directories exist] ***************************",
                                "Tuesday 24 February 2026  00:49:38 +0000 (0:00:00.501)       0:00:02.392 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=memcached)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=memcached)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=memcached)\u001b[0m",
                                "",
                                "TASK [memcached : Copying over config.json files for services] *****************",
                                "Tuesday 24 February 2026  00:49:39 +0000 (0:00:00.877)       0:00:03.270 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=memcached)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=memcached)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=memcached)\u001b[0m",
                                "",
                                "TASK [memcached : Check memcached container] ***********************************",
                                "Tuesday 24 February 2026  00:49:41 +0000 (0:00:02.108)       0:00:05.378 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "",
                                "RUNNING HANDLER [memcached : Restart memcached container] **********************",
                                "Tuesday 24 February 2026  00:49:43 +0000 (0:00:01.897)       0:00:07.276 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=7   \u001b[0m \u001b[0;33mchanged=4   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=7   \u001b[0m \u001b[0;33mchanged=4   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=7   \u001b[0m \u001b[0;33mchanged=4   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Tuesday 24 February 2026  00:49:50 +0000 (0:00:07.173)       0:00:14.449 ****** ",
                                "=============================================================================== ",
                                "memcached : Restart memcached container --------------------------------- 7.17s",
                                "memcached : Copying over config.json files for services ----------------- 2.11s",
                                "memcached : Check memcached container ----------------------------------- 1.90s",
                                "memcached : Ensuring config directories exist --------------------------- 0.88s",
                                "Group hosts based on enabled services ----------------------------------- 0.80s",
                                "Group hosts based on Kolla action --------------------------------------- 0.62s",
                                "memcached : include_tasks ----------------------------------------------- 0.50s",
                                "",
                                "",
                                "PLAY [Group hosts based on configuration] **************************************",
                                "",
                                "TASK [Group hosts based on Kolla action] ***************************************",
                                "Tuesday 24 February 2026  00:49:35 +0000 (0:00:00.429)       0:00:00.429 ****** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [Group hosts based on enabled services] ***********************************",
                                "Tuesday 24 February 2026  00:49:35 +0000 (0:00:00.265)       0:00:00.695 ****** ",
                                "\u001b[0;32mok: [testbed-node-0] => (item=enable_redis_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item=enable_redis_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item=enable_redis_True)\u001b[0m",
                                "",
                                "PLAY [Apply role redis] ********************************************************",
                                "",
                                "TASK [redis : include_tasks] ***************************************************",
                                "Tuesday 24 February 2026  00:49:36 +0000 (0:00:00.799)       0:00:01.495 ****** ",
                                "\u001b[0;36mincluded: /ansible/roles/redis/tasks/deploy.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [redis : Ensuring config directories exist] *******************************",
                                "Tuesday 24 February 2026  00:49:37 +0000 (0:00:00.547)       0:00:02.042 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'redis', 'value': {'container_name': 'redis', 'group': 'redis', 'enabled': True, 'image': 'registry.osism.tech/kolla/redis:2024.2', 'volumes': ['/etc/kolla/redis/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'redis:/var/lib/redis/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-server 6379'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'redis', 'value': {'container_name': 'redis', 'group': 'redis', 'enabled': True, 'image': 'registry.osism.tech/kolla/redis:2024.2', 'volumes': ['/etc/kolla/redis/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'redis:/var/lib/redis/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-server 6379'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'redis', 'value': {'container_name': 'redis', 'group': 'redis', 'enabled': True, 'image': 'registry.osism.tech/kolla/redis:2024.2', 'volumes': ['/etc/kolla/redis/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'redis:/var/lib/redis/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-server 6379'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'redis-sentinel', 'value': {'container_name': 'redis_sentinel', 'group': 'redis', 'environment': {'REDIS_CONF': '/etc/redis/redis.conf', 'REDIS_GEN_CONF': '/etc/redis/redis-regenerated-by-config-rewrite.conf'}, 'enabled': True, 'image': 'registry.osism.tech/kolla/redis-sentinel:2024.2', 'volumes': ['/etc/kolla/redis-sentinel/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-sentinel 26379'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'redis-sentinel', 'value': {'container_name': 'redis_sentinel', 'group': 'redis', 'environment': {'REDIS_CONF': '/etc/redis/redis.conf', 'REDIS_GEN_CONF': '/etc/redis/redis-regenerated-by-config-rewrite.conf'}, 'enabled': True, 'image': 'registry.osism.tech/kolla/redis-sentinel:2024.2', 'volumes': ['/etc/kolla/redis-sentinel/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-sentinel 26379'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'redis-sentinel', 'value': {'container_name': 'redis_sentinel', 'group': 'redis', 'environment': {'REDIS_CONF': '/etc/redis/redis.conf', 'REDIS_GEN_CONF': '/etc/redis/redis-regenerated-by-config-rewrite.conf'}, 'enabled': True, 'image': 'registry.osism.tech/kolla/redis-sentinel:2024.2', 'volumes': ['/etc/kolla/redis-sentinel/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-sentinel 26379'], 'timeout': '30'}}})\u001b[0m",
                                "",
                                "TASK [redis : Copying over default config.json files] **************************",
                                "Tuesday 24 February 2026  00:49:38 +0000 (0:00:01.585)       0:00:03.628 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'redis', 'value': {'container_name': 'redis', 'group': 'redis', 'enabled': True, 'image': 'registry.osism.tech/kolla/redis:2024.2', 'volumes': ['/etc/kolla/redis/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'redis:/var/lib/redis/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-server 6379'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'redis', 'value': {'container_name': 'redis', 'group': 'redis', 'enabled': True, 'image': 'registry.osism.tech/kolla/redis:2024.2', 'volumes': ['/etc/kolla/redis/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'redis:/var/lib/redis/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-server 6379'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'redis', 'value': {'container_name': 'redis', 'group': 'redis', 'enabled': True, 'image': 'registry.osism.tech/kolla/redis:2024.2', 'volumes': ['/etc/kolla/redis/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'redis:/var/lib/redis/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-server 6379'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'redis-sentinel', 'value': {'container_name': 'redis_sentinel', 'group': 'redis', 'environment': {'REDIS_CONF': '/etc/redis/redis.conf', 'REDIS_GEN_CONF': '/etc/redis/redis-regenerated-by-config-rewrite.conf'}, 'enabled': True, 'image': 'registry.osism.tech/kolla/redis-sentinel:2024.2', 'volumes': ['/etc/kolla/redis-sentinel/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-sentinel 26379'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'redis-sentinel', 'value': {'container_name': 'redis_sentinel', 'group': 'redis', 'environment': {'REDIS_CONF': '/etc/redis/redis.conf', 'REDIS_GEN_CONF': '/etc/redis/redis-regenerated-by-config-rewrite.conf'}, 'enabled': True, 'image': 'registry.osism.tech/kolla/redis-sentinel:2024.2', 'volumes': ['/etc/kolla/redis-sentinel/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-sentinel 26379'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'redis-sentinel', 'value': {'container_name': 'redis_sentinel', 'group': 'redis', 'environment': {'REDIS_CONF': '/etc/redis/redis.conf', 'REDIS_GEN_CONF': '/etc/redis/redis-regenerated-by-config-rewrite.conf'}, 'enabled': True, 'image': 'registry.osism.tech/kolla/redis-sentinel:2024.2', 'volumes': ['/etc/kolla/redis-sentinel/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-sentinel 26379'], 'timeout': '30'}}})\u001b[0m",
                                "",
                                "TASK [redis : Copying over redis config files] *********************************",
                                "Tuesday 24 February 2026  00:49:41 +0000 (0:00:03.088)       0:00:06.717 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'redis', 'value': {'container_name': 'redis', 'group': 'redis', 'enabled': True, 'image': 'registry.osism.tech/kolla/redis:2024.2', 'volumes': ['/etc/kolla/redis/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'redis:/var/lib/redis/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-server 6379'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'redis', 'value': {'container_name': 'redis', 'group': 'redis', 'enabled': True, 'image': 'registry.osism.tech/kolla/redis:2024.2', 'volumes': ['/etc/kolla/redis/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'redis:/var/lib/redis/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-server 6379'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'redis', 'value': {'container_name': 'redis', 'group': 'redis', 'enabled': True, 'image': 'registry.osism.tech/kolla/redis:2024.2', 'volumes': ['/etc/kolla/redis/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'redis:/var/lib/redis/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-server 6379'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'redis-sentinel', 'value': {'container_name': 'redis_sentinel', 'group': 'redis', 'environment': {'REDIS_CONF': '/etc/redis/redis.conf', 'REDIS_GEN_CONF': '/etc/redis/redis-regenerated-by-config-rewrite.conf'}, 'enabled': True, 'image': 'registry.osism.tech/kolla/redis-sentinel:2024.2', 'volumes': ['/etc/kolla/redis-sentinel/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-sentinel 26379'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'redis-sentinel', 'value': {'container_name': 'redis_sentinel', 'group': 'redis', 'environment': {'REDIS_CONF': '/etc/redis/redis.conf', 'REDIS_GEN_CONF': '/etc/redis/redis-regenerated-by-config-rewrite.conf'}, 'enabled': True, 'image': 'registry.osism.tech/kolla/redis-sentinel:2024.2', 'volumes': ['/etc/kolla/redis-sentinel/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-sentinel 26379'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'redis-sentinel', 'value': {'container_name': 'redis_sentinel', 'group': 'redis', 'environment': {'REDIS_CONF': '/etc/redis/redis.conf', 'REDIS_GEN_CONF': '/etc/redis/redis-regenerated-by-config-rewrite.conf'}, 'enabled': True, 'image': 'registry.osism.tech/kolla/redis-sentinel:2024.2', 'volumes': ['/etc/kolla/redis-sentinel/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-sentinel 26379'], 'timeout': '30'}}})\u001b[0m",
                                "",
                                "TASK [redis : Check redis containers] ******************************************",
                                "Tuesday 24 February 2026  00:49:44 +0000 (0:00:02.893)       0:00:09.610 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'redis', 'value': {'container_name': 'redis', 'group': 'redis', 'enabled': True, 'image': 'registry.osism.tech/kolla/redis:2024.2', 'volumes': ['/etc/kolla/redis/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'redis:/var/lib/redis/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-server 6379'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'redis', 'value': {'container_name': 'redis', 'group': 'redis', 'enabled': True, 'image': 'registry.osism.tech/kolla/redis:2024.2', 'volumes': ['/etc/kolla/redis/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'redis:/var/lib/redis/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-server 6379'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'redis', 'value': {'container_name': 'redis', 'group': 'redis', 'enabled': True, 'image': 'registry.osism.tech/kolla/redis:2024.2', 'volumes': ['/etc/kolla/redis/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'redis:/var/lib/redis/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-server 6379'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'redis-sentinel', 'value': {'container_name': 'redis_sentinel', 'group': 'redis', 'environment': {'REDIS_CONF': '/etc/redis/redis.conf', 'REDIS_GEN_CONF': '/etc/redis/redis-regenerated-by-config-rewrite.conf'}, 'enabled': True, 'image': 'registry.osism.tech/kolla/redis-sentinel:2024.2', 'volumes': ['/etc/kolla/redis-sentinel/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-sentinel 26379'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'redis-sentinel', 'value': {'container_name': 'redis_sentinel', 'group': 'redis', 'environment': {'REDIS_CONF': '/etc/redis/redis.conf', 'REDIS_GEN_CONF': '/etc/redis/redis-regenerated-by-config-rewrite.conf'}, 'enabled': True, 'image': 'registry.osism.tech/kolla/redis-sentinel:2024.2', 'volumes': ['/etc/kolla/redis-sentinel/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-sentinel 26379'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'redis-sentinel', 'value': {'container_name': 'redis_sentinel', 'group': 'redis', 'environment': {'REDIS_CONF': '/etc/redis/redis.conf', 'REDIS_GEN_CONF': '/etc/redis/redis-regenerated-by-config-rewrite.conf'}, 'enabled': True, 'image': 'registry.osism.tech/kolla/redis-sentinel:2024.2', 'volumes': ['/etc/kolla/redis-sentinel/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen redis-sentinel 26379'], 'timeout': '30'}}})\u001b[0m",
                                "",
                                "TASK [redis : Flush handlers] **************************************************",
                                "Tuesday 24 February 2026  00:49:46 +0000 (0:00:01.754)       0:00:11.365 ****** ",
                                "",
                                "TASK [redis : Flush handlers] **************************************************",
                                "Tuesday 24 February 2026  00:49:46 +0000 (0:00:00.143)       0:00:11.508 ****** ",
                                "",
                                "TASK [redis : Flush handlers] **************************************************",
                                "Tuesday 24 February 2026  00:49:46 +0000 (0:00:00.113)       0:00:11.622 ****** ",
                                "",
                                "RUNNING HANDLER [redis : Restart redis container] ******************************",
                                "Tuesday 24 February 2026  00:49:46 +0000 (0:00:00.077)       0:00:11.699 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "",
                                "RUNNING HANDLER [redis : Restart redis-sentinel container] *********************",
                                "Tuesday 24 February 2026  00:49:55 +0000 (0:00:09.001)       0:00:20.701 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=9   \u001b[0m \u001b[0;33mchanged=6   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=9   \u001b[0m \u001b[0;33mchanged=6   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=9   \u001b[0m \u001b[0;33mchanged=6   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Tuesday 24 February 2026  00:49:58 +0000 (0:00:03.207)       0:00:23.908 ****** ",
                                "=============================================================================== ",
                                "redis : Restart redis container ----------------------------------------- 9.00s",
                                "redis : Restart redis-sentinel container -------------------------------- 3.21s",
                                "redis : Copying over default config.json files -------------------------- 3.09s",
                                "redis : Copying over redis config files --------------------------------- 2.89s",
                                "redis : Check redis containers ------------------------------------------ 1.75s",
                                "redis : Ensuring config directories exist ------------------------------- 1.59s",
                                "Group hosts based on enabled services ----------------------------------- 0.80s",
                                "redis : include_tasks --------------------------------------------------- 0.55s",
                                "redis : Flush handlers -------------------------------------------------- 0.33s",
                                "Group hosts based on Kolla action --------------------------------------- 0.27s",
                                "\u001b[32m2026-02-24 00:50:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 0f568c96-7c6a-46f2-9e12-c74cfe06cd40 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:50:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 00:50:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ff4e45b7-3128-4ce1-95d9-909f33d92301 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:50:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e7b4207c-f95e-49ed-815c-ab81afbfeb3f is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:50:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:50:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 404270fa-50db-4a78-8c44-21df48acba39 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:50:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 0f568c96-7c6a-46f2-9e12-c74cfe06cd40 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:50:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 00:50:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ff4e45b7-3128-4ce1-95d9-909f33d92301 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:50:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e7b4207c-f95e-49ed-815c-ab81afbfeb3f is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:50:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:50:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 404270fa-50db-4a78-8c44-21df48acba39 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:50:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 0f568c96-7c6a-46f2-9e12-c74cfe06cd40 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:50:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 00:50:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ff4e45b7-3128-4ce1-95d9-909f33d92301 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:50:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e7b4207c-f95e-49ed-815c-ab81afbfeb3f is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:50:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:50:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 404270fa-50db-4a78-8c44-21df48acba39 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:50:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 0f568c96-7c6a-46f2-9e12-c74cfe06cd40 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:50:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 00:50:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ff4e45b7-3128-4ce1-95d9-909f33d92301 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:50:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e7b4207c-f95e-49ed-815c-ab81afbfeb3f is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:50:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:50:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 404270fa-50db-4a78-8c44-21df48acba39 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:50:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 0f568c96-7c6a-46f2-9e12-c74cfe06cd40 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:50:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 00:50:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ff4e45b7-3128-4ce1-95d9-909f33d92301 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:50:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e7b4207c-f95e-49ed-815c-ab81afbfeb3f is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:50:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:50:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 404270fa-50db-4a78-8c44-21df48acba39 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:50:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 0f568c96-7c6a-46f2-9e12-c74cfe06cd40 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:50:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 00:50:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ff4e45b7-3128-4ce1-95d9-909f33d92301 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:50:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e7b4207c-f95e-49ed-815c-ab81afbfeb3f is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:50:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:50:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 404270fa-50db-4a78-8c44-21df48acba39 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:50:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 0f568c96-7c6a-46f2-9e12-c74cfe06cd40 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:50:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 00:50:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ff4e45b7-3128-4ce1-95d9-909f33d92301 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:50:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e7b4207c-f95e-49ed-815c-ab81afbfeb3f is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:50:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:50:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 404270fa-50db-4a78-8c44-21df48acba39 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:50:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 0f568c96-7c6a-46f2-9e12-c74cfe06cd40 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:50:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 00:50:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ff4e45b7-3128-4ce1-95d9-909f33d92301 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:50:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e7b4207c-f95e-49ed-815c-ab81afbfeb3f is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:50:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:50:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 404270fa-50db-4a78-8c44-21df48acba39 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:50:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 0f568c96-7c6a-46f2-9e12-c74cfe06cd40 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:50:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 00:50:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ff4e45b7-3128-4ce1-95d9-909f33d92301 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:50:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e7b4207c-f95e-49ed-815c-ab81afbfeb3f is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:50:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:50:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 404270fa-50db-4a78-8c44-21df48acba39 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:50:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 0f568c96-7c6a-46f2-9e12-c74cfe06cd40 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:50:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 00:50:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ff4e45b7-3128-4ce1-95d9-909f33d92301 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:50:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e7b4207c-f95e-49ed-815c-ab81afbfeb3f is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:50:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:50:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 404270fa-50db-4a78-8c44-21df48acba39 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:50:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 0f568c96-7c6a-46f2-9e12-c74cfe06cd40 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:50:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 00:50:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ff4e45b7-3128-4ce1-95d9-909f33d92301 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:50:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e7b4207c-f95e-49ed-815c-ab81afbfeb3f is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:50:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:50:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 404270fa-50db-4a78-8c44-21df48acba39 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:50:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 0f568c96-7c6a-46f2-9e12-c74cfe06cd40 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:50:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 00:50:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ff4e45b7-3128-4ce1-95d9-909f33d92301 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:50:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e7b4207c-f95e-49ed-815c-ab81afbfeb3f is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:50:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:50:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 404270fa-50db-4a78-8c44-21df48acba39 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:50:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 0f568c96-7c6a-46f2-9e12-c74cfe06cd40 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:50:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 00:50:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ff4e45b7-3128-4ce1-95d9-909f33d92301 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:50:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e7b4207c-f95e-49ed-815c-ab81afbfeb3f is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:50:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ba7d4726-e54a-4b5b-b060-92175ab177a2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:50:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:50:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 404270fa-50db-4a78-8c44-21df48acba39 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:50:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 0f568c96-7c6a-46f2-9e12-c74cfe06cd40 is in state SUCCESS\u001b[0m",
                                "",
                                "",
                                "PLAY [Group hosts based on configuration] **************************************",
                                "",
                                "TASK [Group hosts based on Kolla action] ***************************************",
                                "Tuesday 24 February 2026  00:49:35 +0000 (0:00:00.236)       0:00:00.236 ****** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Group hosts based on enabled services] ***********************************",
                                "Tuesday 24 February 2026  00:49:36 +0000 (0:00:00.798)       0:00:01.034 ****** ",
                                "\u001b[0;32mok: [testbed-node-0] => (item=enable_openvswitch_True_enable_ovs_dpdk_False)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item=enable_openvswitch_True_enable_ovs_dpdk_False)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item=enable_openvswitch_True_enable_ovs_dpdk_False)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item=enable_openvswitch_True_enable_ovs_dpdk_False)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item=enable_openvswitch_True_enable_ovs_dpdk_False)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item=enable_openvswitch_True_enable_ovs_dpdk_False)\u001b[0m",
                                "",
                                "PLAY [Apply role openvswitch] **************************************************",
                                "",
                                "TASK [openvswitch : include_tasks] *********************************************",
                                "Tuesday 24 February 2026  00:49:37 +0000 (0:00:01.031)       0:00:02.066 ****** ",
                                "\u001b[0;36mincluded: /ansible/roles/openvswitch/tasks/deploy.yml for testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m",
                                "",
                                "TASK [module-load : Load modules] **********************************************",
                                "Tuesday 24 February 2026  00:49:38 +0000 (0:00:01.459)       0:00:03.526 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=openvswitch)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=openvswitch)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=openvswitch)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=openvswitch)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=openvswitch)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=openvswitch)\u001b[0m",
                                "",
                                "TASK [module-load : Persist modules via modules-load.d] ************************",
                                "Tuesday 24 February 2026  00:49:40 +0000 (0:00:01.801)       0:00:05.327 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=openvswitch)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=openvswitch)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=openvswitch)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=openvswitch)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=openvswitch)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=openvswitch)\u001b[0m",
                                "",
                                "TASK [module-load : Drop module persistence] ***********************************",
                                "Tuesday 24 February 2026  00:49:42 +0000 (0:00:01.659)       0:00:06.987 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=openvswitch) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=openvswitch) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=openvswitch) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=openvswitch) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=openvswitch) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=openvswitch) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [openvswitch : Create /run/openvswitch directory on host] *****************",
                                "Tuesday 24 February 2026  00:49:43 +0000 (0:00:01.095)       0:00:08.082 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [openvswitch : Ensuring config directories exist] *************************",
                                "Tuesday 24 February 2026  00:49:44 +0000 (0:00:00.912)       0:00:08.994 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'openvswitch-db-server', 'value': {'container_name': 'openvswitch_db', 'image': 'registry.osism.tech/kolla/openvswitch-db-server:2024.2', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'volumes': ['/etc/kolla/openvswitch-db-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', 'openvswitch_db:/var/lib/openvswitch/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovsdb-client list-dbs'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'openvswitch-db-server', 'value': {'container_name': 'openvswitch_db', 'image': 'registry.osism.tech/kolla/openvswitch-db-server:2024.2', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'volumes': ['/etc/kolla/openvswitch-db-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', 'openvswitch_db:/var/lib/openvswitch/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovsdb-client list-dbs'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'openvswitch-db-server', 'value': {'container_name': 'openvswitch_db', 'image': 'registry.osism.tech/kolla/openvswitch-db-server:2024.2', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'volumes': ['/etc/kolla/openvswitch-db-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', 'openvswitch_db:/var/lib/openvswitch/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovsdb-client list-dbs'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'openvswitch-db-server', 'value': {'container_name': 'openvswitch_db', 'image': 'registry.osism.tech/kolla/openvswitch-db-server:2024.2', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'volumes': ['/etc/kolla/openvswitch-db-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', 'openvswitch_db:/var/lib/openvswitch/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovsdb-client list-dbs'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'openvswitch-vswitchd', 'value': {'container_name': 'openvswitch_vswitchd', 'image': 'registry.osism.tech/kolla/openvswitch-vswitchd:2024.2', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'privileged': True, 'volumes': ['/etc/kolla/openvswitch-vswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovs-appctl version'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'openvswitch-db-server', 'value': {'container_name': 'openvswitch_db', 'image': 'registry.osism.tech/kolla/openvswitch-db-server:2024.2', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'volumes': ['/etc/kolla/openvswitch-db-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', 'openvswitch_db:/var/lib/openvswitch/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovsdb-client list-dbs'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'openvswitch-vswitchd', 'value': {'container_name': 'openvswitch_vswitchd', 'image': 'registry.osism.tech/kolla/openvswitch-vswitchd:2024.2', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'privileged': True, 'volumes': ['/etc/kolla/openvswitch-vswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovs-appctl version'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'openvswitch-vswitchd', 'value': {'container_name': 'openvswitch_vswitchd', 'image': 'registry.osism.tech/kolla/openvswitch-vswitchd:2024.2', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'privileged': True, 'volumes': ['/etc/kolla/openvswitch-vswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovs-appctl version'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'openvswitch-db-server', 'value': {'container_name': 'openvswitch_db', 'image': 'registry.osism.tech/kolla/openvswitch-db-server:2024.2', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'volumes': ['/etc/kolla/openvswitch-db-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', 'openvswitch_db:/var/lib/openvswitch/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovsdb-client list-dbs'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'openvswitch-vswitchd', 'value': {'container_name': 'openvswitch_vswitchd', 'image': 'registry.osism.tech/kolla/openvswitch-vswitchd:2024.2', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'privileged': True, 'volumes': ['/etc/kolla/openvswitch-vswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovs-appctl version'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'openvswitch-vswitchd', 'value': {'container_name': 'openvswitch_vswitchd', 'image': 'registry.osism.tech/kolla/openvswitch-vswitchd:2024.2', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'privileged': True, 'volumes': ['/etc/kolla/openvswitch-vswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovs-appctl version'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'openvswitch-vswitchd', 'value': {'container_name': 'openvswitch_vswitchd', 'image': 'registry.osism.tech/kolla/openvswitch-vswitchd:2024.2', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'privileged': True, 'volumes': ['/etc/kolla/openvswitch-vswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovs-appctl version'], 'timeout': '30'}}})\u001b[0m",
                                "",
                                "TASK [openvswitch : Copying over config.json files for services] ***************",
                                "Tuesday 24 February 2026  00:49:45 +0000 (0:00:01.723)       0:00:10.717 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'openvswitch-db-server', 'value': {'container_name': 'openvswitch_db', 'image': 'registry.osism.tech/kolla/openvswitch-db-server:2024.2', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'volumes': ['/etc/kolla/openvswitch-db-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', 'openvswitch_db:/var/lib/openvswitch/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovsdb-client list-dbs'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'openvswitch-db-server', 'value': {'container_name': 'openvswitch_db', 'image': 'registry.osism.tech/kolla/openvswitch-db-server:2024.2', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'volumes': ['/etc/kolla/openvswitch-db-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', 'openvswitch_db:/var/lib/openvswitch/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovsdb-client list-dbs'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'openvswitch-db-server', 'value': {'container_name': 'openvswitch_db', 'image': 'registry.osism.tech/kolla/openvswitch-db-server:2024.2', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'volumes': ['/etc/kolla/openvswitch-db-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', 'openvswitch_db:/var/lib/openvswitch/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovsdb-client list-dbs'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'openvswitch-db-server', 'value': {'container_name': 'openvswitch_db', 'image': 'registry.osism.tech/kolla/openvswitch-db-server:2024.2', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'volumes': ['/etc/kolla/openvswitch-db-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', 'openvswitch_db:/var/lib/openvswitch/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovsdb-client list-dbs'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'openvswitch-db-server', 'value': {'container_name': 'openvswitch_db', 'image': 'registry.osism.tech/kolla/openvswitch-db-server:2024.2', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'volumes': ['/etc/kolla/openvswitch-db-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', 'openvswitch_db:/var/lib/openvswitch/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovsdb-client list-dbs'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'openvswitch-db-server', 'value': {'container_name': 'openvswitch_db', 'image': 'registry.osism.tech/kolla/openvswitch-db-server:2024.2', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'volumes': ['/etc/kolla/openvswitch-db-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', 'openvswitch_db:/var/lib/openvswitch/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovsdb-client list-dbs'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'openvswitch-vswitchd', 'value': {'container_name': 'openvswitch_vswitchd', 'image': 'registry.osism.tech/kolla/openvswitch-vswitchd:2024.2', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'privileged': True, 'volumes': ['/etc/kolla/openvswitch-vswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovs-appctl version'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'openvswitch-vswitchd', 'value': {'container_name': 'openvswitch_vswitchd', 'image': 'registry.osism.tech/kolla/openvswitch-vswitchd:2024.2', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'privileged': True, 'volumes': ['/etc/kolla/openvswitch-vswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovs-appctl version'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'openvswitch-vswitchd', 'value': {'container_name': 'openvswitch_vswitchd', 'image': 'registry.osism.tech/kolla/openvswitch-vswitchd:2024.2', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'privileged': True, 'volumes': ['/etc/kolla/openvswitch-vswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovs-appctl version'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'openvswitch-vswitchd', 'value': {'container_name': 'openvswitch_vswitchd', 'image': 'registry.osism.tech/kolla/openvswitch-vswitchd:2024.2', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'privileged': True, 'volumes': ['/etc/kolla/openvswitch-vswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovs-appctl version'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'openvswitch-vswitchd', 'value': {'container_name': 'openvswitch_vswitchd', 'image': 'registry.osism.tech/kolla/openvswitch-vswitchd:2024.2', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'privileged': True, 'volumes': ['/etc/kolla/openvswitch-vswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovs-appctl version'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'openvswitch-vswitchd', 'value': {'container_name': 'openvswitch_vswitchd', 'image': 'registry.osism.tech/kolla/openvswitch-vswitchd:2024.2', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'privileged': True, 'volumes': ['/etc/kolla/openvswitch-vswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovs-appctl version'], 'timeout': '30'}}})\u001b[0m",
                                "",
                                "TASK [openvswitch : Copying over ovs-vsctl wrapper] ****************************",
                                "Tuesday 24 February 2026  00:49:48 +0000 (0:00:02.791)       0:00:13.509 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [openvswitch : Check openvswitch containers] ******************************",
                                "Tuesday 24 February 2026  00:49:49 +0000 (0:00:00.825)       0:00:14.335 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'openvswitch-db-server', 'value': {'container_name': 'openvswitch_db', 'image': 'registry.osism.tech/kolla/openvswitch-db-server:2024.2', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'volumes': ['/etc/kolla/openvswitch-db-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', 'openvswitch_db:/var/lib/openvswitch/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovsdb-client list-dbs'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'openvswitch-db-server', 'value': {'container_name': 'openvswitch_db', 'image': 'registry.osism.tech/kolla/openvswitch-db-server:2024.2', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'volumes': ['/etc/kolla/openvswitch-db-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', 'openvswitch_db:/var/lib/openvswitch/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovsdb-client list-dbs'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'openvswitch-db-server', 'value': {'container_name': 'openvswitch_db', 'image': 'registry.osism.tech/kolla/openvswitch-db-server:2024.2', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'volumes': ['/etc/kolla/openvswitch-db-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', 'openvswitch_db:/var/lib/openvswitch/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovsdb-client list-dbs'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'openvswitch-db-server', 'value': {'container_name': 'openvswitch_db', 'image': 'registry.osism.tech/kolla/openvswitch-db-server:2024.2', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'volumes': ['/etc/kolla/openvswitch-db-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', 'openvswitch_db:/var/lib/openvswitch/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovsdb-client list-dbs'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'openvswitch-db-server', 'value': {'container_name': 'openvswitch_db', 'image': 'registry.osism.tech/kolla/openvswitch-db-server:2024.2', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'volumes': ['/etc/kolla/openvswitch-db-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', 'openvswitch_db:/var/lib/openvswitch/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovsdb-client list-dbs'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'openvswitch-db-server', 'value': {'container_name': 'openvswitch_db', 'image': 'registry.osism.tech/kolla/openvswitch-db-server:2024.2', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'volumes': ['/etc/kolla/openvswitch-db-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', 'openvswitch_db:/var/lib/openvswitch/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovsdb-client list-dbs'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'openvswitch-vswitchd', 'value': {'container_name': 'openvswitch_vswitchd', 'image': 'registry.osism.tech/kolla/openvswitch-vswitchd:2024.2', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'privileged': True, 'volumes': ['/etc/kolla/openvswitch-vswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovs-appctl version'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'openvswitch-vswitchd', 'value': {'container_name': 'openvswitch_vswitchd', 'image': 'registry.osism.tech/kolla/openvswitch-vswitchd:2024.2', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'privileged': True, 'volumes': ['/etc/kolla/openvswitch-vswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovs-appctl version'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'openvswitch-vswitchd', 'value': {'container_name': 'openvswitch_vswitchd', 'image': 'registry.osism.tech/kolla/openvswitch-vswitchd:2024.2', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'privileged': True, 'volumes': ['/etc/kolla/openvswitch-vswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovs-appctl version'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'openvswitch-vswitchd', 'value': {'container_name': 'openvswitch_vswitchd', 'image': 'registry.osism.tech/kolla/openvswitch-vswitchd:2024.2', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'privileged': True, 'volumes': ['/etc/kolla/openvswitch-vswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovs-appctl version'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'openvswitch-vswitchd', 'value': {'container_name': 'openvswitch_vswitchd', 'image': 'registry.osism.tech/kolla/openvswitch-vswitchd:2024.2', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'privileged': True, 'volumes': ['/etc/kolla/openvswitch-vswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovs-appctl version'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'openvswitch-vswitchd', 'value': {'container_name': 'openvswitch_vswitchd', 'image': 'registry.osism.tech/kolla/openvswitch-vswitchd:2024.2', 'enabled': True, 'group': 'openvswitch', 'host_in_groups': True, 'privileged': True, 'volumes': ['/etc/kolla/openvswitch-vswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'ovs-appctl version'], 'timeout': '30'}}})\u001b[0m",
                                "",
                                "TASK [openvswitch : Flush Handlers] ********************************************",
                                "Tuesday 24 February 2026  00:49:51 +0000 (0:00:02.158)       0:00:16.494 ****** ",
                                "",
                                "TASK [openvswitch : Flush Handlers] ********************************************",
                                "Tuesday 24 February 2026  00:49:52 +0000 (0:00:00.345)       0:00:16.840 ****** ",
                                "",
                                "TASK [openvswitch : Flush Handlers] ********************************************",
                                "Tuesday 24 February 2026  00:49:52 +0000 (0:00:00.147)       0:00:16.987 ****** ",
                                "",
                                "TASK [openvswitch : Flush Handlers] ********************************************",
                                "Tuesday 24 February 2026  00:49:52 +0000 (0:00:00.153)       0:00:17.140 ****** ",
                                "",
                                "TASK [openvswitch : Flush Handlers] ********************************************",
                                "Tuesday 24 February 2026  00:49:52 +0000 (0:00:00.138)       0:00:17.279 ****** ",
                                "",
                                "TASK [openvswitch : Flush Handlers] ********************************************",
                                "Tuesday 24 February 2026  00:49:52 +0000 (0:00:00.229)       0:00:17.509 ****** ",
                                "",
                                "RUNNING HANDLER [openvswitch : Restart openvswitch-db-server container] ********",
                                "Tuesday 24 February 2026  00:49:52 +0000 (0:00:00.162)       0:00:17.671 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "",
                                "RUNNING HANDLER [openvswitch : Waiting for openvswitch_db service to be ready] ***",
                                "Tuesday 24 February 2026  00:50:02 +0000 (0:00:10.090)       0:00:27.762 ****** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "RUNNING HANDLER [openvswitch : Restart openvswitch-vswitchd container] *********",
                                "Tuesday 24 February 2026  00:50:05 +0000 (0:00:02.400)       0:00:30.162 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [openvswitch : Set system-id, hostname and hw-offload] ********************",
                                "Tuesday 24 February 2026  00:50:15 +0000 (0:00:10.030)       0:00:40.192 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'col': 'external_ids', 'name': 'system-id', 'value': 'testbed-node-2'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'col': 'external_ids', 'name': 'system-id', 'value': 'testbed-node-0'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'col': 'external_ids', 'name': 'system-id', 'value': 'testbed-node-4'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'col': 'external_ids', 'name': 'system-id', 'value': 'testbed-node-5'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'col': 'external_ids', 'name': 'system-id', 'value': 'testbed-node-3'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'col': 'external_ids', 'name': 'system-id', 'value': 'testbed-node-1'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'col': 'external_ids', 'name': 'hostname', 'value': 'testbed-node-4'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'col': 'external_ids', 'name': 'hostname', 'value': 'testbed-node-2'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'col': 'external_ids', 'name': 'hostname', 'value': 'testbed-node-0'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'col': 'external_ids', 'name': 'hostname', 'value': 'testbed-node-1'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'col': 'external_ids', 'name': 'hostname', 'value': 'testbed-node-3'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'col': 'external_ids', 'name': 'hostname', 'value': 'testbed-node-5'})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item={'col': 'other_config', 'name': 'hw-offload', 'value': True, 'state': 'absent'})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item={'col': 'other_config', 'name': 'hw-offload', 'value': True, 'state': 'absent'})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0] => (item={'col': 'other_config', 'name': 'hw-offload', 'value': True, 'state': 'absent'})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item={'col': 'other_config', 'name': 'hw-offload', 'value': True, 'state': 'absent'})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item={'col': 'other_config', 'name': 'hw-offload', 'value': True, 'state': 'absent'})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item={'col': 'other_config', 'name': 'hw-offload', 'value': True, 'state': 'absent'})\u001b[0m",
                                "",
                                "TASK [openvswitch : Ensuring OVS bridge is properly setup] *********************",
                                "Tuesday 24 February 2026  00:50:22 +0000 (0:00:06.907)       0:00:47.100 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=br-ex) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=br-ex) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=br-ex) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=br-ex)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=br-ex)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=br-ex)\u001b[0m",
                                "",
                                "TASK [openvswitch : Ensuring OVS ports are properly setup] *********************",
                                "Tuesday 24 February 2026  00:50:25 +0000 (0:00:03.009)       0:00:50.109 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=['br-ex', 'vxlan0']) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=['br-ex', 'vxlan0']) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=['br-ex', 'vxlan0']) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=['br-ex', 'vxlan0'])\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=['br-ex', 'vxlan0'])\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=['br-ex', 'vxlan0'])\u001b[0m",
                                "",
                                "RUNNING HANDLER [openvswitch : Restart openvswitch-vswitchd container] *********",
                                "Tuesday 24 February 2026  00:50:29 +0000 (0:00:03.997)       0:00:54.107 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=15  \u001b[0m \u001b[0;33mchanged=11  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=3   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=15  \u001b[0m \u001b[0;33mchanged=11  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=3   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=15  \u001b[0m \u001b[0;33mchanged=11  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=3   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-3\u001b[0m             : \u001b[0;32mok=13  \u001b[0m \u001b[0;33mchanged=9   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=5   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-4\u001b[0m             : \u001b[0;32mok=13  \u001b[0m \u001b[0;33mchanged=9   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=5   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-5\u001b[0m             : \u001b[0;32mok=13  \u001b[0m \u001b[0;33mchanged=9   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=5   \u001b[0m rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Tuesday 24 February 2026  00:50:38 +0000 (0:00:08.774)       0:01:02.881 ****** ",
                                "=============================================================================== ",
                                "openvswitch : Restart openvswitch-vswitchd container ------------------- 18.81s",
                                "openvswitch : Restart openvswitch-db-server container ------------------ 10.09s",
                                "openvswitch : Set system-id, hostname and hw-offload -------------------- 6.91s",
                                "openvswitch : Ensuring OVS ports are properly setup --------------------- 4.00s",
                                "openvswitch : Ensuring OVS bridge is properly setup --------------------- 3.01s",
                                "openvswitch : Copying over config.json files for services --------------- 2.79s",
                                "openvswitch : Waiting for openvswitch_db service to be ready ------------ 2.40s",
                                "openvswitch : Check openvswitch containers ------------------------------ 2.16s",
                                "module-load : Load modules ---------------------------------------------- 1.80s",
                                "openvswitch : Ensuring config directories exist ------------------------- 1.72s",
                                "module-load : Persist modules via modules-load.d ------------------------ 1.66s",
                                "openvswitch : include_tasks --------------------------------------------- 1.46s",
                                "openvswitch : Flush Handlers -------------------------------------------- 1.18s",
                                "module-load : Drop module persistence ----------------------------------- 1.10s",
                                "Group hosts based on enabled services ----------------------------------- 1.03s",
                                "openvswitch : Create /run/openvswitch directory on host ----------------- 0.91s",
                                "openvswitch : Copying over ovs-vsctl wrapper ---------------------------- 0.83s",
                                "Group hosts based on Kolla action --------------------------------------- 0.80s",
                                "\u001b[32m2026-02-24 00:50:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 00:50:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ff4e45b7-3128-4ce1-95d9-909f33d92301 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:50:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e7b4207c-f95e-49ed-815c-ab81afbfeb3f is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:50:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ba7d4726-e54a-4b5b-b060-92175ab177a2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:50:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:50:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 404270fa-50db-4a78-8c44-21df48acba39 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:50:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 00:50:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ff4e45b7-3128-4ce1-95d9-909f33d92301 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:50:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e7b4207c-f95e-49ed-815c-ab81afbfeb3f is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:50:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ba7d4726-e54a-4b5b-b060-92175ab177a2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:50:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:50:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 404270fa-50db-4a78-8c44-21df48acba39 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:50:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 00:50:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ff4e45b7-3128-4ce1-95d9-909f33d92301 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:50:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e7b4207c-f95e-49ed-815c-ab81afbfeb3f is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:50:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ba7d4726-e54a-4b5b-b060-92175ab177a2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:50:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:50:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 404270fa-50db-4a78-8c44-21df48acba39 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:50:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 00:50:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ff4e45b7-3128-4ce1-95d9-909f33d92301 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:50:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e7b4207c-f95e-49ed-815c-ab81afbfeb3f is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:50:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ba7d4726-e54a-4b5b-b060-92175ab177a2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:50:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:50:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 404270fa-50db-4a78-8c44-21df48acba39 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:50:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 00:50:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ff4e45b7-3128-4ce1-95d9-909f33d92301 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:50:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e7b4207c-f95e-49ed-815c-ab81afbfeb3f is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:50:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ba7d4726-e54a-4b5b-b060-92175ab177a2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:50:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:50:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 404270fa-50db-4a78-8c44-21df48acba39 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:50:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 00:50:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ff4e45b7-3128-4ce1-95d9-909f33d92301 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:50:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e7b4207c-f95e-49ed-815c-ab81afbfeb3f is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:50:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ba7d4726-e54a-4b5b-b060-92175ab177a2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:50:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:50:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 404270fa-50db-4a78-8c44-21df48acba39 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:50:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 00:51:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ff4e45b7-3128-4ce1-95d9-909f33d92301 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:51:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e7b4207c-f95e-49ed-815c-ab81afbfeb3f is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:51:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ba7d4726-e54a-4b5b-b060-92175ab177a2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:51:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:51:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 404270fa-50db-4a78-8c44-21df48acba39 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:51:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 00:51:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ff4e45b7-3128-4ce1-95d9-909f33d92301 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:51:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e7b4207c-f95e-49ed-815c-ab81afbfeb3f is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:51:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ba7d4726-e54a-4b5b-b060-92175ab177a2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:51:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:51:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 404270fa-50db-4a78-8c44-21df48acba39 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:51:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 00:51:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ff4e45b7-3128-4ce1-95d9-909f33d92301 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:51:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e7b4207c-f95e-49ed-815c-ab81afbfeb3f is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:51:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ba7d4726-e54a-4b5b-b060-92175ab177a2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:51:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:51:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 404270fa-50db-4a78-8c44-21df48acba39 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:51:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 00:51:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ff4e45b7-3128-4ce1-95d9-909f33d92301 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:51:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e7b4207c-f95e-49ed-815c-ab81afbfeb3f is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:51:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ba7d4726-e54a-4b5b-b060-92175ab177a2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:51:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:51:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 404270fa-50db-4a78-8c44-21df48acba39 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:51:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 00:51:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ff4e45b7-3128-4ce1-95d9-909f33d92301 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:51:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e7b4207c-f95e-49ed-815c-ab81afbfeb3f is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:51:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ba7d4726-e54a-4b5b-b060-92175ab177a2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:51:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:51:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 404270fa-50db-4a78-8c44-21df48acba39 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:51:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 00:51:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ff4e45b7-3128-4ce1-95d9-909f33d92301 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:51:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e7b4207c-f95e-49ed-815c-ab81afbfeb3f is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:51:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ba7d4726-e54a-4b5b-b060-92175ab177a2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:51:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:51:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 404270fa-50db-4a78-8c44-21df48acba39 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:51:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 00:51:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ff4e45b7-3128-4ce1-95d9-909f33d92301 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:51:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e7b4207c-f95e-49ed-815c-ab81afbfeb3f is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:51:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ba7d4726-e54a-4b5b-b060-92175ab177a2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:51:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:51:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 404270fa-50db-4a78-8c44-21df48acba39 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:51:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 00:51:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ff4e45b7-3128-4ce1-95d9-909f33d92301 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:51:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e7b4207c-f95e-49ed-815c-ab81afbfeb3f is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:51:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ba7d4726-e54a-4b5b-b060-92175ab177a2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:51:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:51:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 404270fa-50db-4a78-8c44-21df48acba39 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:51:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 00:51:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ff4e45b7-3128-4ce1-95d9-909f33d92301 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:51:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e7b4207c-f95e-49ed-815c-ab81afbfeb3f is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:51:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ba7d4726-e54a-4b5b-b060-92175ab177a2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:51:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:51:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 404270fa-50db-4a78-8c44-21df48acba39 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:51:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 00:51:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ff4e45b7-3128-4ce1-95d9-909f33d92301 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:51:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e7b4207c-f95e-49ed-815c-ab81afbfeb3f is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:51:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ba7d4726-e54a-4b5b-b060-92175ab177a2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:51:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:51:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 404270fa-50db-4a78-8c44-21df48acba39 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:51:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 00:51:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ff4e45b7-3128-4ce1-95d9-909f33d92301 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:51:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e7b4207c-f95e-49ed-815c-ab81afbfeb3f is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:51:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ba7d4726-e54a-4b5b-b060-92175ab177a2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:51:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:51:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 404270fa-50db-4a78-8c44-21df48acba39 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:51:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 00:51:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ff4e45b7-3128-4ce1-95d9-909f33d92301 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:51:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e7b4207c-f95e-49ed-815c-ab81afbfeb3f is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:51:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ba7d4726-e54a-4b5b-b060-92175ab177a2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:51:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:51:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 404270fa-50db-4a78-8c44-21df48acba39 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:51:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 00:51:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ff4e45b7-3128-4ce1-95d9-909f33d92301 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:51:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e7b4207c-f95e-49ed-815c-ab81afbfeb3f is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:51:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ba7d4726-e54a-4b5b-b060-92175ab177a2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:51:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:51:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 404270fa-50db-4a78-8c44-21df48acba39 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:51:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 00:51:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ff4e45b7-3128-4ce1-95d9-909f33d92301 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:51:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e7b4207c-f95e-49ed-815c-ab81afbfeb3f is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:51:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ba7d4726-e54a-4b5b-b060-92175ab177a2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:51:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:51:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 404270fa-50db-4a78-8c44-21df48acba39 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:51:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 00:51:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ff4e45b7-3128-4ce1-95d9-909f33d92301 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:51:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e7b4207c-f95e-49ed-815c-ab81afbfeb3f is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:51:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ba7d4726-e54a-4b5b-b060-92175ab177a2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:51:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:51:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 404270fa-50db-4a78-8c44-21df48acba39 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:51:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 00:51:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ff4e45b7-3128-4ce1-95d9-909f33d92301 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:51:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e7b4207c-f95e-49ed-815c-ab81afbfeb3f is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:51:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ba7d4726-e54a-4b5b-b060-92175ab177a2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:51:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:51:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 404270fa-50db-4a78-8c44-21df48acba39 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:51:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 00:51:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ff4e45b7-3128-4ce1-95d9-909f33d92301 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:51:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e7b4207c-f95e-49ed-815c-ab81afbfeb3f is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:51:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ba7d4726-e54a-4b5b-b060-92175ab177a2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:51:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:51:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 404270fa-50db-4a78-8c44-21df48acba39 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:51:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 00:51:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ff4e45b7-3128-4ce1-95d9-909f33d92301 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:51:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e7b4207c-f95e-49ed-815c-ab81afbfeb3f is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:51:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ba7d4726-e54a-4b5b-b060-92175ab177a2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:51:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:51:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 404270fa-50db-4a78-8c44-21df48acba39 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:51:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 00:51:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ff4e45b7-3128-4ce1-95d9-909f33d92301 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:51:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e7b4207c-f95e-49ed-815c-ab81afbfeb3f is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:51:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ba7d4726-e54a-4b5b-b060-92175ab177a2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:51:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:51:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 404270fa-50db-4a78-8c44-21df48acba39 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:51:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 00:51:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ff4e45b7-3128-4ce1-95d9-909f33d92301 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:51:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e7b4207c-f95e-49ed-815c-ab81afbfeb3f is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:51:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ba7d4726-e54a-4b5b-b060-92175ab177a2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:51:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:51:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 404270fa-50db-4a78-8c44-21df48acba39 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:51:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 00:52:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ff4e45b7-3128-4ce1-95d9-909f33d92301 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:52:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e7b4207c-f95e-49ed-815c-ab81afbfeb3f is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:52:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ba7d4726-e54a-4b5b-b060-92175ab177a2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:52:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:52:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 404270fa-50db-4a78-8c44-21df48acba39 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:52:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 00:52:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ff4e45b7-3128-4ce1-95d9-909f33d92301 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:52:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e7b4207c-f95e-49ed-815c-ab81afbfeb3f is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:52:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ba7d4726-e54a-4b5b-b060-92175ab177a2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:52:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:52:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 404270fa-50db-4a78-8c44-21df48acba39 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:52:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 00:52:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ff4e45b7-3128-4ce1-95d9-909f33d92301 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:52:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e7b4207c-f95e-49ed-815c-ab81afbfeb3f is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:52:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ba7d4726-e54a-4b5b-b060-92175ab177a2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:52:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:52:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 404270fa-50db-4a78-8c44-21df48acba39 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:52:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 00:52:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ff4e45b7-3128-4ce1-95d9-909f33d92301 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:52:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e7b4207c-f95e-49ed-815c-ab81afbfeb3f is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:52:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ba7d4726-e54a-4b5b-b060-92175ab177a2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:52:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:52:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 404270fa-50db-4a78-8c44-21df48acba39 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:52:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "",
                                "",
                                "PLAY [Set kolla_action_rabbitmq] ***********************************************",
                                "",
                                "TASK [Inform the user about the following task] ********************************",
                                "Tuesday 24 February 2026  00:49:57 +0000 (0:00:00.086)       0:00:00.086 ****** ",
                                "\u001b[0;32mok: [localhost] => {\u001b[0m",
                                "\u001b[0;32m    \"msg\": \"The task 'Check RabbitMQ service' fails if the RabbitMQ service has not yet been deployed. This is fine.\"\u001b[0m",
                                "\u001b[0;32m}\u001b[0m",
                                "",
                                "TASK [Check RabbitMQ service] **************************************************",
                                "Tuesday 24 February 2026  00:49:57 +0000 (0:00:00.040)       0:00:00.127 ****** ",
                                "\u001b[0;31mfatal: [localhost]: FAILED! => {\"changed\": false, \"elapsed\": 2, \"msg\": \"Timeout when waiting for search string RabbitMQ Management in 192.168.16.9:15672\"}\u001b[0m",
                                "\u001b[0;36m...ignoring\u001b[0m",
                                "",
                                "TASK [Set kolla_action_rabbitmq = upgrade if RabbitMQ is already running] ******",
                                "Tuesday 24 February 2026  00:50:00 +0000 (0:00:02.801)       0:00:02.929 ****** ",
                                "\u001b[0;36mskipping: [localhost]\u001b[0m",
                                "",
                                "TASK [Set kolla_action_rabbitmq = kolla_action_ng] *****************************",
                                "Tuesday 24 February 2026  00:50:00 +0000 (0:00:00.038)       0:00:02.968 ****** ",
                                "\u001b[0;32mok: [localhost]\u001b[0m",
                                "",
                                "PLAY [Group hosts based on configuration] **************************************",
                                "",
                                "TASK [Group hosts based on Kolla action] ***************************************",
                                "Tuesday 24 February 2026  00:50:00 +0000 (0:00:00.169)       0:00:03.137 ****** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [Group hosts based on enabled services] ***********************************",
                                "Tuesday 24 February 2026  00:50:00 +0000 (0:00:00.233)       0:00:03.370 ****** ",
                                "\u001b[0;32mok: [testbed-node-0] => (item=enable_rabbitmq_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item=enable_rabbitmq_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item=enable_rabbitmq_True)\u001b[0m",
                                "",
                                "PLAY [Apply role rabbitmq] *****************************************************",
                                "",
                                "TASK [rabbitmq : include_tasks] ************************************************",
                                "Tuesday 24 February 2026  00:50:01 +0000 (0:00:00.603)       0:00:03.974 ****** ",
                                "\u001b[0;36mincluded: /ansible/roles/rabbitmq/tasks/deploy.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [rabbitmq : Get container facts] ******************************************",
                                "Tuesday 24 February 2026  00:50:01 +0000 (0:00:00.593)       0:00:04.567 ****** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [rabbitmq : Get current RabbitMQ version] *********************************",
                                "Tuesday 24 February 2026  00:50:03 +0000 (0:00:01.205)       0:00:05.773 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [rabbitmq : Get new RabbitMQ version] *************************************",
                                "Tuesday 24 February 2026  00:50:04 +0000 (0:00:00.935)       0:00:06.708 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [rabbitmq : Check if running RabbitMQ is at most one version behind] ******",
                                "Tuesday 24 February 2026  00:50:04 +0000 (0:00:00.713)       0:00:07.422 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [rabbitmq : Catch when RabbitMQ is being downgraded] **********************",
                                "Tuesday 24 February 2026  00:50:05 +0000 (0:00:00.436)       0:00:07.858 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [rabbitmq : include_tasks] ************************************************",
                                "Tuesday 24 February 2026  00:50:07 +0000 (0:00:02.218)       0:00:10.077 ****** ",
                                "\u001b[0;36mincluded: /ansible/roles/rabbitmq/tasks/remove-ha-all-policy.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [rabbitmq : Get container facts] ******************************************",
                                "Tuesday 24 February 2026  00:50:09 +0000 (0:00:01.657)       0:00:11.735 ****** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [rabbitmq : List RabbitMQ policies] ***************************************",
                                "Tuesday 24 February 2026  00:50:09 +0000 (0:00:00.705)       0:00:12.440 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [rabbitmq : Remove ha-all policy from RabbitMQ] ***************************",
                                "Tuesday 24 February 2026  00:50:10 +0000 (0:00:00.360)       0:00:12.801 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [rabbitmq : Ensuring config directories exist] ****************************",
                                "Tuesday 24 February 2026  00:50:10 +0000 (0:00:00.384)       0:00:13.185 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'rabbitmq', 'value': {'container_name': 'rabbitmq', 'group': 'rabbitmq', 'enabled': True, 'image': 'registry.osism.tech/kolla/rabbitmq:2024.2', 'bootstrap_environment': {'KOLLA_BOOTSTRAP': None, 'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'volumes': ['/etc/kolla/rabbitmq/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'rabbitmq:/var/lib/rabbitmq/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_rabbitmq'], 'timeout': '30'}, 'haproxy': {'rabbitmq_management': {'enabled': 'yes', 'mode': 'http', 'port': '15672', 'host_group': 'rabbitmq'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'rabbitmq', 'value': {'container_name': 'rabbitmq', 'group': 'rabbitmq', 'enabled': True, 'image': 'registry.osism.tech/kolla/rabbitmq:2024.2', 'bootstrap_environment': {'KOLLA_BOOTSTRAP': None, 'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'volumes': ['/etc/kolla/rabbitmq/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'rabbitmq:/var/lib/rabbitmq/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_rabbitmq'], 'timeout': '30'}, 'haproxy': {'rabbitmq_management': {'enabled': 'yes', 'mode': 'http', 'port': '15672', 'host_group': 'rabbitmq'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'rabbitmq', 'value': {'container_name': 'rabbitmq', 'group': 'rabbitmq', 'enabled': True, 'image': 'registry.osism.tech/kolla/rabbitmq:2024.2', 'bootstrap_environment': {'KOLLA_BOOTSTRAP': None, 'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'volumes': ['/etc/kolla/rabbitmq/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'rabbitmq:/var/lib/rabbitmq/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_rabbitmq'], 'timeout': '30'}, 'haproxy': {'rabbitmq_management': {'enabled': 'yes', 'mode': 'http', 'port': '15672', 'host_group': 'rabbitmq'}}}})\u001b[0m",
                                "",
                                "TASK [rabbitmq : Copying over config.json files for services] ******************",
                                "Tuesday 24 February 2026  00:50:11 +0000 (0:00:01.189)       0:00:14.375 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'rabbitmq', 'value': {'container_name': 'rabbitmq', 'group': 'rabbitmq', 'enabled': True, 'image': 'registry.osism.tech/kolla/rabbitmq:2024.2', 'bootstrap_environment': {'KOLLA_BOOTSTRAP': None, 'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'volumes': ['/etc/kolla/rabbitmq/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'rabbitmq:/var/lib/rabbitmq/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_rabbitmq'], 'timeout': '30'}, 'haproxy': {'rabbitmq_management': {'enabled': 'yes', 'mode': 'http', 'port': '15672', 'host_group': 'rabbitmq'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'rabbitmq', 'value': {'container_name': 'rabbitmq', 'group': 'rabbitmq', 'enabled': True, 'image': 'registry.osism.tech/kolla/rabbitmq:2024.2', 'bootstrap_environment': {'KOLLA_BOOTSTRAP': None, 'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'volumes': ['/etc/kolla/rabbitmq/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'rabbitmq:/var/lib/rabbitmq/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_rabbitmq'], 'timeout': '30'}, 'haproxy': {'rabbitmq_management': {'enabled': 'yes', 'mode': 'http', 'port': '15672', 'host_group': 'rabbitmq'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'rabbitmq', 'value': {'container_name': 'rabbitmq', 'group': 'rabbitmq', 'enabled': True, 'image': 'registry.osism.tech/kolla/rabbitmq:2024.2', 'bootstrap_environment': {'KOLLA_BOOTSTRAP': None, 'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'volumes': ['/etc/kolla/rabbitmq/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'rabbitmq:/var/lib/rabbitmq/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_rabbitmq'], 'timeout': '30'}, 'haproxy': {'rabbitmq_management': {'enabled': 'yes', 'mode': 'http', 'port': '15672', 'host_group': 'rabbitmq'}}}})\u001b[0m",
                                "",
                                "TASK [rabbitmq : Copying over rabbitmq-env.conf] *******************************",
                                "Tuesday 24 February 2026  00:50:14 +0000 (0:00:02.430)       0:00:16.806 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=/ansible/roles/rabbitmq/templates/rabbitmq-env.conf.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=/ansible/roles/rabbitmq/templates/rabbitmq-env.conf.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=/ansible/roles/rabbitmq/templates/rabbitmq-env.conf.j2)\u001b[0m",
                                "",
                                "TASK [rabbitmq : Copying over rabbitmq.conf] ***********************************",
                                "Tuesday 24 February 2026  00:50:16 +0000 (0:00:01.812)       0:00:18.618 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=/ansible/roles/rabbitmq/templates/rabbitmq.conf.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=/ansible/roles/rabbitmq/templates/rabbitmq.conf.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=/ansible/roles/rabbitmq/templates/rabbitmq.conf.j2)\u001b[0m",
                                "",
                                "TASK [rabbitmq : Copying over erl_inetrc] **************************************",
                                "Tuesday 24 February 2026  00:50:17 +0000 (0:00:01.843)       0:00:20.462 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=/ansible/roles/rabbitmq/templates/erl_inetrc.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=/ansible/roles/rabbitmq/templates/erl_inetrc.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=/ansible/roles/rabbitmq/templates/erl_inetrc.j2)\u001b[0m",
                                "",
                                "TASK [rabbitmq : Copying over advanced.config] *********************************",
                                "Tuesday 24 February 2026  00:50:19 +0000 (0:00:01.294)       0:00:21.756 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=/ansible/roles/rabbitmq/templates/advanced.config.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=/ansible/roles/rabbitmq/templates/advanced.config.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=/ansible/roles/rabbitmq/templates/advanced.config.j2)\u001b[0m",
                                "",
                                "TASK [rabbitmq : Copying over definitions.json] ********************************",
                                "Tuesday 24 February 2026  00:50:20 +0000 (0:00:01.697)       0:00:23.454 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=/ansible/roles/rabbitmq/templates/definitions.json.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=/ansible/roles/rabbitmq/templates/definitions.json.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=/ansible/roles/rabbitmq/templates/definitions.json.j2)\u001b[0m",
                                "",
                                "TASK [rabbitmq : Copying over enabled_plugins] *********************************",
                                "Tuesday 24 February 2026  00:50:22 +0000 (0:00:01.489)       0:00:24.943 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=/ansible/roles/rabbitmq/templates/enabled_plugins.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=/ansible/roles/rabbitmq/templates/enabled_plugins.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=/ansible/roles/rabbitmq/templates/enabled_plugins.j2)\u001b[0m",
                                "",
                                "TASK [rabbitmq : include_tasks] ************************************************",
                                "Tuesday 24 February 2026  00:50:23 +0000 (0:00:01.649)       0:00:26.592 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [rabbitmq : Check rabbitmq containers] ************************************",
                                "Tuesday 24 February 2026  00:50:24 +0000 (0:00:00.427)       0:00:27.019 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'rabbitmq', 'value': {'container_name': 'rabbitmq', 'group': 'rabbitmq', 'enabled': True, 'image': 'registry.osism.tech/kolla/rabbitmq:2024.2', 'bootstrap_environment': {'KOLLA_BOOTSTRAP': None, 'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'volumes': ['/etc/kolla/rabbitmq/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'rabbitmq:/var/lib/rabbitmq/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_rabbitmq'], 'timeout': '30'}, 'haproxy': {'rabbitmq_management': {'enabled': 'yes', 'mode': 'http', 'port': '15672', 'host_group': 'rabbitmq'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'rabbitmq', 'value': {'container_name': 'rabbitmq', 'group': 'rabbitmq', 'enabled': True, 'image': 'registry.osism.tech/kolla/rabbitmq:2024.2', 'bootstrap_environment': {'KOLLA_BOOTSTRAP': None, 'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'volumes': ['/etc/kolla/rabbitmq/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'rabbitmq:/var/lib/rabbitmq/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_rabbitmq'], 'timeout': '30'}, 'haproxy': {'rabbitmq_management': {'enabled': 'yes', 'mode': 'http', 'port': '15672', 'host_group': 'rabbitmq'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'rabbitmq', 'value': {'container_name': 'rabbitmq', 'group': 'rabbitmq', 'enabled': True, 'image': 'registry.osism.tech/kolla/rabbitmq:2024.2', 'bootstrap_environment': {'KOLLA_BOOTSTRAP': None, 'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': 'zdd6geSBXefcI7IoHnP1U1fxtRWS3u5QtnPCvQTT', 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'volumes': ['/etc/kolla/rabbitmq/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'rabbitmq:/var/lib/rabbitmq/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_rabbitmq'], 'timeout': '30'}, 'haproxy': {'rabbitmq_management': {'enabled': 'yes', 'mode': 'http', 'port': '15672', 'host_group': 'rabbitmq'}}}})\u001b[0m",
                                "",
                                "TASK [rabbitmq : Creating rabbitmq volume] *************************************",
                                "Tuesday 24 February 2026  00:50:26 +0000 (0:00:02.066)       0:00:29.086 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [rabbitmq : Running RabbitMQ bootstrap container] *************************",
                                "Tuesday 24 February 2026  00:50:27 +0000 (0:00:01.174)       0:00:30.260 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "",
                                "RUNNING HANDLER [rabbitmq : Restart rabbitmq container] ************************",
                                "Tuesday 24 February 2026  00:50:35 +0000 (0:00:07.674)       0:00:37.934 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "PLAY [Restart rabbitmq services] ***********************************************",
                                "",
                                "TASK [rabbitmq : Get info on RabbitMQ container] *******************************",
                                "Tuesday 24 February 2026  00:50:35 +0000 (0:00:00.282)       0:00:38.217 ****** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [rabbitmq : Put RabbitMQ node into maintenance mode] **********************",
                                "Tuesday 24 February 2026  00:50:36 +0000 (0:00:00.614)       0:00:38.831 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [rabbitmq : Restart rabbitmq container] ***********************************",
                                "Tuesday 24 February 2026  00:50:36 +0000 (0:00:00.213)       0:00:39.044 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [rabbitmq : Waiting for rabbitmq to start] ********************************",
                                "Tuesday 24 February 2026  00:50:38 +0000 (0:00:02.185)       0:00:41.230 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "PLAY [Restart rabbitmq services] ***********************************************",
                                "",
                                "TASK [rabbitmq : Get info on RabbitMQ container] *******************************",
                                "Tuesday 24 February 2026  00:51:34 +0000 (0:00:55.674)       0:01:36.905 ****** ",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [rabbitmq : Put RabbitMQ node into maintenance mode] **********************",
                                "Tuesday 24 February 2026  00:51:35 +0000 (0:00:00.758)       0:01:37.664 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [rabbitmq : Restart rabbitmq container] ***********************************",
                                "Tuesday 24 February 2026  00:51:35 +0000 (0:00:00.237)       0:01:37.901 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [rabbitmq : Waiting for rabbitmq to start] ********************************",
                                "Tuesday 24 February 2026  00:51:37 +0000 (0:00:02.552)       0:01:40.454 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "",
                                "PLAY [Restart rabbitmq services] ***********************************************",
                                "",
                                "TASK [rabbitmq : Get info on RabbitMQ container] *******************************",
                                "Tuesday 24 February 2026  00:51:52 +0000 (0:00:14.407)       0:01:54.862 ****** ",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [rabbitmq : Put RabbitMQ node into maintenance mode] **********************",
                                "Tuesday 24 February 2026  00:51:52 +0000 (0:00:00.728)       0:01:55.590 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [rabbitmq : Restart rabbitmq container] ***********************************",
                                "Tuesday 24 February 2026  00:51:53 +0000 (0:00:00.351)       0:01:55.941 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [rabbitmq : Waiting for rabbitmq to start] ********************************",
                                "Tuesday 24 February 2026  00:52:00 +0000 (0:00:06.886)       0:02:02.828 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "PLAY [Apply rabbitmq post-configuration] ***************************************",
                                "",
                                "TASK [Include rabbitmq post-deploy.yml] ****************************************",
                                "Tuesday 24 February 2026  00:52:10 +0000 (0:00:10.477)       0:02:13.306 ****** ",
                                "\u001b[0;36mincluded: rabbitmq for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [rabbitmq : Enable all stable feature flags] ******************************",
                                "Tuesday 24 February 2026  00:52:11 +0000 (0:00:00.731)       0:02:14.037 ****** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[1;35m[WARNING]: Could not match supplied host pattern, ignoring:\u001b[0m",
                                "\u001b[1;35menable_outward_rabbitmq_True\u001b[0m",
                                "",
                                "PLAY [Apply role rabbitmq (outward)] *******************************************",
                                "\u001b[0;36mskipping: no hosts matched\u001b[0m",
                                "\u001b[1;35m[WARNING]: Could not match supplied host pattern, ignoring:\u001b[0m",
                                "\u001b[1;35moutward_rabbitmq_restart\u001b[0m",
                                "",
                                "PLAY [Restart rabbitmq (outward) services] *************************************",
                                "\u001b[0;36mskipping: no hosts matched\u001b[0m",
                                "",
                                "PLAY [Apply rabbitmq (outward) post-configuration] *****************************",
                                "\u001b[0;36mskipping: no hosts matched\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;32mlocalhost\u001b[0m                  : \u001b[0;32mok=3   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=1   \u001b[0m rescued=0    \u001b[1;35mignored=1   \u001b[0m",
                                "\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=23  \u001b[0m \u001b[0;33mchanged=14  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=8   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=21  \u001b[0m \u001b[0;33mchanged=14  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=21  \u001b[0m \u001b[0;33mchanged=14  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Tuesday 24 February 2026  00:52:13 +0000 (0:00:02.397)       0:02:16.434 ****** ",
                                "=============================================================================== ",
                                "rabbitmq : Waiting for rabbitmq to start ------------------------------- 80.56s",
                                "rabbitmq : Restart rabbitmq container ---------------------------------- 11.63s",
                                "rabbitmq : Running RabbitMQ bootstrap container ------------------------- 7.67s",
                                "Check RabbitMQ service -------------------------------------------------- 2.80s",
                                "rabbitmq : Copying over config.json files for services ------------------ 2.43s",
                                "rabbitmq : Enable all stable feature flags ------------------------------ 2.40s",
                                "rabbitmq : Catch when RabbitMQ is being downgraded ---------------------- 2.22s",
                                "rabbitmq : Get info on RabbitMQ container ------------------------------- 2.10s",
                                "rabbitmq : Check rabbitmq containers ------------------------------------ 2.07s",
                                "rabbitmq : Copying over rabbitmq.conf ----------------------------------- 1.84s",
                                "rabbitmq : Copying over rabbitmq-env.conf ------------------------------- 1.81s",
                                "rabbitmq : Copying over advanced.config --------------------------------- 1.70s",
                                "rabbitmq : include_tasks ------------------------------------------------ 1.66s",
                                "rabbitmq : Copying over enabled_plugins --------------------------------- 1.65s",
                                "rabbitmq : Copying over definitions.json -------------------------------- 1.49s",
                                "rabbitmq : Copying over erl_inetrc -------------------------------------- 1.29s",
                                "rabbitmq : Get container facts ------------------------------------------ 1.21s",
                                "rabbitmq : Ensuring config directories exist ---------------------------- 1.19s",
                                "rabbitmq : Creating rabbitmq volume ------------------------------------- 1.17s",
                                "rabbitmq : Get current RabbitMQ version --------------------------------- 0.94s",
                                "\u001b[32m2026-02-24 00:52:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ff4e45b7-3128-4ce1-95d9-909f33d92301 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:52:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask e7b4207c-f95e-49ed-815c-ab81afbfeb3f is in state SUCCESS\u001b[0m",
                                "\u001b[32m2026-02-24 00:52:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ba7d4726-e54a-4b5b-b060-92175ab177a2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:52:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:52:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 404270fa-50db-4a78-8c44-21df48acba39 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:52:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "",
                                "\u001b[32m2026-02-24 00:52:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ff4e45b7-3128-4ce1-95d9-909f33d92301 is in state SUCCESS\u001b[0m",
                                "",
                                "PLAY [Prepare all k3s nodes] ***************************************************",
                                "",
                                "TASK [k3s_prereq : Validating arguments against arg spec 'main' - Prerequisites] ***",
                                "Tuesday 24 February 2026  00:47:14 +0000 (0:00:00.200)       0:00:00.200 ****** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [k3s_prereq : Set same timezone on every Server] **************************",
                                "Tuesday 24 February 2026  00:47:15 +0000 (0:00:00.745)       0:00:00.946 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [k3s_prereq : Set SELinux to disabled state] ******************************",
                                "Tuesday 24 February 2026  00:47:16 +0000 (0:00:00.757)       0:00:01.703 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [k3s_prereq : Enable IPv4 forwarding] *************************************",
                                "Tuesday 24 February 2026  00:47:16 +0000 (0:00:00.778)       0:00:02.482 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [k3s_prereq : Enable IPv6 forwarding] *************************************",
                                "Tuesday 24 February 2026  00:47:19 +0000 (0:00:02.159)       0:00:04.642 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [k3s_prereq : Enable IPv6 router advertisements] **************************",
                                "Tuesday 24 February 2026  00:47:20 +0000 (0:00:01.258)       0:00:05.900 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [k3s_prereq : Add br_netfilter to /etc/modules-load.d/] *******************",
                                "Tuesday 24 February 2026  00:47:22 +0000 (0:00:01.954)       0:00:07.854 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [k3s_prereq : Load br_netfilter] ******************************************",
                                "Tuesday 24 February 2026  00:47:23 +0000 (0:00:00.839)       0:00:08.694 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [k3s_prereq : Set bridge-nf-call-iptables (just to be sure)] **************",
                                "Tuesday 24 February 2026  00:47:24 +0000 (0:00:00.837)       0:00:09.532 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=net.bridge.bridge-nf-call-iptables) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=net.bridge.bridge-nf-call-ip6tables) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=net.bridge.bridge-nf-call-iptables) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=net.bridge.bridge-nf-call-ip6tables) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=net.bridge.bridge-nf-call-iptables) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=net.bridge.bridge-nf-call-ip6tables) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=net.bridge.bridge-nf-call-iptables) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=net.bridge.bridge-nf-call-ip6tables) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=net.bridge.bridge-nf-call-iptables) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=net.bridge.bridge-nf-call-ip6tables) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=net.bridge.bridge-nf-call-iptables) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=net.bridge.bridge-nf-call-ip6tables) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [k3s_prereq : Add /usr/local/bin to sudo secure_path] *********************",
                                "Tuesday 24 February 2026  00:47:24 +0000 (0:00:00.889)       0:00:10.422 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [k3s_download : Validating arguments against arg spec 'main' - Manage the downloading of K3S binaries] ***",
                                "Tuesday 24 February 2026  00:47:26 +0000 (0:00:01.499)       0:00:11.921 ****** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [k3s_download : Download k3s binary x64] **********************************",
                                "Tuesday 24 February 2026  00:47:27 +0000 (0:00:01.145)       0:00:13.067 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [k3s_download : Download k3s binary arm64] ********************************",
                                "Tuesday 24 February 2026  00:47:33 +0000 (0:00:05.707)       0:00:18.774 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [k3s_download : Download k3s binary armhf] ********************************",
                                "Tuesday 24 February 2026  00:47:35 +0000 (0:00:01.833)       0:00:20.608 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [k3s_custom_registries : Validating arguments against arg spec 'main' - Configure the use of a custom container registry] ***",
                                "Tuesday 24 February 2026  00:47:37 +0000 (0:00:02.094)       0:00:22.702 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [k3s_custom_registries : Create directory /etc/rancher/k3s] ***************",
                                "Tuesday 24 February 2026  00:47:38 +0000 (0:00:01.638)       0:00:24.340 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=rancher) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=rancher/k3s) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=rancher) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=rancher/k3s) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=rancher) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=rancher/k3s) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=rancher) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=rancher/k3s) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=rancher) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=rancher/k3s) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=rancher) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=rancher/k3s) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [k3s_custom_registries : Insert registries into /etc/rancher/k3s/registries.yaml] ***",
                                "Tuesday 24 February 2026  00:47:40 +0000 (0:00:01.815)       0:00:26.155 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [k3s_custom_registries : Remove /etc/rancher/k3s/registries.yaml when no registries configured] ***",
                                "Tuesday 24 February 2026  00:47:41 +0000 (0:00:00.953)       0:00:27.109 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "PLAY [Deploy k3s master nodes] *************************************************",
                                "",
                                "TASK [k3s_server : Validating arguments against arg spec 'main' - Setup k3s servers] ***",
                                "Tuesday 24 February 2026  00:47:43 +0000 (0:00:01.457)       0:00:28.567 ****** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [k3s_server : Stop k3s-init] **********************************************",
                                "Tuesday 24 February 2026  00:47:44 +0000 (0:00:01.351)       0:00:29.919 ****** ",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [k3s_server : Stop k3s] ***************************************************",
                                "Tuesday 24 February 2026  00:47:45 +0000 (0:00:01.528)       0:00:31.447 ****** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [k3s_server : Clean previous runs of k3s-init] ****************************",
                                "Tuesday 24 February 2026  00:47:47 +0000 (0:00:01.110)       0:00:32.557 ****** ",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [k3s_server : Deploy K3s http_proxy conf] *********************************",
                                "Tuesday 24 February 2026  00:47:48 +0000 (0:00:01.019)       0:00:33.576 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [k3s_server : Create /etc/rancher/k3s directory] **************************",
                                "Tuesday 24 February 2026  00:47:48 +0000 (0:00:00.498)       0:00:34.075 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [k3s_server : Create custom resolv.conf for k3s] **************************",
                                "Tuesday 24 February 2026  00:47:49 +0000 (0:00:01.037)       0:00:35.113 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [k3s_server : Deploy vip manifest] ****************************************",
                                "Tuesday 24 February 2026  00:47:51 +0000 (0:00:01.960)       0:00:37.073 ****** ",
                                "\u001b[0;36mincluded: /ansible/roles/k3s_server/tasks/vip.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [k3s_server : Set _kube_vip_bgp_peers fact] *******************************",
                                "Tuesday 24 February 2026  00:47:52 +0000 (0:00:00.618)       0:00:37.692 ****** ",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [k3s_server : Create manifests directory on first master] *****************",
                                "Tuesday 24 February 2026  00:47:55 +0000 (0:00:02.833)       0:00:40.525 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [k3s_server : Download vip rbac manifest to first master] *****************",
                                "Tuesday 24 February 2026  00:47:56 +0000 (0:00:01.085)       0:00:41.611 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [k3s_server : Copy vip manifest to first master] **************************",
                                "Tuesday 24 February 2026  00:47:57 +0000 (0:00:01.741)       0:00:43.353 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [k3s_server : Deploy metallb manifest] ************************************",
                                "Tuesday 24 February 2026  00:47:59 +0000 (0:00:01.809)       0:00:45.162 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [k3s_server : Deploy kube-vip manifest] ***********************************",
                                "Tuesday 24 February 2026  00:48:00 +0000 (0:00:00.689)       0:00:45.852 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [k3s_server : Init cluster inside the transient k3s-init service] *********",
                                "Tuesday 24 February 2026  00:48:00 +0000 (0:00:00.420)       0:00:46.272 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [k3s_server : Detect Kubernetes version for label compatibility] **********",
                                "Tuesday 24 February 2026  00:48:03 +0000 (0:00:02.332)       0:00:48.604 ****** ",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [k3s_server : Set node role label selector based on Kubernetes version] ***",
                                "Tuesday 24 February 2026  00:48:05 +0000 (0:00:02.802)       0:00:51.407 ****** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [k3s_server : Verify that all nodes actually joined (check k3s-init.service if this fails)] ***",
                                "Tuesday 24 February 2026  00:48:06 +0000 (0:00:00.862)       0:00:52.270 ****** ",
                                "\u001b[1;30mFAILED - RETRYING: [testbed-node-1]: Verify that all nodes actually joined (check k3s-init.service if this fails) (20 retries left).\u001b[0m",
                                "\u001b[1;30mFAILED - RETRYING: [testbed-node-0]: Verify that all nodes actually joined (check k3s-init.service if this fails) (20 retries left).\u001b[0m",
                                "\u001b[1;30mFAILED - RETRYING: [testbed-node-2]: Verify that all nodes actually joined (check k3s-init.service if this fails) (20 retries left).\u001b[0m",
                                "\u001b[1;30mFAILED - RETRYING: [testbed-node-0]: Verify that all nodes actually joined (check k3s-init.service if this fails) (19 retries left).\u001b[0m",
                                "\u001b[1;30mFAILED - RETRYING: [testbed-node-1]: Verify that all nodes actually joined (check k3s-init.service if this fails) (19 retries left).\u001b[0m",
                                "\u001b[1;30mFAILED - RETRYING: [testbed-node-2]: Verify that all nodes actually joined (check k3s-init.service if this fails) (19 retries left).\u001b[0m",
                                "\u001b[1;30mFAILED - RETRYING: [testbed-node-0]: Verify that all nodes actually joined (check k3s-init.service if this fails) (18 retries left).\u001b[0m",
                                "\u001b[1;30mFAILED - RETRYING: [testbed-node-1]: Verify that all nodes actually joined (check k3s-init.service if this fails) (18 retries left).\u001b[0m",
                                "\u001b[1;30mFAILED - RETRYING: [testbed-node-2]: Verify that all nodes actually joined (check k3s-init.service if this fails) (18 retries left).\u001b[0m",
                                "\u001b[1;30mFAILED - RETRYING: [testbed-node-0]: Verify that all nodes actually joined (check k3s-init.service if this fails) (17 retries left).\u001b[0m",
                                "\u001b[1;30mFAILED - RETRYING: [testbed-node-1]: Verify that all nodes actually joined (check k3s-init.service if this fails) (17 retries left).\u001b[0m",
                                "\u001b[1;30mFAILED - RETRYING: [testbed-node-2]: Verify that all nodes actually joined (check k3s-init.service if this fails) (17 retries left).\u001b[0m",
                                "\u001b[1;30mFAILED - RETRYING: [testbed-node-0]: Verify that all nodes actually joined (check k3s-init.service if this fails) (16 retries left).\u001b[0m",
                                "\u001b[1;30mFAILED - RETRYING: [testbed-node-1]: Verify that all nodes actually joined (check k3s-init.service if this fails) (16 retries left).\u001b[0m",
                                "\u001b[1;30mFAILED - RETRYING: [testbed-node-2]: Verify that all nodes actually joined (check k3s-init.service if this fails) (16 retries left).\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [k3s_server : Save logs of k3s-init.service] ******************************",
                                "Tuesday 24 February 2026  00:49:00 +0000 (0:00:54.020)       0:01:46.290 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [k3s_server : Kill the temporary service used for initialization] *********",
                                "Tuesday 24 February 2026  00:49:01 +0000 (0:00:00.623)       0:01:46.913 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [k3s_server : Copy K3s service file] **************************************",
                                "Tuesday 24 February 2026  00:49:02 +0000 (0:00:01.184)       0:01:48.099 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [k3s_server : Enable and check K3s service] *******************************",
                                "Tuesday 24 February 2026  00:49:03 +0000 (0:00:01.320)       0:01:49.419 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [k3s_server : Wait for node-token] ****************************************",
                                "Tuesday 24 February 2026  00:49:57 +0000 (0:00:53.123)       0:02:42.542 ****** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [k3s_server : Register node-token file access mode] ***********************",
                                "Tuesday 24 February 2026  00:49:57 +0000 (0:00:00.633)       0:02:43.176 ****** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [k3s_server : Change file access node-token] ******************************",
                                "Tuesday 24 February 2026  00:49:58 +0000 (0:00:00.632)       0:02:43.809 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [k3s_server : Read node-token from master] ********************************",
                                "Tuesday 24 February 2026  00:49:58 +0000 (0:00:00.534)       0:02:44.343 ****** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [k3s_server : Store Master node-token] ************************************",
                                "Tuesday 24 February 2026  00:49:59 +0000 (0:00:00.760)       0:02:45.103 ****** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [k3s_server : Restore node-token file access] *****************************",
                                "Tuesday 24 February 2026  00:49:59 +0000 (0:00:00.340)       0:02:45.443 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [k3s_server : Create directory .kube] *************************************",
                                "Tuesday 24 February 2026  00:50:00 +0000 (0:00:00.636)       0:02:46.080 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [k3s_server : Copy config file to user home directory] ********************",
                                "Tuesday 24 February 2026  00:50:01 +0000 (0:00:00.563)       0:02:46.644 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [k3s_server : Configure kubectl cluster to https://192.168.16.8:6443] *****",
                                "Tuesday 24 February 2026  00:50:01 +0000 (0:00:00.852)       0:02:47.497 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [k3s_server : Create kubectl symlink] *************************************",
                                "Tuesday 24 February 2026  00:50:02 +0000 (0:00:00.665)       0:02:48.163 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [k3s_server : Create crictl symlink] **************************************",
                                "Tuesday 24 February 2026  00:50:02 +0000 (0:00:00.326)       0:02:48.490 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [k3s_server : Get contents of manifests folder] ***************************",
                                "Tuesday 24 February 2026  00:50:03 +0000 (0:00:00.391)       0:02:48.881 ****** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [k3s_server : Get sub dirs of manifests folder] ***************************",
                                "Tuesday 24 February 2026  00:50:04 +0000 (0:00:01.028)       0:02:49.910 ****** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [k3s_server : Remove manifests and folders that are only needed for bootstrapping cluster so k3s doesn't auto apply on start] ***",
                                "Tuesday 24 February 2026  00:50:05 +0000 (0:00:00.678)       0:02:50.588 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=/var/lib/rancher/k3s/server/manifests/rolebindings.yaml)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=/var/lib/rancher/k3s/server/manifests/rolebindings.yaml)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=/var/lib/rancher/k3s/server/manifests/rolebindings.yaml)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=/var/lib/rancher/k3s/server/manifests/local-storage.yaml)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=/var/lib/rancher/k3s/server/manifests/local-storage.yaml)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=/var/lib/rancher/k3s/server/manifests/local-storage.yaml)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=/var/lib/rancher/k3s/server/manifests/coredns.yaml)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=/var/lib/rancher/k3s/server/manifests/coredns.yaml)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=/var/lib/rancher/k3s/server/manifests/vip.yaml)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=/var/lib/rancher/k3s/server/manifests/coredns.yaml)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=/var/lib/rancher/k3s/server/manifests/runtimes.yaml)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=/var/lib/rancher/k3s/server/manifests/vip-rbac.yaml)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=/var/lib/rancher/k3s/server/manifests/runtimes.yaml)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=/var/lib/rancher/k3s/server/manifests/ccm.yaml)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=/var/lib/rancher/k3s/server/manifests/runtimes.yaml)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=/var/lib/rancher/k3s/server/manifests/ccm.yaml)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=/var/lib/rancher/k3s/server/manifests/ccm.yaml)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=/var/lib/rancher/k3s/server/manifests/metrics-server)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=/var/lib/rancher/k3s/server/manifests/metrics-server)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=/var/lib/rancher/k3s/server/manifests/metrics-server)\u001b[0m",
                                "",
                                "PLAY [Deploy k3s worker nodes] *************************************************",
                                "",
                                "TASK [k3s_agent : Validating arguments against arg spec 'main' - Setup k3s agents] ***",
                                "Tuesday 24 February 2026  00:50:08 +0000 (0:00:03.029)       0:02:53.617 ****** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [k3s_agent : Check if system is PXE-booted] *******************************",
                                "Tuesday 24 February 2026  00:50:08 +0000 (0:00:00.489)       0:02:54.107 ****** ",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [k3s_agent : Set fact for PXE-booted system] ******************************",
                                "Tuesday 24 February 2026  00:50:10 +0000 (0:00:01.680)       0:02:55.787 ****** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [k3s_agent : Include http_proxy configuration tasks] **********************",
                                "Tuesday 24 February 2026  00:50:10 +0000 (0:00:00.397)       0:02:56.185 ****** ",
                                "\u001b[0;36mincluded: /ansible/roles/k3s_agent/tasks/http_proxy.yml for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m",
                                "",
                                "TASK [k3s_agent : Create k3s-node.service.d directory] *************************",
                                "Tuesday 24 February 2026  00:50:11 +0000 (0:00:00.700)       0:02:56.885 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [k3s_agent : Copy K3s http_proxy conf file] *******************************",
                                "Tuesday 24 February 2026  00:50:11 +0000 (0:00:00.228)       0:02:57.113 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [k3s_agent : Deploy K3s http_proxy conf] **********************************",
                                "Tuesday 24 February 2026  00:50:11 +0000 (0:00:00.286)       0:02:57.400 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [k3s_agent : Create /etc/rancher/k3s directory] ***************************",
                                "Tuesday 24 February 2026  00:50:12 +0000 (0:00:00.279)       0:02:57.680 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [k3s_agent : Create custom resolv.conf for k3s] ***************************",
                                "Tuesday 24 February 2026  00:50:13 +0000 (0:00:00.861)       0:02:58.541 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [k3s_agent : Configure the k3s service] ***********************************",
                                "Tuesday 24 February 2026  00:50:14 +0000 (0:00:01.366)       0:02:59.908 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [k3s_agent : Manage k3s service] ******************************************",
                                "Tuesday 24 February 2026  00:50:16 +0000 (0:00:01.848)       0:03:01.757 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "",
                                "PLAY [Prepare kubeconfig file] *************************************************",
                                "",
                                "TASK [Get home directory of operator user] *************************************",
                                "Tuesday 24 February 2026  00:50:26 +0000 (0:00:10.203)       0:03:11.961 ****** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [Create .kube directory] **************************************************",
                                "Tuesday 24 February 2026  00:50:27 +0000 (0:00:00.753)       0:03:12.714 ****** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [Get kubeconfig file] *****************************************************",
                                "Tuesday 24 February 2026  00:50:27 +0000 (0:00:00.397)       0:03:13.112 ****** ",
                                "\u001b[0;32mok: [testbed-manager -> testbed-node-0(192.168.16.10)]\u001b[0m",
                                "",
                                "TASK [Write kubeconfig file] ***************************************************",
                                "Tuesday 24 February 2026  00:50:28 +0000 (0:00:00.543)       0:03:13.656 ****** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [Change server address in the kubeconfig] *********************************",
                                "Tuesday 24 February 2026  00:50:28 +0000 (0:00:00.749)       0:03:14.406 ****** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [Make kubeconfig available for use inside the manager service] ************",
                                "Tuesday 24 February 2026  00:50:29 +0000 (0:00:00.519)       0:03:14.926 ****** ",
                                "\u001b[0;33mchanged: [testbed-manager -> localhost]\u001b[0m",
                                "",
                                "TASK [Change server address in the kubeconfig inside the manager service] ******",
                                "Tuesday 24 February 2026  00:50:31 +0000 (0:00:01.615)       0:03:16.541 ****** ",
                                "\u001b[0;33mchanged: [testbed-manager -> localhost]\u001b[0m",
                                "",
                                "TASK [Set KUBECONFIG environment variable] *************************************",
                                "Tuesday 24 February 2026  00:50:31 +0000 (0:00:00.829)       0:03:17.371 ****** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [Enable kubectl command line completion] **********************************",
                                "Tuesday 24 February 2026  00:50:32 +0000 (0:00:00.492)       0:03:17.863 ****** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "PLAY [Apply role kubectl] ******************************************************",
                                "",
                                "TASK [kubectl : Gather variables for each operating system] ********************",
                                "Tuesday 24 February 2026  00:50:32 +0000 (0:00:00.420)       0:03:18.283 ****** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [kubectl : Include distribution specific install tasks] *******************",
                                "Tuesday 24 February 2026  00:50:32 +0000 (0:00:00.129)       0:03:18.412 ****** ",
                                "\u001b[0;36mincluded: /ansible/roles/kubectl/tasks/install-Debian-family.yml for testbed-manager\u001b[0m",
                                "",
                                "TASK [kubectl : Remove old architecture-dependent repository] ******************",
                                "Tuesday 24 February 2026  00:50:33 +0000 (0:00:00.229)       0:03:18.642 ****** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [kubectl : Install apt-transport-https package] ***************************",
                                "Tuesday 24 February 2026  00:50:33 +0000 (0:00:00.838)       0:03:19.481 ****** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [kubectl : Add repository gpg key] ****************************************",
                                "Tuesday 24 February 2026  00:50:35 +0000 (0:00:01.468)       0:03:20.949 ****** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [kubectl : Set permissions of gpg key] ************************************",
                                "Tuesday 24 February 2026  00:50:36 +0000 (0:00:00.718)       0:03:21.667 ****** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [kubectl : Add repository Debian] *****************************************",
                                "Tuesday 24 February 2026  00:50:36 +0000 (0:00:00.423)       0:03:22.091 ****** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [kubectl : Install required packages] *************************************",
                                "Tuesday 24 February 2026  00:50:43 +0000 (0:00:06.676)       0:03:28.768 ****** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [kubectl : Remove kubectl symlink] ****************************************",
                                "Tuesday 24 February 2026  00:50:55 +0000 (0:00:12.522)       0:03:41.291 ****** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "",
                                "PLAY [Run post actions on master nodes] ****************************************",
                                "",
                                "TASK [k3s_server_post : Validating arguments against arg spec 'main' - Configure k3s cluster] ***",
                                "Tuesday 24 February 2026  00:50:56 +0000 (0:00:00.520)       0:03:41.811 ****** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [k3s_server_post : Deploy calico] *****************************************",
                                "Tuesday 24 February 2026  00:50:56 +0000 (0:00:00.306)       0:03:42.118 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [k3s_server_post : Deploy cilium] *****************************************",
                                "Tuesday 24 February 2026  00:50:57 +0000 (0:00:00.616)       0:03:42.735 ****** ",
                                "\u001b[0;36mincluded: /ansible/roles/k3s_server_post/tasks/cilium.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [k3s_server_post : Create tmp directory on first master] ******************",
                                "Tuesday 24 February 2026  00:50:57 +0000 (0:00:00.763)       0:03:43.498 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0 -> localhost]\u001b[0m",
                                "",
                                "TASK [k3s_server_post : Wait for connectivity to kube VIP] *********************",
                                "Tuesday 24 February 2026  00:50:59 +0000 (0:00:01.044)       0:03:44.543 ****** ",
                                "\u001b[0;32mok: [testbed-node-0 -> localhost]\u001b[0m",
                                "",
                                "TASK [k3s_server_post : Fail if kube VIP not reachable] ************************",
                                "Tuesday 24 February 2026  00:50:59 +0000 (0:00:00.873)       0:03:45.416 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [k3s_server_post : Test for existing Cilium install] **********************",
                                "Tuesday 24 February 2026  00:51:00 +0000 (0:00:00.121)       0:03:45.538 ****** ",
                                "\u001b[0;32mok: [testbed-node-0 -> localhost]\u001b[0m",
                                "",
                                "TASK [k3s_server_post : Check Cilium version] **********************************",
                                "Tuesday 24 February 2026  00:51:01 +0000 (0:00:00.997)       0:03:46.535 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [k3s_server_post : Parse installed Cilium version] ************************",
                                "Tuesday 24 February 2026  00:51:01 +0000 (0:00:00.124)       0:03:46.660 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [k3s_server_post : Determine if Cilium needs update] **********************",
                                "Tuesday 24 February 2026  00:51:01 +0000 (0:00:00.114)       0:03:46.774 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [k3s_server_post : Log result] ********************************************",
                                "Tuesday 24 February 2026  00:51:01 +0000 (0:00:00.121)       0:03:46.896 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [k3s_server_post : Install Cilium] ****************************************",
                                "Tuesday 24 February 2026  00:51:01 +0000 (0:00:00.115)       0:03:47.011 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0 -> localhost]\u001b[0m",
                                "",
                                "TASK [k3s_server_post : Wait for Cilium resources] *****************************",
                                "Tuesday 24 February 2026  00:51:06 +0000 (0:00:05.132)       0:03:52.144 ****** ",
                                "\u001b[0;32mok: [testbed-node-0 -> localhost] => (item=deployment/cilium-operator)\u001b[0m",
                                "\u001b[1;30mFAILED - RETRYING: [testbed-node-0 -> localhost]: Wait for Cilium resources (30 retries left).\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0 -> localhost] => (item=daemonset/cilium)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0 -> localhost] => (item=deployment/hubble-relay)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0 -> localhost] => (item=deployment/hubble-ui)\u001b[0m",
                                "",
                                "TASK [k3s_server_post : Set _cilium_bgp_neighbors fact] ************************",
                                "Tuesday 24 February 2026  00:51:48 +0000 (0:00:42.155)       0:04:34.300 ****** ",
                                "\u001b[0;32mok: [testbed-node-0 -> localhost]\u001b[0m",
                                "",
                                "TASK [k3s_server_post : Copy BGP manifests to first master] ********************",
                                "Tuesday 24 February 2026  00:51:50 +0000 (0:00:01.220)       0:04:35.520 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0 -> localhost]\u001b[0m",
                                "",
                                "TASK [k3s_server_post : Apply BGP manifests] ***********************************",
                                "Tuesday 24 February 2026  00:51:51 +0000 (0:00:01.540)       0:04:37.060 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0 -> localhost]\u001b[0m",
                                "",
                                "TASK [k3s_server_post : Print error message if BGP manifests application fails] ***",
                                "Tuesday 24 February 2026  00:51:52 +0000 (0:00:00.842)       0:04:37.903 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [k3s_server_post : Test for BGP config resources] *************************",
                                "Tuesday 24 February 2026  00:51:52 +0000 (0:00:00.104)       0:04:38.008 ****** ",
                                "\u001b[0;32mok: [testbed-node-0 -> localhost] => (item=kubectl get CiliumBGPPeeringPolicy.cilium.io)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0 -> localhost] => (item=kubectl get CiliumLoadBalancerIPPool.cilium.io)\u001b[0m",
                                "",
                                "TASK [k3s_server_post : Deploy metallb pool] ***********************************",
                                "Tuesday 24 February 2026  00:51:54 +0000 (0:00:01.645)       0:04:39.654 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [k3s_server_post : Remove tmp directory used for manifests] ***************",
                                "Tuesday 24 February 2026  00:51:54 +0000 (0:00:00.258)       0:04:39.913 ****** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "PLAY [Apply role k9s] **********************************************************",
                                "",
                                "TASK [k9s : Gather variables for each operating system] ************************",
                                "Tuesday 24 February 2026  00:51:55 +0000 (0:00:00.974)       0:04:40.887 ****** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [k9s : Include distribution specific install tasks] ***********************",
                                "Tuesday 24 February 2026  00:51:55 +0000 (0:00:00.133)       0:04:41.021 ****** ",
                                "\u001b[0;36mincluded: /ansible/roles/k9s/tasks/install-Debian-family.yml for testbed-manager\u001b[0m",
                                "",
                                "TASK [k9s : Install k9s packages] **********************************************",
                                "Tuesday 24 February 2026  00:51:55 +0000 (0:00:00.192)       0:04:41.214 ****** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "PLAY [Manage labels, annotations, and taints on all k3s nodes] *****************",
                                "",
                                "TASK [Merge labels, annotations, and taints] ***********************************",
                                "Tuesday 24 February 2026  00:52:01 +0000 (0:00:05.433)       0:04:46.648 ****** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [Manage labels] ***********************************************************",
                                "Tuesday 24 February 2026  00:52:01 +0000 (0:00:00.673)       0:04:47.321 ****** ",
                                "\u001b[0;32mok: [testbed-node-3 -> localhost] => (item=node-role.osism.tech/compute-plane=true)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4 -> localhost] => (item=node-role.osism.tech/compute-plane=true)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5 -> localhost] => (item=node-role.osism.tech/compute-plane=true)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0 -> localhost] => (item=node-role.osism.tech/control-plane=true)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1 -> localhost] => (item=node-role.osism.tech/control-plane=true)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5 -> localhost] => (item=node-role.kubernetes.io/worker=worker)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2 -> localhost] => (item=node-role.osism.tech/control-plane=true)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3 -> localhost] => (item=node-role.kubernetes.io/worker=worker)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4 -> localhost] => (item=node-role.kubernetes.io/worker=worker)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0 -> localhost] => (item=openstack-control-plane=enabled)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1 -> localhost] => (item=openstack-control-plane=enabled)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5 -> localhost] => (item=node-role.osism.tech/rook-osd=true)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2 -> localhost] => (item=openstack-control-plane=enabled)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4 -> localhost] => (item=node-role.osism.tech/rook-osd=true)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1 -> localhost] => (item=node-role.osism.tech/network-plane=true)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3 -> localhost] => (item=node-role.osism.tech/rook-osd=true)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0 -> localhost] => (item=node-role.osism.tech/network-plane=true)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2 -> localhost] => (item=node-role.osism.tech/network-plane=true)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1 -> localhost] => (item=node-role.osism.tech/rook-mds=true)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0 -> localhost] => (item=node-role.osism.tech/rook-mds=true)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1 -> localhost] => (item=node-role.osism.tech/rook-mgr=true)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2 -> localhost] => (item=node-role.osism.tech/rook-mds=true)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2 -> localhost] => (item=node-role.osism.tech/rook-mgr=true)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1 -> localhost] => (item=node-role.osism.tech/rook-mon=true)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0 -> localhost] => (item=node-role.osism.tech/rook-mgr=true)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2 -> localhost] => (item=node-role.osism.tech/rook-mon=true)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1 -> localhost] => (item=node-role.osism.tech/rook-rgw=true)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0 -> localhost] => (item=node-role.osism.tech/rook-mon=true)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0 -> localhost] => (item=node-role.osism.tech/rook-rgw=true)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2 -> localhost] => (item=node-role.osism.tech/rook-rgw=true)\u001b[0m",
                                "",
                                "TASK [Manage annotations] ******************************************************",
                                "Tuesday 24 February 2026  00:52:15 +0000 (0:00:13.202)       0:05:00.523 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [Manage taints] ***********************************************************",
                                "Tuesday 24 February 2026  00:52:15 +0000 (0:00:00.533)       0:05:01.057 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;33mtestbed-manager\u001b[0m            : \u001b[0;32mok=21  \u001b[0m \u001b[0;33mchanged=11  \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=50  \u001b[0m \u001b[0;33mchanged=23  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=28  \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=38  \u001b[0m \u001b[0;33mchanged=16  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=25  \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=38  \u001b[0m \u001b[0;33mchanged=16  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=25  \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-3\u001b[0m             : \u001b[0;32mok=16  \u001b[0m \u001b[0;33mchanged=8   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=17  \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-4\u001b[0m             : \u001b[0;32mok=16  \u001b[0m \u001b[0;33mchanged=8   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=17  \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-5\u001b[0m             : \u001b[0;32mok=16  \u001b[0m \u001b[0;33mchanged=8   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=17  \u001b[0m rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Tuesday 24 February 2026  00:52:15 +0000 (0:00:00.386)       0:05:01.443 ****** ",
                                "=============================================================================== ",
                                "k3s_server : Verify that all nodes actually joined (check k3s-init.service if this fails) -- 54.02s",
                                "k3s_server : Enable and check K3s service ------------------------------ 53.12s",
                                "k3s_server_post : Wait for Cilium resources ---------------------------- 42.16s",
                                "Manage labels ---------------------------------------------------------- 13.20s",
                                "kubectl : Install required packages ------------------------------------ 12.52s",
                                "k3s_agent : Manage k3s service ----------------------------------------- 10.20s",
                                "kubectl : Add repository Debian ----------------------------------------- 6.68s",
                                "k3s_download : Download k3s binary x64 ---------------------------------- 5.71s",
                                "k9s : Install k9s packages ---------------------------------------------- 5.43s",
                                "k3s_server_post : Install Cilium ---------------------------------------- 5.13s",
                                "k3s_server : Remove manifests and folders that are only needed for bootstrapping cluster so k3s doesn't auto apply on start --- 3.03s",
                                "k3s_server : Set _kube_vip_bgp_peers fact ------------------------------- 2.83s",
                                "k3s_server : Detect Kubernetes version for label compatibility ---------- 2.80s",
                                "k3s_server : Init cluster inside the transient k3s-init service --------- 2.33s",
                                "k3s_prereq : Enable IPv4 forwarding ------------------------------------- 2.16s",
                                "k3s_download : Download k3s binary armhf -------------------------------- 2.09s",
                                "k3s_server : Create custom resolv.conf for k3s -------------------------- 1.96s",
                                "k3s_prereq : Enable IPv6 router advertisements -------------------------- 1.95s",
                                "k3s_agent : Configure the k3s service ----------------------------------- 1.85s",
                                "k3s_download : Download k3s binary arm64 -------------------------------- 1.83s",
                                "\u001b[32m2026-02-24 00:52:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask feb6adf1-2c00-4649-9a4f-39e67ac78a53 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:52:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ba7d4726-e54a-4b5b-b060-92175ab177a2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:52:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6bfc963b-db63-4b91-bc2e-bbd0ebbc84e3 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:52:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:52:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 404270fa-50db-4a78-8c44-21df48acba39 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:52:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 00:52:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask feb6adf1-2c00-4649-9a4f-39e67ac78a53 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:52:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ba7d4726-e54a-4b5b-b060-92175ab177a2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:52:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6bfc963b-db63-4b91-bc2e-bbd0ebbc84e3 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:52:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:52:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 404270fa-50db-4a78-8c44-21df48acba39 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:52:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 00:52:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask feb6adf1-2c00-4649-9a4f-39e67ac78a53 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:52:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ba7d4726-e54a-4b5b-b060-92175ab177a2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:52:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6bfc963b-db63-4b91-bc2e-bbd0ebbc84e3 is in state SUCCESS\u001b[0m",
                                "\u001b[32m2026-02-24 00:52:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:52:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 404270fa-50db-4a78-8c44-21df48acba39 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:52:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 00:52:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask feb6adf1-2c00-4649-9a4f-39e67ac78a53 is in state SUCCESS\u001b[0m",
                                "\u001b[32m2026-02-24 00:52:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ba7d4726-e54a-4b5b-b060-92175ab177a2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:52:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:52:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 404270fa-50db-4a78-8c44-21df48acba39 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:52:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 00:52:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ba7d4726-e54a-4b5b-b060-92175ab177a2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:52:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:52:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 404270fa-50db-4a78-8c44-21df48acba39 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:52:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 00:52:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ba7d4726-e54a-4b5b-b060-92175ab177a2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:52:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:52:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 404270fa-50db-4a78-8c44-21df48acba39 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:52:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 00:52:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ba7d4726-e54a-4b5b-b060-92175ab177a2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:52:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:52:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 404270fa-50db-4a78-8c44-21df48acba39 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:52:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 00:52:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ba7d4726-e54a-4b5b-b060-92175ab177a2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:52:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:52:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 404270fa-50db-4a78-8c44-21df48acba39 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:52:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 00:52:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ba7d4726-e54a-4b5b-b060-92175ab177a2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:52:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:52:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 404270fa-50db-4a78-8c44-21df48acba39 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:52:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 00:52:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ba7d4726-e54a-4b5b-b060-92175ab177a2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:52:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:52:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 404270fa-50db-4a78-8c44-21df48acba39 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:52:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 00:52:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ba7d4726-e54a-4b5b-b060-92175ab177a2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:52:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:52:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 404270fa-50db-4a78-8c44-21df48acba39 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:52:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 00:52:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ba7d4726-e54a-4b5b-b060-92175ab177a2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:52:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:52:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 404270fa-50db-4a78-8c44-21df48acba39 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:52:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 00:52:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ba7d4726-e54a-4b5b-b060-92175ab177a2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:52:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:52:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 404270fa-50db-4a78-8c44-21df48acba39 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:52:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 00:52:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ba7d4726-e54a-4b5b-b060-92175ab177a2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:52:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:52:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 404270fa-50db-4a78-8c44-21df48acba39 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:52:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 00:53:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ba7d4726-e54a-4b5b-b060-92175ab177a2 is in state SUCCESS\u001b[0m",
                                "",
                                "",
                                "PLAY [Copy kubeconfig to the configuration repository] *************************",
                                "",
                                "TASK [Get kubeconfig file] *****************************************************",
                                "Tuesday 24 February 2026  00:52:20 +0000 (0:00:00.142)       0:00:00.142 ****** ",
                                "\u001b[0;32mok: [testbed-manager -> testbed-node-0(192.168.16.10)]\u001b[0m",
                                "",
                                "TASK [Write kubeconfig file] ***************************************************",
                                "Tuesday 24 February 2026  00:52:21 +0000 (0:00:00.665)       0:00:00.808 ****** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [Change server address in the kubeconfig file] ****************************",
                                "Tuesday 24 February 2026  00:52:22 +0000 (0:00:00.956)       0:00:01.764 ****** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;33mtestbed-manager\u001b[0m            : \u001b[0;32mok=3   \u001b[0m \u001b[0;33mchanged=2   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Tuesday 24 February 2026  00:52:22 +0000 (0:00:00.368)       0:00:02.133 ****** ",
                                "=============================================================================== ",
                                "Write kubeconfig file --------------------------------------------------- 0.96s",
                                "Get kubeconfig file ----------------------------------------------------- 0.67s",
                                "Change server address in the kubeconfig file ---------------------------- 0.37s",
                                "",
                                "",
                                "PLAY [Prepare kubeconfig file] *************************************************",
                                "",
                                "TASK [Get home directory of operator user] *************************************",
                                "Tuesday 24 February 2026  00:52:19 +0000 (0:00:00.119)       0:00:00.119 ****** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [Create .kube directory] **************************************************",
                                "Tuesday 24 February 2026  00:52:20 +0000 (0:00:00.474)       0:00:00.594 ****** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [Get kubeconfig file] *****************************************************",
                                "Tuesday 24 February 2026  00:52:20 +0000 (0:00:00.530)       0:00:01.125 ****** ",
                                "\u001b[0;32mok: [testbed-manager -> testbed-node-0(192.168.16.10)]\u001b[0m",
                                "",
                                "TASK [Write kubeconfig file] ***************************************************",
                                "Tuesday 24 February 2026  00:52:21 +0000 (0:00:00.646)       0:00:01.772 ****** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [Change server address in the kubeconfig] *********************************",
                                "Tuesday 24 February 2026  00:52:22 +0000 (0:00:01.278)       0:00:03.050 ****** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [Make kubeconfig available for use inside the manager service] ************",
                                "Tuesday 24 February 2026  00:52:23 +0000 (0:00:00.454)       0:00:03.505 ****** ",
                                "\u001b[0;33mchanged: [testbed-manager -> localhost]\u001b[0m",
                                "",
                                "TASK [Change server address in the kubeconfig inside the manager service] ******",
                                "Tuesday 24 February 2026  00:52:24 +0000 (0:00:01.527)       0:00:05.032 ****** ",
                                "\u001b[0;33mchanged: [testbed-manager -> localhost]\u001b[0m",
                                "",
                                "TASK [Set KUBECONFIG environment variable] *************************************",
                                "Tuesday 24 February 2026  00:52:25 +0000 (0:00:00.782)       0:00:05.815 ****** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [Enable kubectl command line completion] **********************************",
                                "Tuesday 24 February 2026  00:52:25 +0000 (0:00:00.359)       0:00:06.174 ****** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;33mtestbed-manager\u001b[0m            : \u001b[0;32mok=9   \u001b[0m \u001b[0;33mchanged=4   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Tuesday 24 February 2026  00:52:26 +0000 (0:00:00.340)       0:00:06.515 ****** ",
                                "=============================================================================== ",
                                "Make kubeconfig available for use inside the manager service ------------ 1.53s",
                                "Write kubeconfig file --------------------------------------------------- 1.28s",
                                "Change server address in the kubeconfig inside the manager service ------ 0.78s",
                                "Get kubeconfig file ----------------------------------------------------- 0.65s",
                                "Create .kube directory -------------------------------------------------- 0.53s",
                                "Get home directory of operator user ------------------------------------- 0.47s",
                                "Change server address in the kubeconfig --------------------------------- 0.45s",
                                "Set KUBECONFIG environment variable ------------------------------------- 0.36s",
                                "Enable kubectl command line completion ---------------------------------- 0.34s",
                                "",
                                "",
                                "PLAY [Group hosts based on configuration] **************************************",
                                "",
                                "TASK [Group hosts based on Kolla action] ***************************************",
                                "Tuesday 24 February 2026  00:50:43 +0000 (0:00:00.140)       0:00:00.140 ****** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [Group hosts based on enabled services] ***********************************",
                                "Tuesday 24 February 2026  00:50:43 +0000 (0:00:00.849)       0:00:00.989 ****** ",
                                "\u001b[0;32mok: [testbed-node-3] => (item=enable_ovn_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item=enable_ovn_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item=enable_ovn_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0] => (item=enable_ovn_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item=enable_ovn_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item=enable_ovn_True)\u001b[0m",
                                "",
                                "PLAY [Apply role ovn-controller] ***********************************************",
                                "",
                                "TASK [ovn-controller : include_tasks] ******************************************",
                                "Tuesday 24 February 2026  00:50:44 +0000 (0:00:01.006)       0:00:01.995 ****** ",
                                "\u001b[0;36mincluded: /ansible/roles/ovn-controller/tasks/deploy.yml for testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [ovn-controller : Ensuring config directories exist] **********************",
                                "Tuesday 24 February 2026  00:50:46 +0000 (0:00:01.030)       0:00:03.026 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-controller:2024.2', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-controller:2024.2', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-controller:2024.2', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-controller:2024.2', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-controller:2024.2', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-controller:2024.2', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "",
                                "TASK [ovn-controller : Copying over config.json files for services] ************",
                                "Tuesday 24 February 2026  00:50:47 +0000 (0:00:01.504)       0:00:04.530 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-controller:2024.2', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-controller:2024.2', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-controller:2024.2', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-controller:2024.2', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-controller:2024.2', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-controller:2024.2', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "",
                                "TASK [ovn-controller : Ensuring systemd override directory exists] *************",
                                "Tuesday 24 February 2026  00:50:49 +0000 (0:00:01.684)       0:00:06.215 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-controller:2024.2', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-controller:2024.2', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-controller:2024.2', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-controller:2024.2', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-controller:2024.2', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-controller:2024.2', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "",
                                "TASK [ovn-controller : Copying over systemd override] **************************",
                                "Tuesday 24 February 2026  00:50:50 +0000 (0:00:01.372)       0:00:07.587 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-controller:2024.2', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-controller:2024.2', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-controller:2024.2', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-controller:2024.2', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-controller:2024.2', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-controller:2024.2', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "",
                                "TASK [ovn-controller : Check ovn-controller containers] ************************",
                                "Tuesday 24 February 2026  00:50:52 +0000 (0:00:01.445)       0:00:09.033 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-controller:2024.2', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-controller:2024.2', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-controller:2024.2', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-controller:2024.2', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-controller:2024.2', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ovn-controller', 'value': {'container_name': 'ovn_controller', 'group': 'ovn-controller', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-controller:2024.2', 'volumes': ['/etc/kolla/ovn-controller/:/var/lib/kolla/config_files/:ro', '/run/openvswitch:/run/openvswitch:shared', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "",
                                "TASK [ovn-controller : Create br-int bridge on OpenvSwitch] ********************",
                                "Tuesday 24 February 2026  00:50:53 +0000 (0:00:01.520)       0:00:10.553 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ovn-controller : Configure OVN in OVSDB] *********************************",
                                "Tuesday 24 February 2026  00:50:56 +0000 (0:00:02.732)       0:00:13.285 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'name': 'ovn-encap-ip', 'value': '192.168.16.13'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'name': 'ovn-encap-ip', 'value': '192.168.16.14'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'name': 'ovn-encap-ip', 'value': '192.168.16.12'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'name': 'ovn-encap-ip', 'value': '192.168.16.15'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'name': 'ovn-encap-ip', 'value': '192.168.16.11'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'name': 'ovn-encap-ip', 'value': '192.168.16.10'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'name': 'ovn-encap-type', 'value': 'geneve'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'name': 'ovn-encap-type', 'value': 'geneve'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'name': 'ovn-encap-type', 'value': 'geneve'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'name': 'ovn-encap-type', 'value': 'geneve'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'name': 'ovn-encap-type', 'value': 'geneve'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'name': 'ovn-encap-type', 'value': 'geneve'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'name': 'ovn-remote', 'value': 'tcp:192.168.16.10:6642,tcp:192.168.16.11:6642,tcp:192.168.16.12:6642'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'name': 'ovn-remote', 'value': 'tcp:192.168.16.10:6642,tcp:192.168.16.11:6642,tcp:192.168.16.12:6642'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'name': 'ovn-remote', 'value': 'tcp:192.168.16.10:6642,tcp:192.168.16.11:6642,tcp:192.168.16.12:6642'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'name': 'ovn-remote', 'value': 'tcp:192.168.16.10:6642,tcp:192.168.16.11:6642,tcp:192.168.16.12:6642'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'name': 'ovn-remote', 'value': 'tcp:192.168.16.10:6642,tcp:192.168.16.11:6642,tcp:192.168.16.12:6642'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'name': 'ovn-remote', 'value': 'tcp:192.168.16.10:6642,tcp:192.168.16.11:6642,tcp:192.168.16.12:6642'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'name': 'ovn-remote-probe-interval', 'value': '60000'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'name': 'ovn-remote-probe-interval', 'value': '60000'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'name': 'ovn-remote-probe-interval', 'value': '60000'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'name': 'ovn-remote-probe-interval', 'value': '60000'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'name': 'ovn-remote-probe-interval', 'value': '60000'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'name': 'ovn-remote-probe-interval', 'value': '60000'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'name': 'ovn-openflow-probe-interval', 'value': '60'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'name': 'ovn-openflow-probe-interval', 'value': '60'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'name': 'ovn-openflow-probe-interval', 'value': '60'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'name': 'ovn-openflow-probe-interval', 'value': '60'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'name': 'ovn-openflow-probe-interval', 'value': '60'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'name': 'ovn-openflow-probe-interval', 'value': '60'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'name': 'ovn-monitor-all', 'value': False})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'name': 'ovn-monitor-all', 'value': False})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'name': 'ovn-monitor-all', 'value': False})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'name': 'ovn-monitor-all', 'value': False})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'name': 'ovn-monitor-all', 'value': False})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'name': 'ovn-monitor-all', 'value': False})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item={'name': 'ovn-bridge-mappings', 'value': 'physnet1:br-ex', 'state': 'absent'})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item={'name': 'ovn-bridge-mappings', 'value': 'physnet1:br-ex', 'state': 'absent'})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item={'name': 'ovn-bridge-mappings', 'value': 'physnet1:br-ex', 'state': 'absent'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'name': 'ovn-bridge-mappings', 'value': 'physnet1:br-ex', 'state': 'present'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'name': 'ovn-bridge-mappings', 'value': 'physnet1:br-ex', 'state': 'present'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'name': 'ovn-bridge-mappings', 'value': 'physnet1:br-ex', 'state': 'present'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'name': 'ovn-chassis-mac-mappings', 'value': 'physnet1:52:54:00:2f:fa:44', 'state': 'present'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'name': 'ovn-chassis-mac-mappings', 'value': 'physnet1:52:54:00:71:3a:c3', 'state': 'present'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'name': 'ovn-chassis-mac-mappings', 'value': 'physnet1:52:54:00:89:18:56', 'state': 'present'})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item={'name': 'ovn-chassis-mac-mappings', 'value': 'physnet1:52:54:00:29:4a:9b', 'state': 'absent'})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item={'name': 'ovn-chassis-mac-mappings', 'value': 'physnet1:52:54:00:33:12:50', 'state': 'absent'})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0] => (item={'name': 'ovn-chassis-mac-mappings', 'value': 'physnet1:52:54:00:52:c1:40', 'state': 'absent'})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item={'name': 'ovn-cms-options', 'value': '', 'state': 'absent'})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item={'name': 'ovn-cms-options', 'value': '', 'state': 'absent'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'name': 'ovn-cms-options', 'value': 'enable-chassis-as-gw,availability-zones=nova', 'state': 'present'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'name': 'ovn-cms-options', 'value': 'enable-chassis-as-gw,availability-zones=nova', 'state': 'present'})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item={'name': 'ovn-cms-options', 'value': '', 'state': 'absent'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'name': 'ovn-cms-options', 'value': 'enable-chassis-as-gw,availability-zones=nova', 'state': 'present'})\u001b[0m",
                                "",
                                "TASK [ovn-controller : Flush handlers] *****************************************",
                                "Tuesday 24 February 2026  00:51:18 +0000 (0:00:22.571)       0:00:35.857 ****** ",
                                "",
                                "TASK [ovn-controller : Flush handlers] *****************************************",
                                "Tuesday 24 February 2026  00:51:18 +0000 (0:00:00.063)       0:00:35.921 ****** ",
                                "",
                                "TASK [ovn-controller : Flush handlers] *****************************************",
                                "Tuesday 24 February 2026  00:51:18 +0000 (0:00:00.063)       0:00:35.984 ****** ",
                                "",
                                "TASK [ovn-controller : Flush handlers] *****************************************",
                                "Tuesday 24 February 2026  00:51:19 +0000 (0:00:00.060)       0:00:36.045 ****** ",
                                "",
                                "TASK [ovn-controller : Flush handlers] *****************************************",
                                "Tuesday 24 February 2026  00:51:19 +0000 (0:00:00.065)       0:00:36.111 ****** ",
                                "",
                                "TASK [ovn-controller : Flush handlers] *****************************************",
                                "Tuesday 24 February 2026  00:51:19 +0000 (0:00:00.062)       0:00:36.173 ****** ",
                                "",
                                "RUNNING HANDLER [ovn-controller : Reload systemd config] ***********************",
                                "Tuesday 24 February 2026  00:51:19 +0000 (0:00:00.062)       0:00:36.236 ****** ",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ovn-controller : Restart ovn-controller container] ************",
                                "Tuesday 24 February 2026  00:51:20 +0000 (0:00:01.561)       0:00:37.798 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "PLAY [Apply role ovn-db] *******************************************************",
                                "",
                                "TASK [ovn-db : include_tasks] **************************************************",
                                "Tuesday 24 February 2026  00:51:46 +0000 (0:00:25.870)       0:01:03.668 ****** ",
                                "\u001b[0;36mincluded: /ansible/roles/ovn-db/tasks/deploy.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [ovn-db : include_tasks] **************************************************",
                                "Tuesday 24 February 2026  00:51:47 +0000 (0:00:00.583)       0:01:04.251 ****** ",
                                "\u001b[0;36mincluded: /ansible/roles/ovn-db/tasks/lookup_cluster.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [ovn-db : Checking for any existing OVN DB container volumes] *************",
                                "Tuesday 24 February 2026  00:51:47 +0000 (0:00:00.696)       0:01:04.947 ****** ",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ovn-db : Divide hosts by their OVN NB volume availability] ***************",
                                "Tuesday 24 February 2026  00:51:48 +0000 (0:00:00.975)       0:01:05.923 ****** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ovn-db : Divide hosts by their OVN SB volume availability] ***************",
                                "Tuesday 24 February 2026  00:51:49 +0000 (0:00:00.462)       0:01:06.385 ****** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ovn-db : Establish whether the OVN NB cluster has already existed] *******",
                                "Tuesday 24 February 2026  00:51:49 +0000 (0:00:00.419)       0:01:06.805 ****** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ovn-db : Establish whether the OVN SB cluster has already existed] *******",
                                "Tuesday 24 February 2026  00:51:50 +0000 (0:00:00.597)       0:01:07.403 ****** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ovn-db : Check if running on all OVN NB DB hosts] ************************",
                                "Tuesday 24 February 2026  00:51:51 +0000 (0:00:00.667)       0:01:08.070 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ovn-db : Check OVN NB service port liveness] *****************************",
                                "Tuesday 24 February 2026  00:51:51 +0000 (0:00:00.298)       0:01:08.369 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ovn-db : Divide hosts by their OVN NB service port liveness] *************",
                                "Tuesday 24 February 2026  00:51:51 +0000 (0:00:00.270)       0:01:08.640 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ovn-db : Get OVN NB database information] ********************************",
                                "Tuesday 24 February 2026  00:51:51 +0000 (0:00:00.259)       0:01:08.899 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ovn-db : Divide hosts by their OVN NB leader/follower role] **************",
                                "Tuesday 24 February 2026  00:51:52 +0000 (0:00:00.417)       0:01:09.317 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ovn-db : Fail on existing OVN NB cluster with no leader] *****************",
                                "Tuesday 24 February 2026  00:51:52 +0000 (0:00:00.300)       0:01:09.617 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ovn-db : Check if running on all OVN SB DB hosts] ************************",
                                "Tuesday 24 February 2026  00:51:52 +0000 (0:00:00.262)       0:01:09.879 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ovn-db : Check OVN SB service port liveness] *****************************",
                                "Tuesday 24 February 2026  00:51:53 +0000 (0:00:00.306)       0:01:10.185 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ovn-db : Divide hosts by their OVN SB service port liveness] *************",
                                "Tuesday 24 February 2026  00:51:53 +0000 (0:00:00.442)       0:01:10.628 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ovn-db : Get OVN SB database information] ********************************",
                                "Tuesday 24 February 2026  00:51:53 +0000 (0:00:00.244)       0:01:10.872 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ovn-db : Divide hosts by their OVN SB leader/follower role] **************",
                                "Tuesday 24 February 2026  00:51:54 +0000 (0:00:00.269)       0:01:11.142 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ovn-db : Fail on existing OVN SB cluster with no leader] *****************",
                                "Tuesday 24 February 2026  00:51:54 +0000 (0:00:00.269)       0:01:11.411 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ovn-db : include_tasks] **************************************************",
                                "Tuesday 24 February 2026  00:51:54 +0000 (0:00:00.260)       0:01:11.672 ****** ",
                                "\u001b[0;36mincluded: /ansible/roles/ovn-db/tasks/bootstrap-initial.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [ovn-db : Set bootstrap args fact for NB (new cluster)] *******************",
                                "Tuesday 24 February 2026  00:51:55 +0000 (0:00:00.688)       0:01:12.360 ****** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ovn-db : Set bootstrap args fact for SB (new cluster)] *******************",
                                "Tuesday 24 February 2026  00:51:55 +0000 (0:00:00.382)       0:01:12.743 ****** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ovn-db : Check NB cluster status] ****************************************",
                                "Tuesday 24 February 2026  00:51:56 +0000 (0:00:00.361)       0:01:13.104 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ovn-db : Check SB cluster status] ****************************************",
                                "Tuesday 24 February 2026  00:51:56 +0000 (0:00:00.467)       0:01:13.571 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ovn-db : Remove an old node with the same ip address as the new node in NB DB] ***",
                                "Tuesday 24 February 2026  00:51:56 +0000 (0:00:00.340)       0:01:13.911 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ovn-db : Remove an old node with the same ip address as the new node in SB DB] ***",
                                "Tuesday 24 February 2026  00:51:57 +0000 (0:00:00.339)       0:01:14.250 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ovn-db : Set bootstrap args fact for NB (new member)] ********************",
                                "Tuesday 24 February 2026  00:51:57 +0000 (0:00:00.275)       0:01:14.526 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ovn-db : Set bootstrap args fact for SB (new member)] ********************",
                                "Tuesday 24 February 2026  00:51:57 +0000 (0:00:00.463)       0:01:14.989 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ovn-db : Ensuring config directories exist] ******************************",
                                "Tuesday 24 February 2026  00:51:58 +0000 (0:00:00.346)       0:01:15.336 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ovn-northd', 'value': {'container_name': 'ovn_northd', 'group': 'ovn-northd', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-northd:2024.2', 'volumes': ['/etc/kolla/ovn-northd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ovn-northd', 'value': {'container_name': 'ovn_northd', 'group': 'ovn-northd', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-northd:2024.2', 'volumes': ['/etc/kolla/ovn-northd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ovn-northd', 'value': {'container_name': 'ovn_northd', 'group': 'ovn-northd', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-northd:2024.2', 'volumes': ['/etc/kolla/ovn-northd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ovn-nb-db', 'value': {'container_name': 'ovn_nb_db', 'group': 'ovn-nb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-nb-db-server:2024.2', 'volumes': ['/etc/kolla/ovn-nb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_nb_db:/var/lib/openvswitch/ovn-nb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ovn-nb-db', 'value': {'container_name': 'ovn_nb_db', 'group': 'ovn-nb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-nb-db-server:2024.2', 'volumes': ['/etc/kolla/ovn-nb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_nb_db:/var/lib/openvswitch/ovn-nb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ovn-nb-db', 'value': {'container_name': 'ovn_nb_db', 'group': 'ovn-nb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-nb-db-server:2024.2', 'volumes': ['/etc/kolla/ovn-nb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_nb_db:/var/lib/openvswitch/ovn-nb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ovn-sb-db', 'value': {'container_name': 'ovn_sb_db', 'group': 'ovn-sb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-sb-db-server:2024.2', 'volumes': ['/etc/kolla/ovn-sb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_sb_db:/var/lib/openvswitch/ovn-sb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ovn-sb-db', 'value': {'container_name': 'ovn_sb_db', 'group': 'ovn-sb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-sb-db-server:2024.2', 'volumes': ['/etc/kolla/ovn-sb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_sb_db:/var/lib/openvswitch/ovn-sb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ovn-sb-db', 'value': {'container_name': 'ovn_sb_db', 'group': 'ovn-sb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-sb-db-server:2024.2', 'volumes': ['/etc/kolla/ovn-sb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_sb_db:/var/lib/openvswitch/ovn-sb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "",
                                "TASK [ovn-db : Copying over config.json files for services] ********************",
                                "Tuesday 24 February 2026  00:52:00 +0000 (0:00:01.830)       0:01:17.167 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ovn-northd', 'value': {'container_name': 'ovn_northd', 'group': 'ovn-northd', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-northd:2024.2', 'volumes': ['/etc/kolla/ovn-northd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ovn-northd', 'value': {'container_name': 'ovn_northd', 'group': 'ovn-northd', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-northd:2024.2', 'volumes': ['/etc/kolla/ovn-northd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ovn-northd', 'value': {'container_name': 'ovn_northd', 'group': 'ovn-northd', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-northd:2024.2', 'volumes': ['/etc/kolla/ovn-northd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ovn-nb-db', 'value': {'container_name': 'ovn_nb_db', 'group': 'ovn-nb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-nb-db-server:2024.2', 'volumes': ['/etc/kolla/ovn-nb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_nb_db:/var/lib/openvswitch/ovn-nb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ovn-nb-db', 'value': {'container_name': 'ovn_nb_db', 'group': 'ovn-nb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-nb-db-server:2024.2', 'volumes': ['/etc/kolla/ovn-nb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_nb_db:/var/lib/openvswitch/ovn-nb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ovn-nb-db', 'value': {'container_name': 'ovn_nb_db', 'group': 'ovn-nb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-nb-db-server:2024.2', 'volumes': ['/etc/kolla/ovn-nb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_nb_db:/var/lib/openvswitch/ovn-nb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ovn-sb-db', 'value': {'container_name': 'ovn_sb_db', 'group': 'ovn-sb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-sb-db-server:2024.2', 'volumes': ['/etc/kolla/ovn-sb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_sb_db:/var/lib/openvswitch/ovn-sb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ovn-sb-db', 'value': {'container_name': 'ovn_sb_db', 'group': 'ovn-sb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-sb-db-server:2024.2', 'volumes': ['/etc/kolla/ovn-sb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_sb_db:/var/lib/openvswitch/ovn-sb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ovn-sb-db', 'value': {'container_name': 'ovn_sb_db', 'group': 'ovn-sb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-sb-db-server:2024.2', 'volumes': ['/etc/kolla/ovn-sb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_sb_db:/var/lib/openvswitch/ovn-sb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "",
                                "TASK [ovn-db : Check ovn containers] *******************************************",
                                "Tuesday 24 February 2026  00:52:05 +0000 (0:00:05.458)       0:01:22.625 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ovn-northd', 'value': {'container_name': 'ovn_northd', 'group': 'ovn-northd', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-northd:2024.2', 'volumes': ['/etc/kolla/ovn-northd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ovn-northd', 'value': {'container_name': 'ovn_northd', 'group': 'ovn-northd', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-northd:2024.2', 'volumes': ['/etc/kolla/ovn-northd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ovn-northd', 'value': {'container_name': 'ovn_northd', 'group': 'ovn-northd', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-northd:2024.2', 'volumes': ['/etc/kolla/ovn-northd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ovn-nb-db', 'value': {'container_name': 'ovn_nb_db', 'group': 'ovn-nb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-nb-db-server:2024.2', 'volumes': ['/etc/kolla/ovn-nb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_nb_db:/var/lib/openvswitch/ovn-nb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ovn-nb-db', 'value': {'container_name': 'ovn_nb_db', 'group': 'ovn-nb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-nb-db-server:2024.2', 'volumes': ['/etc/kolla/ovn-nb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_nb_db:/var/lib/openvswitch/ovn-nb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ovn-nb-db', 'value': {'container_name': 'ovn_nb_db', 'group': 'ovn-nb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-nb-db-server:2024.2', 'volumes': ['/etc/kolla/ovn-nb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_nb_db:/var/lib/openvswitch/ovn-nb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ovn-sb-db', 'value': {'container_name': 'ovn_sb_db', 'group': 'ovn-sb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-sb-db-server:2024.2', 'volumes': ['/etc/kolla/ovn-sb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_sb_db:/var/lib/openvswitch/ovn-sb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ovn-sb-db', 'value': {'container_name': 'ovn_sb_db', 'group': 'ovn-sb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-sb-db-server:2024.2', 'volumes': ['/etc/kolla/ovn-sb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_sb_db:/var/lib/openvswitch/ovn-sb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ovn-sb-db', 'value': {'container_name': 'ovn_sb_db', 'group': 'ovn-sb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-sb-db-server:2024.2', 'volumes': ['/etc/kolla/ovn-sb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_sb_db:/var/lib/openvswitch/ovn-sb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "",
                                "TASK [ovn-db : Flush handlers] *************************************************",
                                "Tuesday 24 February 2026  00:52:08 +0000 (0:00:02.544)       0:01:25.169 ****** ",
                                "",
                                "TASK [ovn-db : Flush handlers] *************************************************",
                                "Tuesday 24 February 2026  00:52:08 +0000 (0:00:00.089)       0:01:25.259 ****** ",
                                "",
                                "TASK [ovn-db : Flush handlers] *************************************************",
                                "Tuesday 24 February 2026  00:52:08 +0000 (0:00:00.081)       0:01:25.340 ****** ",
                                "",
                                "RUNNING HANDLER [ovn-db : Restart ovn-nb-db container] *************************",
                                "Tuesday 24 February 2026  00:52:08 +0000 (0:00:00.092)       0:01:25.433 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ovn-db : Restart ovn-sb-db container] *************************",
                                "Tuesday 24 February 2026  00:52:15 +0000 (0:00:07.022)       0:01:32.455 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ovn-db : Restart ovn-northd container] ************************",
                                "Tuesday 24 February 2026  00:52:18 +0000 (0:00:03.035)       0:01:35.490 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ovn-db : Wait for leader election] ***************************************",
                                "Tuesday 24 February 2026  00:52:21 +0000 (0:00:02.988)       0:01:38.478 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ovn-db : Get OVN_Northbound cluster leader] ******************************",
                                "Tuesday 24 February 2026  00:52:21 +0000 (0:00:00.101)       0:01:38.580 ****** ",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ovn-db : Configure OVN NB connection settings] ***************************",
                                "Tuesday 24 February 2026  00:52:22 +0000 (0:00:00.753)       0:01:39.334 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ovn-db : Get OVN_Southbound cluster leader] ******************************",
                                "Tuesday 24 February 2026  00:52:22 +0000 (0:00:00.651)       0:01:39.985 ****** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ovn-db : Configure OVN SB connection settings] ***************************",
                                "Tuesday 24 February 2026  00:52:23 +0000 (0:00:00.889)       0:01:40.874 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ovn-db : Wait for ovn-nb-db] *********************************************",
                                "Tuesday 24 February 2026  00:52:24 +0000 (0:00:00.734)       0:01:41.609 ****** ",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ovn-db : Wait for ovn-sb-db] *********************************************",
                                "Tuesday 24 February 2026  00:52:25 +0000 (0:00:00.919)       0:01:42.529 ****** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ovn-db : Unset bootstrap args fact] **************************************",
                                "Tuesday 24 February 2026  00:52:26 +0000 (0:00:00.720)       0:01:43.249 ****** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ovn-db : Ensuring config directories exist] ******************************",
                                "Tuesday 24 February 2026  00:52:26 +0000 (0:00:00.280)       0:01:43.530 ****** ",
                                "\u001b[0;32mok: [testbed-node-0] => (item={'key': 'ovn-northd', 'value': {'container_name': 'ovn_northd', 'group': 'ovn-northd', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-northd:2024.2', 'volumes': ['/etc/kolla/ovn-northd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item={'key': 'ovn-northd', 'value': {'container_name': 'ovn_northd', 'group': 'ovn-northd', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-northd:2024.2', 'volumes': ['/etc/kolla/ovn-northd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item={'key': 'ovn-northd', 'value': {'container_name': 'ovn_northd', 'group': 'ovn-northd', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-northd:2024.2', 'volumes': ['/etc/kolla/ovn-northd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0] => (item={'key': 'ovn-nb-db', 'value': {'container_name': 'ovn_nb_db', 'group': 'ovn-nb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-nb-db-server:2024.2', 'volumes': ['/etc/kolla/ovn-nb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_nb_db:/var/lib/openvswitch/ovn-nb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item={'key': 'ovn-nb-db', 'value': {'container_name': 'ovn_nb_db', 'group': 'ovn-nb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-nb-db-server:2024.2', 'volumes': ['/etc/kolla/ovn-nb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_nb_db:/var/lib/openvswitch/ovn-nb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item={'key': 'ovn-nb-db', 'value': {'container_name': 'ovn_nb_db', 'group': 'ovn-nb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-nb-db-server:2024.2', 'volumes': ['/etc/kolla/ovn-nb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_nb_db:/var/lib/openvswitch/ovn-nb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0] => (item={'key': 'ovn-sb-db', 'value': {'container_name': 'ovn_sb_db', 'group': 'ovn-sb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-sb-db-server:2024.2', 'volumes': ['/etc/kolla/ovn-sb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_sb_db:/var/lib/openvswitch/ovn-sb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item={'key': 'ovn-sb-db', 'value': {'container_name': 'ovn_sb_db', 'group': 'ovn-sb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-sb-db-server:2024.2', 'volumes': ['/etc/kolla/ovn-sb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_sb_db:/var/lib/openvswitch/ovn-sb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item={'key': 'ovn-sb-db', 'value': {'container_name': 'ovn_sb_db', 'group': 'ovn-sb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-sb-db-server:2024.2', 'volumes': ['/etc/kolla/ovn-sb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_sb_db:/var/lib/openvswitch/ovn-sb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "",
                                "TASK [ovn-db : Copying over config.json files for services] ********************",
                                "Tuesday 24 February 2026  00:52:27 +0000 (0:00:01.430)       0:01:44.960 ****** ",
                                "\u001b[0;32mok: [testbed-node-0] => (item={'key': 'ovn-northd', 'value': {'container_name': 'ovn_northd', 'group': 'ovn-northd', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-northd:2024.2', 'volumes': ['/etc/kolla/ovn-northd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item={'key': 'ovn-northd', 'value': {'container_name': 'ovn_northd', 'group': 'ovn-northd', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-northd:2024.2', 'volumes': ['/etc/kolla/ovn-northd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item={'key': 'ovn-northd', 'value': {'container_name': 'ovn_northd', 'group': 'ovn-northd', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-northd:2024.2', 'volumes': ['/etc/kolla/ovn-northd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0] => (item={'key': 'ovn-nb-db', 'value': {'container_name': 'ovn_nb_db', 'group': 'ovn-nb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-nb-db-server:2024.2', 'volumes': ['/etc/kolla/ovn-nb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_nb_db:/var/lib/openvswitch/ovn-nb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ovn-nb-db', 'value': {'container_name': 'ovn_nb_db', 'group': 'ovn-nb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-nb-db-server:2024.2', 'volumes': ['/etc/kolla/ovn-nb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_nb_db:/var/lib/openvswitch/ovn-nb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ovn-nb-db', 'value': {'container_name': 'ovn_nb_db', 'group': 'ovn-nb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-nb-db-server:2024.2', 'volumes': ['/etc/kolla/ovn-nb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_nb_db:/var/lib/openvswitch/ovn-nb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0] => (item={'key': 'ovn-sb-db', 'value': {'container_name': 'ovn_sb_db', 'group': 'ovn-sb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-sb-db-server:2024.2', 'volumes': ['/etc/kolla/ovn-sb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_sb_db:/var/lib/openvswitch/ovn-sb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ovn-sb-db', 'value': {'container_name': 'ovn_sb_db', 'group': 'ovn-sb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-sb-db-server:2024.2', 'volumes': ['/etc/kolla/ovn-sb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_sb_db:/var/lib/openvswitch/ovn-sb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ovn-sb-db', 'value': {'container_name': 'ovn_sb_db', 'group': 'ovn-sb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-sb-db-server:2024.2', 'volumes': ['/etc/kolla/ovn-sb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_sb_db:/var/lib/openvswitch/ovn-sb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "",
                                "TASK [ovn-db : Check ovn containers] *******************************************",
                                "Tuesday 24 February 2026  00:52:31 +0000 (0:00:03.749)       0:01:48.710 ****** ",
                                "\u001b[0;32mok: [testbed-node-0] => (item={'key': 'ovn-northd', 'value': {'container_name': 'ovn_northd', 'group': 'ovn-northd', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-northd:2024.2', 'volumes': ['/etc/kolla/ovn-northd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item={'key': 'ovn-northd', 'value': {'container_name': 'ovn_northd', 'group': 'ovn-northd', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-northd:2024.2', 'volumes': ['/etc/kolla/ovn-northd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item={'key': 'ovn-northd', 'value': {'container_name': 'ovn_northd', 'group': 'ovn-northd', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-northd:2024.2', 'volumes': ['/etc/kolla/ovn-northd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0] => (item={'key': 'ovn-nb-db', 'value': {'container_name': 'ovn_nb_db', 'group': 'ovn-nb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-nb-db-server:2024.2', 'volumes': ['/etc/kolla/ovn-nb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_nb_db:/var/lib/openvswitch/ovn-nb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ovn-nb-db', 'value': {'container_name': 'ovn_nb_db', 'group': 'ovn-nb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-nb-db-server:2024.2', 'volumes': ['/etc/kolla/ovn-nb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_nb_db:/var/lib/openvswitch/ovn-nb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ovn-nb-db', 'value': {'container_name': 'ovn_nb_db', 'group': 'ovn-nb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-nb-db-server:2024.2', 'volumes': ['/etc/kolla/ovn-nb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_nb_db:/var/lib/openvswitch/ovn-nb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0] => (item={'key': 'ovn-sb-db', 'value': {'container_name': 'ovn_sb_db', 'group': 'ovn-sb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-sb-db-server:2024.2', 'volumes': ['/etc/kolla/ovn-sb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_sb_db:/var/lib/openvswitch/ovn-sb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ovn-sb-db', 'value': {'container_name': 'ovn_sb_db', 'group': 'ovn-sb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-sb-db-server:2024.2', 'volumes': ['/etc/kolla/ovn-sb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_sb_db:/var/lib/openvswitch/ovn-sb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ovn-sb-db', 'value': {'container_name': 'ovn_sb_db', 'group': 'ovn-sb-db', 'enabled': True, 'image': 'registry.osism.tech/kolla/ovn-sb-db-server:2024.2', 'volumes': ['/etc/kolla/ovn-sb-db/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'ovn_sb_db:/var/lib/openvswitch/ovn-sb/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "",
                                "TASK [ovn-db : Flush handlers] *************************************************",
                                "Tuesday 24 February 2026  00:52:34 +0000 (0:00:03.027)       0:01:51.737 ****** ",
                                "",
                                "TASK [ovn-db : Flush handlers] *************************************************",
                                "Tuesday 24 February 2026  00:52:34 +0000 (0:00:00.105)       0:01:51.843 ****** ",
                                "",
                                "TASK [ovn-db : Flush handlers] *************************************************",
                                "Tuesday 24 February 2026  00:52:34 +0000 (0:00:00.069)       0:01:51.912 ****** ",
                                "",
                                "RUNNING HANDLER [ovn-db : Restart ovn-nb-db container] *************************",
                                "Tuesday 24 February 2026  00:52:34 +0000 (0:00:00.066)       0:01:51.979 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ovn-db : Restart ovn-sb-db container] *************************",
                                "Tuesday 24 February 2026  00:52:41 +0000 (0:00:06.527)       0:01:58.508 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ovn-db : Restart ovn-northd container] ************************",
                                "Tuesday 24 February 2026  00:52:47 +0000 (0:00:06.167)       0:02:04.675 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ovn-db : Wait for leader election] ***************************************",
                                "Tuesday 24 February 2026  00:52:54 +0000 (0:00:06.811)       0:02:11.486 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ovn-db : Get OVN_Northbound cluster leader] ******************************",
                                "Tuesday 24 February 2026  00:52:54 +0000 (0:00:00.124)       0:02:11.611 ****** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ovn-db : Configure OVN NB connection settings] ***************************",
                                "Tuesday 24 February 2026  00:52:55 +0000 (0:00:00.849)       0:02:12.461 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ovn-db : Get OVN_Southbound cluster leader] ******************************",
                                "Tuesday 24 February 2026  00:52:56 +0000 (0:00:00.804)       0:02:13.265 ****** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ovn-db : Configure OVN SB connection settings] ***************************",
                                "Tuesday 24 February 2026  00:52:57 +0000 (0:00:00.881)       0:02:14.147 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ovn-db : Wait for ovn-nb-db] *********************************************",
                                "Tuesday 24 February 2026  00:52:57 +0000 (0:00:00.666)       0:02:14.813 ****** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ovn-db : Wait for ovn-sb-db] *********************************************",
                                "Tuesday 24 February 2026  00:52:58 +0000 (0:00:00.807)       0:02:15.621 ****** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=44  \u001b[0m \u001b[0;33mchanged=18  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=20  \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=43  \u001b[0m \u001b[0;33mchanged=19  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=22  \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=43  \u001b[0m \u001b[0;33mchanged=19  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=22  \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-3\u001b[0m             : \u001b[0;32mok=12  \u001b[0m \u001b[0;33mchanged=8   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-4\u001b[0m             : \u001b[0;32mok=12  \u001b[0m \u001b[0;33mchanged=8   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-5\u001b[0m             : \u001b[0;32mok=12  \u001b[0m \u001b[0;33mchanged=8   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Tuesday 24 February 2026  00:52:59 +0000 (0:00:01.001)       0:02:16.623 ****** ",
                                "=============================================================================== ",
                                "ovn-controller : Restart ovn-controller container ---------------------- 25.87s",
                                "ovn-controller : Configure OVN in OVSDB -------------------------------- 22.57s",
                                "ovn-db : Restart ovn-nb-db container ----------------------------------- 13.55s",
                                "ovn-db : Restart ovn-northd container ----------------------------------- 9.80s",
                                "ovn-db : Restart ovn-sb-db container ------------------------------------ 9.20s",
                                "ovn-db : Copying over config.json files for services -------------------- 5.46s",
                                "ovn-db : Copying over config.json files for services -------------------- 3.75s",
                                "ovn-db : Check ovn containers ------------------------------------------- 3.03s",
                                "ovn-controller : Create br-int bridge on OpenvSwitch -------------------- 2.73s",
                                "ovn-db : Check ovn containers ------------------------------------------- 2.54s",
                                "ovn-db : Ensuring config directories exist ------------------------------ 1.83s",
                                "ovn-controller : Copying over config.json files for services ------------ 1.68s",
                                "ovn-controller : Reload systemd config ---------------------------------- 1.56s",
                                "ovn-controller : Check ovn-controller containers ------------------------ 1.52s",
                                "ovn-controller : Ensuring config directories exist ---------------------- 1.50s",
                                "ovn-controller : Copying over systemd override -------------------------- 1.45s",
                                "ovn-db : Ensuring config directories exist ------------------------------ 1.43s",
                                "ovn-controller : Ensuring systemd override directory exists ------------- 1.37s",
                                "ovn-controller : include_tasks ------------------------------------------ 1.03s",
                                "Group hosts based on enabled services ----------------------------------- 1.01s",
                                "\u001b[32m2026-02-24 00:53:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:53:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 404270fa-50db-4a78-8c44-21df48acba39 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:53:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 00:53:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:53:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 404270fa-50db-4a78-8c44-21df48acba39 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:53:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 00:53:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:53:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 404270fa-50db-4a78-8c44-21df48acba39 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:53:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 00:53:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:53:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 404270fa-50db-4a78-8c44-21df48acba39 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:53:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 00:53:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:53:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 404270fa-50db-4a78-8c44-21df48acba39 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:53:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 00:53:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:53:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 404270fa-50db-4a78-8c44-21df48acba39 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:53:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 00:53:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:53:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 404270fa-50db-4a78-8c44-21df48acba39 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:53:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 00:53:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:53:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 404270fa-50db-4a78-8c44-21df48acba39 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:53:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 00:53:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:53:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 404270fa-50db-4a78-8c44-21df48acba39 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:53:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 00:53:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:53:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 404270fa-50db-4a78-8c44-21df48acba39 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:53:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 00:53:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:53:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 404270fa-50db-4a78-8c44-21df48acba39 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:53:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 00:53:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:53:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 404270fa-50db-4a78-8c44-21df48acba39 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:53:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 00:53:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:53:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 404270fa-50db-4a78-8c44-21df48acba39 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:53:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 00:53:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:53:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 404270fa-50db-4a78-8c44-21df48acba39 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:53:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 00:53:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:53:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 404270fa-50db-4a78-8c44-21df48acba39 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:53:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 00:53:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:53:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 404270fa-50db-4a78-8c44-21df48acba39 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:53:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 00:53:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:53:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 404270fa-50db-4a78-8c44-21df48acba39 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:53:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 00:53:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:53:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 404270fa-50db-4a78-8c44-21df48acba39 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:53:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 00:53:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:53:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 404270fa-50db-4a78-8c44-21df48acba39 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:53:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 00:53:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:53:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 404270fa-50db-4a78-8c44-21df48acba39 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:53:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 00:54:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:54:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 404270fa-50db-4a78-8c44-21df48acba39 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:54:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 00:54:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:54:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 404270fa-50db-4a78-8c44-21df48acba39 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:54:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 00:54:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:54:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 404270fa-50db-4a78-8c44-21df48acba39 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:54:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 00:54:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:54:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 404270fa-50db-4a78-8c44-21df48acba39 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:54:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 00:54:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:54:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 404270fa-50db-4a78-8c44-21df48acba39 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:54:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 00:54:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:54:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 404270fa-50db-4a78-8c44-21df48acba39 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:54:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 00:54:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:54:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 404270fa-50db-4a78-8c44-21df48acba39 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:54:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 00:54:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:54:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 404270fa-50db-4a78-8c44-21df48acba39 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:54:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 00:54:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:54:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 404270fa-50db-4a78-8c44-21df48acba39 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:54:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 00:54:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:54:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 404270fa-50db-4a78-8c44-21df48acba39 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:54:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 00:54:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:54:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 404270fa-50db-4a78-8c44-21df48acba39 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:54:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 00:54:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:54:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 404270fa-50db-4a78-8c44-21df48acba39 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:54:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 00:54:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:54:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 404270fa-50db-4a78-8c44-21df48acba39 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:54:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 00:54:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:54:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 404270fa-50db-4a78-8c44-21df48acba39 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:54:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 00:54:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:54:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 404270fa-50db-4a78-8c44-21df48acba39 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:54:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 00:54:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:54:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 404270fa-50db-4a78-8c44-21df48acba39 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:54:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 00:54:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:54:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 404270fa-50db-4a78-8c44-21df48acba39 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:54:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 00:54:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:54:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 404270fa-50db-4a78-8c44-21df48acba39 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:54:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 00:54:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:54:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 404270fa-50db-4a78-8c44-21df48acba39 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:54:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 00:54:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:54:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 404270fa-50db-4a78-8c44-21df48acba39 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:54:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 00:55:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:55:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 404270fa-50db-4a78-8c44-21df48acba39 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:55:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 00:55:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:55:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 404270fa-50db-4a78-8c44-21df48acba39 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:55:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 00:55:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:55:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 404270fa-50db-4a78-8c44-21df48acba39 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:55:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 00:55:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:55:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 404270fa-50db-4a78-8c44-21df48acba39 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:55:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 00:55:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:55:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 404270fa-50db-4a78-8c44-21df48acba39 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:55:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 00:55:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:55:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 404270fa-50db-4a78-8c44-21df48acba39 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:55:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 00:55:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:55:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 404270fa-50db-4a78-8c44-21df48acba39 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:55:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 00:55:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:55:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 404270fa-50db-4a78-8c44-21df48acba39 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:55:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 00:55:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:55:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 404270fa-50db-4a78-8c44-21df48acba39 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:55:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 00:55:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:55:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 404270fa-50db-4a78-8c44-21df48acba39 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:55:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 00:55:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:55:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 404270fa-50db-4a78-8c44-21df48acba39 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:55:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 00:55:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:55:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 404270fa-50db-4a78-8c44-21df48acba39 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:55:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 00:55:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:55:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 404270fa-50db-4a78-8c44-21df48acba39 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:55:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 00:55:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:55:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 404270fa-50db-4a78-8c44-21df48acba39 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:55:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 00:55:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:55:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 404270fa-50db-4a78-8c44-21df48acba39 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:55:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 00:55:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:55:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 404270fa-50db-4a78-8c44-21df48acba39 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:55:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 00:55:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:55:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 404270fa-50db-4a78-8c44-21df48acba39 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:55:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 00:55:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:55:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 404270fa-50db-4a78-8c44-21df48acba39 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:55:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 00:55:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:55:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 404270fa-50db-4a78-8c44-21df48acba39 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:55:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 00:56:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:56:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 404270fa-50db-4a78-8c44-21df48acba39 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:56:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 00:56:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:56:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 404270fa-50db-4a78-8c44-21df48acba39 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:56:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 00:56:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:56:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 404270fa-50db-4a78-8c44-21df48acba39 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:56:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 00:56:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:56:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 404270fa-50db-4a78-8c44-21df48acba39 is in state SUCCESS\u001b[0m",
                                "",
                                "",
                                "PLAY [Group hosts based on configuration] **************************************",
                                "",
                                "TASK [Group hosts based on Kolla action] ***************************************",
                                "Tuesday 24 February 2026  00:49:35 +0000 (0:00:00.368)       0:00:00.368 ****** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [Group hosts based on enabled services] ***********************************",
                                "Tuesday 24 February 2026  00:49:35 +0000 (0:00:00.452)       0:00:00.820 ****** ",
                                "\u001b[0;32mok: [testbed-node-0] => (item=enable_loadbalancer_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item=enable_loadbalancer_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item=enable_loadbalancer_True)\u001b[0m",
                                "",
                                "PLAY [Apply role loadbalancer] *************************************************",
                                "",
                                "TASK [loadbalancer : include_tasks] ********************************************",
                                "Tuesday 24 February 2026  00:49:36 +0000 (0:00:00.426)       0:00:01.247 ****** ",
                                "\u001b[0;36mincluded: /ansible/roles/loadbalancer/tasks/deploy.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [loadbalancer : Check IPv6 support] ***************************************",
                                "Tuesday 24 February 2026  00:49:37 +0000 (0:00:00.988)       0:00:02.235 ****** ",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [Setting sysctl values] ***************************************************",
                                "Tuesday 24 February 2026  00:49:38 +0000 (0:00:01.009)       0:00:03.244 ****** ",
                                "\u001b[0;36mincluded: sysctl for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [sysctl : Check IPv6 support] *********************************************",
                                "Tuesday 24 February 2026  00:49:39 +0000 (0:00:01.173)       0:00:04.418 ****** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [sysctl : Setting sysctl values] ******************************************",
                                "Tuesday 24 February 2026  00:49:40 +0000 (0:00:00.666)       0:00:05.084 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'name': 'net.ipv6.ip_nonlocal_bind', 'value': 1})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'name': 'net.ipv6.ip_nonlocal_bind', 'value': 1})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'name': 'net.ipv4.ip_nonlocal_bind', 'value': 1})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'name': 'net.ipv4.ip_nonlocal_bind', 'value': 1})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0] => (item={'name': 'net.ipv4.tcp_retries2', 'value': 'KOLLA_UNSET'})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item={'name': 'net.ipv4.tcp_retries2', 'value': 'KOLLA_UNSET'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'name': 'net.ipv6.ip_nonlocal_bind', 'value': 1})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'name': 'net.unix.max_dgram_qlen', 'value': 128})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'name': 'net.unix.max_dgram_qlen', 'value': 128})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'name': 'net.ipv4.ip_nonlocal_bind', 'value': 1})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item={'name': 'net.ipv4.tcp_retries2', 'value': 'KOLLA_UNSET'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'name': 'net.unix.max_dgram_qlen', 'value': 128})\u001b[0m",
                                "",
                                "TASK [module-load : Load modules] **********************************************",
                                "Tuesday 24 February 2026  00:49:43 +0000 (0:00:03.747)       0:00:08.832 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=ip_vs)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=ip_vs)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=ip_vs)\u001b[0m",
                                "",
                                "TASK [module-load : Persist modules via modules-load.d] ************************",
                                "Tuesday 24 February 2026  00:49:44 +0000 (0:00:00.832)       0:00:09.665 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=ip_vs)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=ip_vs)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=ip_vs)\u001b[0m",
                                "",
                                "TASK [module-load : Drop module persistence] ***********************************",
                                "Tuesday 24 February 2026  00:49:46 +0000 (0:00:01.418)       0:00:11.083 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=ip_vs) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=ip_vs) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=ip_vs) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [loadbalancer : Ensuring config directories exist] ************************",
                                "Tuesday 24 February 2026  00:49:47 +0000 (0:00:01.199)       0:00:12.282 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:61313'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:61313'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:61313'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2024.2', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2024.2', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2024.2', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}})\u001b[0m",
                                "",
                                "TASK [loadbalancer : Ensuring haproxy service config subdir exists] ************",
                                "Tuesday 24 February 2026  00:49:49 +0000 (0:00:02.010)       0:00:14.292 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [loadbalancer : Ensuring proxysql service config subdirectories exist] ****",
                                "Tuesday 24 February 2026  00:49:50 +0000 (0:00:01.202)       0:00:15.495 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=users)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=users)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=users)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=rules)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=rules)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=rules)\u001b[0m",
                                "",
                                "TASK [loadbalancer : Ensuring keepalived checks subdir exists] *****************",
                                "Tuesday 24 February 2026  00:49:52 +0000 (0:00:01.637)       0:00:17.133 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [loadbalancer : Remove mariadb.cfg if proxysql enabled] *******************",
                                "Tuesday 24 February 2026  00:49:53 +0000 (0:00:01.733)       0:00:18.866 ****** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [loadbalancer : Removing checks for services which are disabled] **********",
                                "Tuesday 24 February 2026  00:49:56 +0000 (0:00:02.364)       0:00:21.231 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:61313'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2024.2', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'haproxy-ssh', 'value': {'container_name': 'haproxy_ssh', 'group': 'loadbalancer', 'enabled': False, 'image': 'registry.osism.tech/kolla/haproxy-ssh:2024.2', 'volumes': ['/etc/kolla/haproxy-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', '__omit_place_holder__852810438d06097b6927d4e594558f870baad287', '__omit_place_holder__852810438d06097b6927d4e594558f870baad287'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 2985'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:61313'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:61313'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2024.2', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2024.2', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'haproxy-ssh', 'value': {'container_name': 'haproxy_ssh', 'group': 'loadbalancer', 'enabled': False, 'image': 'registry.osism.tech/kolla/haproxy-ssh:2024.2', 'volumes': ['/etc/kolla/haproxy-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', '__omit_place_holder__852810438d06097b6927d4e594558f870baad287', '__omit_place_holder__852810438d06097b6927d4e594558f870baad287'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 2985'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'haproxy-ssh', 'value': {'container_name': 'haproxy_ssh', 'group': 'loadbalancer', 'enabled': False, 'image': 'registry.osism.tech/kolla/haproxy-ssh:2024.2', 'volumes': ['/etc/kolla/haproxy-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', '__omit_place_holder__852810438d06097b6927d4e594558f870baad287', '__omit_place_holder__852810438d06097b6927d4e594558f870baad287'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 2985'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [loadbalancer : Copying checks for services which are enabled] ************",
                                "Tuesday 24 February 2026  00:49:57 +0000 (0:00:00.748)       0:00:21.979 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:61313'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:61313'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:61313'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2024.2', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'haproxy-ssh', 'value': {'container_name': 'haproxy_ssh', 'group': 'loadbalancer', 'enabled': False, 'image': 'registry.osism.tech/kolla/haproxy-ssh:2024.2', 'volumes': ['/etc/kolla/haproxy-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', '__omit_place_holder__852810438d06097b6927d4e594558f870baad287', '__omit_place_holder__852810438d06097b6927d4e594558f870baad287'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 2985'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2024.2', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'haproxy-ssh', 'value': {'container_name': 'haproxy_ssh', 'group': 'loadbalancer', 'enabled': False, 'image': 'registry.osism.tech/kolla/haproxy-ssh:2024.2', 'volumes': ['/etc/kolla/haproxy-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', '__omit_place_holder__852810438d06097b6927d4e594558f870baad287', '__omit_place_holder__852810438d06097b6927d4e594558f870baad287'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 2985'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2024.2', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'haproxy-ssh', 'value': {'container_name': 'haproxy_ssh', 'group': 'loadbalancer', 'enabled': False, 'image': 'registry.osism.tech/kolla/haproxy-ssh:2024.2', 'volumes': ['/etc/kolla/haproxy-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', '__omit_place_holder__852810438d06097b6927d4e594558f870baad287', '__omit_place_holder__852810438d06097b6927d4e594558f870baad287'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 2985'], 'timeout': '30'}}}) \u001b[0m",
                                "",
                                "TASK [loadbalancer : Copying over config.json files for services] **************",
                                "Tuesday 24 February 2026  00:49:59 +0000 (0:00:02.501)       0:00:24.481 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:61313'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:61313'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:61313'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2024.2', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2024.2', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2024.2', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}})\u001b[0m",
                                "",
                                "TASK [loadbalancer : Copying over haproxy.cfg] *********************************",
                                "Tuesday 24 February 2026  00:50:02 +0000 (0:00:02.886)       0:00:27.367 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=/ansible/roles/loadbalancer/templates/haproxy/haproxy_main.cfg.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=/ansible/roles/loadbalancer/templates/haproxy/haproxy_main.cfg.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=/ansible/roles/loadbalancer/templates/haproxy/haproxy_main.cfg.j2)\u001b[0m",
                                "",
                                "TASK [loadbalancer : Copying over proxysql config] *****************************",
                                "Tuesday 24 February 2026  00:50:05 +0000 (0:00:03.147)       0:00:30.514 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=/ansible/roles/loadbalancer/templates/proxysql/proxysql.yaml.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=/ansible/roles/loadbalancer/templates/proxysql/proxysql.yaml.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=/ansible/roles/loadbalancer/templates/proxysql/proxysql.yaml.j2)\u001b[0m",
                                "",
                                "TASK [loadbalancer : Copying over haproxy single external frontend config] *****",
                                "Tuesday 24 February 2026  00:50:10 +0000 (0:00:04.951)       0:00:35.466 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [loadbalancer : Copying over custom haproxy services configuration] *******",
                                "Tuesday 24 February 2026  00:50:11 +0000 (0:00:00.930)       0:00:36.396 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=/opt/configuration/environments/kolla/files/overlays/haproxy/services.d/haproxy.cfg)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=/opt/configuration/environments/kolla/files/overlays/haproxy/services.d/haproxy.cfg)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=/opt/configuration/environments/kolla/files/overlays/haproxy/services.d/haproxy.cfg)\u001b[0m",
                                "",
                                "TASK [loadbalancer : Copying over keepalived.conf] *****************************",
                                "Tuesday 24 February 2026  00:50:14 +0000 (0:00:02.905)       0:00:39.302 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=/ansible/roles/loadbalancer/templates/keepalived/keepalived.conf.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=/ansible/roles/loadbalancer/templates/keepalived/keepalived.conf.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=/ansible/roles/loadbalancer/templates/keepalived/keepalived.conf.j2)\u001b[0m",
                                "",
                                "TASK [loadbalancer : Copying over haproxy.pem] *********************************",
                                "Tuesday 24 February 2026  00:50:17 +0000 (0:00:02.861)       0:00:42.163 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=haproxy.pem)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=haproxy.pem)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=haproxy.pem)\u001b[0m",
                                "",
                                "TASK [loadbalancer : Copying over haproxy-internal.pem] ************************",
                                "Tuesday 24 February 2026  00:50:18 +0000 (0:00:01.532)       0:00:43.696 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=haproxy-internal.pem)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=haproxy-internal.pem)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=haproxy-internal.pem)\u001b[0m",
                                "",
                                "TASK [loadbalancer : include_tasks] ********************************************",
                                "Tuesday 24 February 2026  00:50:20 +0000 (0:00:01.388)       0:00:45.085 ****** ",
                                "\u001b[0;36mincluded: /ansible/roles/loadbalancer/tasks/copy-certs.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [service-cert-copy : loadbalancer | Copying over extra CA certificates] ***",
                                "Tuesday 24 February 2026  00:50:20 +0000 (0:00:00.805)       0:00:45.891 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:61313'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:61313'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:61313'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2024.2', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2024.2', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2024.2', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}})\u001b[0m",
                                "",
                                "TASK [service-cert-copy : loadbalancer | Copying over backend internal TLS certificate] ***",
                                "Tuesday 24 February 2026  00:50:24 +0000 (0:00:03.628)       0:00:49.519 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:61313'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:61313'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2024.2', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2024.2', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:61313'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2024.2', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [service-cert-copy : loadbalancer | Copying over backend internal TLS key] ***",
                                "Tuesday 24 February 2026  00:50:25 +0000 (0:00:01.014)       0:00:50.533 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:61313'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2024.2', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:61313'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2024.2', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:61313'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2024.2', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [service-cert-copy : mariadb | Copying over extra CA certificates] ********",
                                "Tuesday 24 February 2026  00:50:26 +0000 (0:00:00.845)       0:00:51.379 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:61313'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2024.2', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:61313'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2024.2', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:61313'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2024.2', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [service-cert-copy : mariadb | Copying over backend internal TLS certificate] ***",
                                "Tuesday 24 February 2026  00:50:27 +0000 (0:00:00.848)       0:00:52.227 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:61313'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2024.2', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:61313'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2024.2', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:61313'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2024.2', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [service-cert-copy : mariadb | Copying over backend internal TLS key] *****",
                                "Tuesday 24 February 2026  00:50:28 +0000 (0:00:00.909)       0:00:53.137 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:61313'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2024.2', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:61313'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2024.2', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:61313'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2024.2', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [service-cert-copy : proxysql | Copying over extra CA certificates] *******",
                                "Tuesday 24 February 2026  00:50:29 +0000 (0:00:01.005)       0:00:54.142 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:61313'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2024.2', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:61313'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2024.2', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:61313'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2024.2', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [service-cert-copy : proxysql | Copying over backend internal TLS certificate] ***",
                                "Tuesday 24 February 2026  00:50:31 +0000 (0:00:02.049)       0:00:56.192 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:61313'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2024.2', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:61313'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2024.2', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:61313'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2024.2', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [service-cert-copy : proxysql | Copying over backend internal TLS key] ****",
                                "Tuesday 24 February 2026  00:50:32 +0000 (0:00:00.950)       0:00:57.143 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:61313'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2024.2', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:61313'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2024.2', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:61313'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2024.2', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [loadbalancer : Copying over haproxy start script] ************************",
                                "Tuesday 24 February 2026  00:50:33 +0000 (0:00:00.774)       0:00:57.917 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=/ansible/roles/loadbalancer/templates/haproxy/haproxy_run.sh.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=/ansible/roles/loadbalancer/templates/haproxy/haproxy_run.sh.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=/ansible/roles/loadbalancer/templates/haproxy/haproxy_run.sh.j2)\u001b[0m",
                                "",
                                "TASK [loadbalancer : Copying over proxysql start script] ***********************",
                                "Tuesday 24 February 2026  00:50:34 +0000 (0:00:01.797)       0:00:59.714 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=/ansible/roles/loadbalancer/templates/proxysql/proxysql_run.sh.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=/ansible/roles/loadbalancer/templates/proxysql/proxysql_run.sh.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=/ansible/roles/loadbalancer/templates/proxysql/proxysql_run.sh.j2)\u001b[0m",
                                "",
                                "TASK [loadbalancer : Copying files for haproxy-ssh] ****************************",
                                "Tuesday 24 February 2026  00:50:36 +0000 (0:00:01.728)       0:01:01.443 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'src': 'haproxy-ssh/sshd_config.j2', 'dest': 'sshd_config'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'src': 'haproxy-ssh/sshd_config.j2', 'dest': 'sshd_config'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'src': 'haproxy-ssh/sshd_config.j2', 'dest': 'sshd_config'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'src': 'haproxy-ssh/id_rsa.pub', 'dest': 'id_rsa.pub'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'src': 'haproxy-ssh/id_rsa.pub', 'dest': 'id_rsa.pub'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'src': 'haproxy-ssh/id_rsa.pub', 'dest': 'id_rsa.pub'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [loadbalancer : Check loadbalancer containers] ****************************",
                                "Tuesday 24 February 2026  00:50:37 +0000 (0:00:01.067)       0:01:02.511 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:61313'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:61313'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'haproxy', 'value': {'container_name': 'haproxy', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/haproxy:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/haproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'letsencrypt_certificates:/etc/haproxy/certificates'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:61313'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2024.2', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2024.2', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'proxysql', 'value': {'container_name': 'proxysql', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/proxysql:2024.2', 'privileged': False, 'volumes': ['/etc/kolla/proxysql/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'proxysql:/var/lib/proxysql/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen proxysql 6032'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'keepalived', 'value': {'container_name': 'keepalived', 'group': 'loadbalancer', 'enabled': True, 'image': 'registry.osism.tech/kolla/keepalived:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/keepalived/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'haproxy_socket:/var/lib/kolla/haproxy/', 'proxysql_socket:/var/lib/kolla/proxysql/'], 'dimensions': {}}})\u001b[0m",
                                "",
                                "TASK [include_role : aodh] *****************************************************",
                                "Tuesday 24 February 2026  00:50:40 +0000 (0:00:02.955)       0:01:05.467 ****** ",
                                "\u001b[0;36mincluded: aodh for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [haproxy-config : Copying over aodh haproxy config] ***********************",
                                "Tuesday 24 February 2026  00:50:41 +0000 (0:00:00.551)       0:01:06.019 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'aodh-api', 'value': {'container_name': 'aodh_api', 'group': 'aodh-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/aodh-api:2024.2', 'volumes': ['/etc/kolla/aodh-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'aodh:/var/lib/aodh/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8042'], 'timeout': '30'}, 'haproxy': {'aodh_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8042', 'listen_port': '8042'}, 'aodh_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8042', 'listen_port': '8042'}}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'aodh-evaluator', 'value': {'container_name': 'aodh_evaluator', 'group': 'aodh-evaluator', 'enabled': True, 'image': 'registry.osism.tech/kolla/aodh-evaluator:2024.2', 'volumes': ['/etc/kolla/aodh-evaluator/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-evaluator 3306'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'aodh-listener', 'value': {'container_name': 'aodh_listener', 'group': 'aodh-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/aodh-listener:2024.2', 'volumes': ['/etc/kolla/aodh-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-listener 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'aodh-notifier', 'value': {'container_name': 'aodh_notifier', 'group': 'aodh-notifier', 'enabled': True, 'image': 'registry.osism.tech/kolla/aodh-notifier:2024.2', 'volumes': ['/etc/kolla/aodh-notifier/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-notifier 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'aodh-api', 'value': {'container_name': 'aodh_api', 'group': 'aodh-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/aodh-api:2024.2', 'volumes': ['/etc/kolla/aodh-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'aodh:/var/lib/aodh/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8042'], 'timeout': '30'}, 'haproxy': {'aodh_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8042', 'listen_port': '8042'}, 'aodh_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8042', 'listen_port': '8042'}}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'aodh-evaluator', 'value': {'container_name': 'aodh_evaluator', 'group': 'aodh-evaluator', 'enabled': True, 'image': 'registry.osism.tech/kolla/aodh-evaluator:2024.2', 'volumes': ['/etc/kolla/aodh-evaluator/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-evaluator 3306'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'aodh-listener', 'value': {'container_name': 'aodh_listener', 'group': 'aodh-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/aodh-listener:2024.2', 'volumes': ['/etc/kolla/aodh-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-listener 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'aodh-notifier', 'value': {'container_name': 'aodh_notifier', 'group': 'aodh-notifier', 'enabled': True, 'image': 'registry.osism.tech/kolla/aodh-notifier:2024.2', 'volumes': ['/etc/kolla/aodh-notifier/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-notifier 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'aodh-api', 'value': {'container_name': 'aodh_api', 'group': 'aodh-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/aodh-api:2024.2', 'volumes': ['/etc/kolla/aodh-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'aodh:/var/lib/aodh/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8042'], 'timeout': '30'}, 'haproxy': {'aodh_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8042', 'listen_port': '8042'}, 'aodh_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8042', 'listen_port': '8042'}}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'aodh-evaluator', 'value': {'container_name': 'aodh_evaluator', 'group': 'aodh-evaluator', 'enabled': True, 'image': 'registry.osism.tech/kolla/aodh-evaluator:2024.2', 'volumes': ['/etc/kolla/aodh-evaluator/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-evaluator 3306'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'aodh-listener', 'value': {'container_name': 'aodh_listener', 'group': 'aodh-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/aodh-listener:2024.2', 'volumes': ['/etc/kolla/aodh-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-listener 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'aodh-notifier', 'value': {'container_name': 'aodh_notifier', 'group': 'aodh-notifier', 'enabled': True, 'image': 'registry.osism.tech/kolla/aodh-notifier:2024.2', 'volumes': ['/etc/kolla/aodh-notifier/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-notifier 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "",
                                "TASK [haproxy-config : Add configuration for aodh when using single external frontend] ***",
                                "Tuesday 24 February 2026  00:50:45 +0000 (0:00:04.644)       0:01:10.663 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'aodh-api', 'value': {'container_name': 'aodh_api', 'group': 'aodh-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/aodh-api:2024.2', 'volumes': ['/etc/kolla/aodh-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'aodh:/var/lib/aodh/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8042'], 'timeout': '30'}, 'haproxy': {'aodh_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8042', 'listen_port': '8042'}, 'aodh_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8042', 'listen_port': '8042'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'aodh-evaluator', 'value': {'container_name': 'aodh_evaluator', 'group': 'aodh-evaluator', 'enabled': True, 'image': 'registry.osism.tech/kolla/aodh-evaluator:2024.2', 'volumes': ['/etc/kolla/aodh-evaluator/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-evaluator 3306'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'aodh-listener', 'value': {'container_name': 'aodh_listener', 'group': 'aodh-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/aodh-listener:2024.2', 'volumes': ['/etc/kolla/aodh-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-listener 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'aodh-notifier', 'value': {'container_name': 'aodh_notifier', 'group': 'aodh-notifier', 'enabled': True, 'image': 'registry.osism.tech/kolla/aodh-notifier:2024.2', 'volumes': ['/etc/kolla/aodh-notifier/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-notifier 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'aodh-api', 'value': {'container_name': 'aodh_api', 'group': 'aodh-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/aodh-api:2024.2', 'volumes': ['/etc/kolla/aodh-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'aodh:/var/lib/aodh/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8042'], 'timeout': '30'}, 'haproxy': {'aodh_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8042', 'listen_port': '8042'}, 'aodh_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8042', 'listen_port': '8042'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'aodh-evaluator', 'value': {'container_name': 'aodh_evaluator', 'group': 'aodh-evaluator', 'enabled': True, 'image': 'registry.osism.tech/kolla/aodh-evaluator:2024.2', 'volumes': ['/etc/kolla/aodh-evaluator/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-evaluator 3306'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'aodh-listener', 'value': {'container_name': 'aodh_listener', 'group': 'aodh-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/aodh-listener:2024.2', 'volumes': ['/etc/kolla/aodh-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-listener 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'aodh-notifier', 'value': {'container_name': 'aodh_notifier', 'group': 'aodh-notifier', 'enabled': True, 'image': 'registry.osism.tech/kolla/aodh-notifier:2024.2', 'volumes': ['/etc/kolla/aodh-notifier/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-notifier 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'aodh-api', 'value': {'container_name': 'aodh_api', 'group': 'aodh-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/aodh-api:2024.2', 'volumes': ['/etc/kolla/aodh-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'aodh:/var/lib/aodh/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8042'], 'timeout': '30'}, 'haproxy': {'aodh_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8042', 'listen_port': '8042'}, 'aodh_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8042', 'listen_port': '8042'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'aodh-evaluator', 'value': {'container_name': 'aodh_evaluator', 'group': 'aodh-evaluator', 'enabled': True, 'image': 'registry.osism.tech/kolla/aodh-evaluator:2024.2', 'volumes': ['/etc/kolla/aodh-evaluator/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-evaluator 3306'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'aodh-listener', 'value': {'container_name': 'aodh_listener', 'group': 'aodh-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/aodh-listener:2024.2', 'volumes': ['/etc/kolla/aodh-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-listener 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'aodh-notifier', 'value': {'container_name': 'aodh_notifier', 'group': 'aodh-notifier', 'enabled': True, 'image': 'registry.osism.tech/kolla/aodh-notifier:2024.2', 'volumes': ['/etc/kolla/aodh-notifier/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port aodh-notifier 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [haproxy-config : Configuring firewall for aodh] **************************",
                                "Tuesday 24 February 2026  00:50:47 +0000 (0:00:01.439)       0:01:12.103 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'aodh_api', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8042', 'listen_port': '8042'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'aodh_api_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8042', 'listen_port': '8042'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'aodh_api', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8042', 'listen_port': '8042'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'aodh_api_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8042', 'listen_port': '8042'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'aodh_api', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8042', 'listen_port': '8042'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'aodh_api_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8042', 'listen_port': '8042'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [proxysql-config : Copying over aodh ProxySQL users config] ***************",
                                "Tuesday 24 February 2026  00:50:48 +0000 (0:00:00.937)       0:01:13.041 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [proxysql-config : Copying over aodh ProxySQL rules config] ***************",
                                "Tuesday 24 February 2026  00:50:49 +0000 (0:00:01.461)       0:01:14.502 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [include_role : barbican] *************************************************",
                                "Tuesday 24 February 2026  00:50:51 +0000 (0:00:01.959)       0:01:16.462 ****** ",
                                "\u001b[0;36mincluded: barbican for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [haproxy-config : Copying over barbican haproxy config] *******************",
                                "Tuesday 24 February 2026  00:50:52 +0000 (0:00:00.744)       0:01:17.206 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-api:2024.2', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-keystone-listener:2024.2', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-worker:2024.2', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-api:2024.2', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-keystone-listener:2024.2', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-worker:2024.2', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-api:2024.2', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-keystone-listener:2024.2', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-worker:2024.2', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "",
                                "TASK [haproxy-config : Add configuration for barbican when using single external frontend] ***",
                                "Tuesday 24 February 2026  00:50:56 +0000 (0:00:03.789)       0:01:20.995 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-api:2024.2', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-keystone-listener:2024.2', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-worker:2024.2', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-api:2024.2', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-keystone-listener:2024.2', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-worker:2024.2', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-api:2024.2', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-keystone-listener:2024.2', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-worker:2024.2', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [haproxy-config : Configuring firewall for barbican] **********************",
                                "Tuesday 24 February 2026  00:50:57 +0000 (0:00:01.468)       0:01:22.464 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'barbican_api', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'barbican_api_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'barbican_api', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'barbican_api_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'barbican_api', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'barbican_api_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [proxysql-config : Copying over barbican ProxySQL users config] ***********",
                                "Tuesday 24 February 2026  00:50:59 +0000 (0:00:02.179)       0:01:24.644 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [proxysql-config : Copying over barbican ProxySQL rules config] ***********",
                                "Tuesday 24 February 2026  00:51:01 +0000 (0:00:01.323)       0:01:25.967 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [include_role : blazar] ***************************************************",
                                "Tuesday 24 February 2026  00:51:02 +0000 (0:00:01.888)       0:01:27.856 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [include_role : ceph-rgw] *************************************************",
                                "Tuesday 24 February 2026  00:51:03 +0000 (0:00:00.287)       0:01:28.143 ****** ",
                                "\u001b[0;36mincluded: ceph-rgw for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [haproxy-config : Copying over ceph-rgw haproxy config] *******************",
                                "Tuesday 24 February 2026  00:51:03 +0000 (0:00:00.706)       0:01:28.849 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ceph-rgw', 'value': {'group': 'all', 'enabled': True, 'haproxy': {'radosgw': {'enabled': True, 'mode': 'http', 'external': False, 'port': '6780', 'custom_member_list': ['server testbed-node-3 192.168.16.13:7480 check inter 2000 rise 2 fall 5', 'server testbed-node-4 192.168.16.14:7480 check inter 2000 rise 2 fall 5', 'server testbed-node-5 192.168.16.15:7480 check inter 2000 rise 2 fall 5']}, 'radosgw_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6780', 'custom_member_list': ['server testbed-node-3 192.168.16.13:7480 check inter 2000 rise 2 fall 5', 'server testbed-node-4 192.168.16.14:7480 check inter 2000 rise 2 fall 5', 'server testbed-node-5 192.168.16.15:7480 check inter 2000 rise 2 fall 5']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ceph-rgw', 'value': {'group': 'all', 'enabled': True, 'haproxy': {'radosgw': {'enabled': True, 'mode': 'http', 'external': False, 'port': '6780', 'custom_member_list': ['server testbed-node-3 192.168.16.13:7480 check inter 2000 rise 2 fall 5', 'server testbed-node-4 192.168.16.14:7480 check inter 2000 rise 2 fall 5', 'server testbed-node-5 192.168.16.15:7480 check inter 2000 rise 2 fall 5']}, 'radosgw_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6780', 'custom_member_list': ['server testbed-node-3 192.168.16.13:7480 check inter 2000 rise 2 fall 5', 'server testbed-node-4 192.168.16.14:7480 check inter 2000 rise 2 fall 5', 'server testbed-node-5 192.168.16.15:7480 check inter 2000 rise 2 fall 5']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ceph-rgw', 'value': {'group': 'all', 'enabled': True, 'haproxy': {'radosgw': {'enabled': True, 'mode': 'http', 'external': False, 'port': '6780', 'custom_member_list': ['server testbed-node-3 192.168.16.13:7480 check inter 2000 rise 2 fall 5', 'server testbed-node-4 192.168.16.14:7480 check inter 2000 rise 2 fall 5', 'server testbed-node-5 192.168.16.15:7480 check inter 2000 rise 2 fall 5']}, 'radosgw_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6780', 'custom_member_list': ['server testbed-node-3 192.168.16.13:7480 check inter 2000 rise 2 fall 5', 'server testbed-node-4 192.168.16.14:7480 check inter 2000 rise 2 fall 5', 'server testbed-node-5 192.168.16.15:7480 check inter 2000 rise 2 fall 5']}}}})\u001b[0m",
                                "",
                                "TASK [haproxy-config : Add configuration for ceph-rgw when using single external frontend] ***",
                                "Tuesday 24 February 2026  00:51:06 +0000 (0:00:02.656)       0:01:31.506 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'ceph-rgw', 'value': {'group': 'all', 'enabled': True, 'haproxy': {'radosgw': {'enabled': True, 'mode': 'http', 'external': False, 'port': '6780', 'custom_member_list': ['server testbed-node-3 192.168.16.13:7480 check inter 2000 rise 2 fall 5', 'server testbed-node-4 192.168.16.14:7480 check inter 2000 rise 2 fall 5', 'server testbed-node-5 192.168.16.15:7480 check inter 2000 rise 2 fall 5']}, 'radosgw_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6780', 'custom_member_list': ['server testbed-node-3 192.168.16.13:7480 check inter 2000 rise 2 fall 5', 'server testbed-node-4 192.168.16.14:7480 check inter 2000 rise 2 fall 5', 'server testbed-node-5 192.168.16.15:7480 check inter 2000 rise 2 fall 5']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'ceph-rgw', 'value': {'group': 'all', 'enabled': True, 'haproxy': {'radosgw': {'enabled': True, 'mode': 'http', 'external': False, 'port': '6780', 'custom_member_list': ['server testbed-node-3 192.168.16.13:7480 check inter 2000 rise 2 fall 5', 'server testbed-node-4 192.168.16.14:7480 check inter 2000 rise 2 fall 5', 'server testbed-node-5 192.168.16.15:7480 check inter 2000 rise 2 fall 5']}, 'radosgw_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6780', 'custom_member_list': ['server testbed-node-3 192.168.16.13:7480 check inter 2000 rise 2 fall 5', 'server testbed-node-4 192.168.16.14:7480 check inter 2000 rise 2 fall 5', 'server testbed-node-5 192.168.16.15:7480 check inter 2000 rise 2 fall 5']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'ceph-rgw', 'value': {'group': 'all', 'enabled': True, 'haproxy': {'radosgw': {'enabled': True, 'mode': 'http', 'external': False, 'port': '6780', 'custom_member_list': ['server testbed-node-3 192.168.16.13:7480 check inter 2000 rise 2 fall 5', 'server testbed-node-4 192.168.16.14:7480 check inter 2000 rise 2 fall 5', 'server testbed-node-5 192.168.16.15:7480 check inter 2000 rise 2 fall 5']}, 'radosgw_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6780', 'custom_member_list': ['server testbed-node-3 192.168.16.13:7480 check inter 2000 rise 2 fall 5', 'server testbed-node-4 192.168.16.14:7480 check inter 2000 rise 2 fall 5', 'server testbed-node-5 192.168.16.15:7480 check inter 2000 rise 2 fall 5']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [haproxy-config : Configuring firewall for ceph-rgw] **********************",
                                "Tuesday 24 February 2026  00:51:08 +0000 (0:00:01.986)       0:01:33.493 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'radosgw', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '6780', 'custom_member_list': ['server testbed-node-3 192.168.16.13:7480 check inter 2000 rise 2 fall 5', 'server testbed-node-4 192.168.16.14:7480 check inter 2000 rise 2 fall 5', 'server testbed-node-5 192.168.16.15:7480 check inter 2000 rise 2 fall 5']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'radosgw_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6780', 'custom_member_list': ['server testbed-node-3 192.168.16.13:7480 check inter 2000 rise 2 fall 5', 'server testbed-node-4 192.168.16.14:7480 check inter 2000 rise 2 fall 5', 'server testbed-node-5 192.168.16.15:7480 check inter 2000 rise 2 fall 5']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'radosgw', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '6780', 'custom_member_list': ['server testbed-node-3 192.168.16.13:7480 check inter 2000 rise 2 fall 5', 'server testbed-node-4 192.168.16.14:7480 check inter 2000 rise 2 fall 5', 'server testbed-node-5 192.168.16.15:7480 check inter 2000 rise 2 fall 5']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'radosgw_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6780', 'custom_member_list': ['server testbed-node-3 192.168.16.13:7480 check inter 2000 rise 2 fall 5', 'server testbed-node-4 192.168.16.14:7480 check inter 2000 rise 2 fall 5', 'server testbed-node-5 192.168.16.15:7480 check inter 2000 rise 2 fall 5']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'radosgw', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '6780', 'custom_member_list': ['server testbed-node-3 192.168.16.13:7480 check inter 2000 rise 2 fall 5', 'server testbed-node-4 192.168.16.14:7480 check inter 2000 rise 2 fall 5', 'server testbed-node-5 192.168.16.15:7480 check inter 2000 rise 2 fall 5']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'radosgw_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6780', 'custom_member_list': ['server testbed-node-3 192.168.16.13:7480 check inter 2000 rise 2 fall 5', 'server testbed-node-4 192.168.16.14:7480 check inter 2000 rise 2 fall 5', 'server testbed-node-5 192.168.16.15:7480 check inter 2000 rise 2 fall 5']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [proxysql-config : Copying over ceph-rgw ProxySQL users config] ***********",
                                "Tuesday 24 February 2026  00:51:10 +0000 (0:00:01.615)       0:01:35.109 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [proxysql-config : Copying over ceph-rgw ProxySQL rules config] ***********",
                                "Tuesday 24 February 2026  00:51:10 +0000 (0:00:00.375)       0:01:35.484 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [include_role : cinder] ***************************************************",
                                "Tuesday 24 February 2026  00:51:11 +0000 (0:00:01.057)       0:01:36.542 ****** ",
                                "\u001b[0;36mincluded: cinder for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [haproxy-config : Copying over cinder haproxy config] *********************",
                                "Tuesday 24 February 2026  00:51:12 +0000 (0:00:00.664)       0:01:37.206 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-api:2024.2', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8776'], 'timeout': '30'}, 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-scheduler:2024.2', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-volume:2024.2', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-api:2024.2', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8776'], 'timeout': '30'}, 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-backup:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-scheduler:2024.2', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-volume:2024.2', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-api:2024.2', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8776'], 'timeout': '30'}, 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-backup:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-scheduler:2024.2', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-volume:2024.2', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-backup:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "",
                                "TASK [haproxy-config : Add configuration for cinder when using single external frontend] ***",
                                "Tuesday 24 February 2026  00:51:15 +0000 (0:00:03.380)       0:01:40.587 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-api:2024.2', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8776'], 'timeout': '30'}, 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-scheduler:2024.2', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-volume:2024.2', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-backup:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-api:2024.2', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8776'], 'timeout': '30'}, 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-scheduler:2024.2', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-volume:2024.2', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-backup:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-api:2024.2', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8776'], 'timeout': '30'}, 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-scheduler:2024.2', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-volume:2024.2', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-backup:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [haproxy-config : Configuring firewall for cinder] ************************",
                                "Tuesday 24 February 2026  00:51:16 +0000 (0:00:00.775)       0:01:41.363 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'cinder_api', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'cinder_api_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'cinder_api', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'cinder_api_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'cinder_api', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'cinder_api_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [proxysql-config : Copying over cinder ProxySQL users config] *************",
                                "Tuesday 24 February 2026  00:51:17 +0000 (0:00:00.841)       0:01:42.204 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [proxysql-config : Copying over cinder ProxySQL rules config] *************",
                                "Tuesday 24 February 2026  00:51:18 +0000 (0:00:01.245)       0:01:43.449 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [include_role : cloudkitty] ***********************************************",
                                "Tuesday 24 February 2026  00:51:20 +0000 (0:00:01.903)       0:01:45.353 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [include_role : cyborg] ***************************************************",
                                "Tuesday 24 February 2026  00:51:20 +0000 (0:00:00.513)       0:01:45.867 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [include_role : designate] ************************************************",
                                "Tuesday 24 February 2026  00:51:21 +0000 (0:00:00.288)       0:01:46.155 ****** ",
                                "\u001b[0;36mincluded: designate for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [haproxy-config : Copying over designate haproxy config] ******************",
                                "Tuesday 24 February 2026  00:51:22 +0000 (0:00:00.792)       0:01:46.947 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-api:2024.2', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-api:2024.2', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-backend-bind9:2024.2', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-backend-bind9:2024.2', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-central:2024.2', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-mdns:2024.2', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-producer:2024.2', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-central:2024.2', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-worker:2024.2', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-sink', 'value': {'container_name': 'designate_sink', 'group': 'designate-sink', 'enabled': False, 'image': 'registry.osism.tech/kolla/designate-sink:2024.2', 'volumes': ['/etc/kolla/designate-sink/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-sink 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-mdns:2024.2', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-producer:2024.2', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-worker:2024.2', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-sink', 'value': {'container_name': 'designate_sink', 'group': 'designate-sink', 'enabled': False, 'image': 'registry.osism.tech/kolla/designate-sink:2024.2', 'volumes': ['/etc/kolla/designate-sink/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-sink 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-api:2024.2', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-backend-bind9:2024.2', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-central:2024.2', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-mdns:2024.2', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-producer:2024.2', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-worker:2024.2', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-sink', 'value': {'container_name': 'designate_sink', 'group': 'designate-sink', 'enabled': False, 'image': 'registry.osism.tech/kolla/designate-sink:2024.2', 'volumes': ['/etc/kolla/designate-sink/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-sink 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "",
                                "TASK [haproxy-config : Add configuration for designate when using single external frontend] ***",
                                "Tuesday 24 February 2026  00:51:26 +0000 (0:00:04.593)       0:01:51.541 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-api:2024.2', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-backend-bind9:2024.2', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-central:2024.2', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-mdns:2024.2', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-producer:2024.2', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-worker:2024.2', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-sink', 'value': {'container_name': 'designate_sink', 'group': 'designate-sink', 'enabled': False, 'image': 'registry.osism.tech/kolla/designate-sink:2024.2', 'volumes': ['/etc/kolla/designate-sink/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-sink 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-api:2024.2', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-backend-bind9:2024.2', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-central:2024.2', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-mdns:2024.2', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-producer:2024.2', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-worker:2024.2', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-sink', 'value': {'container_name': 'designate_sink', 'group': 'designate-sink', 'enabled': False, 'image': 'registry.osism.tech/kolla/designate-sink:2024.2', 'volumes': ['/etc/kolla/designate-sink/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-sink 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-api:2024.2', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-backend-bind9:2024.2', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-central:2024.2', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-mdns:2024.2', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-producer:2024.2', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-worker:2024.2', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-sink', 'value': {'container_name': 'designate_sink', 'group': 'designate-sink', 'enabled': False, 'image': 'registry.osism.tech/kolla/designate-sink:2024.2', 'volumes': ['/etc/kolla/designate-sink/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-sink 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [haproxy-config : Configuring firewall for designate] *********************",
                                "Tuesday 24 February 2026  00:51:27 +0000 (0:00:01.064)       0:01:52.606 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate_api', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate_api_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate_api', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate_api_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate_api', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate_api_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [proxysql-config : Copying over designate ProxySQL users config] **********",
                                "Tuesday 24 February 2026  00:51:28 +0000 (0:00:00.950)       0:01:53.556 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [proxysql-config : Copying over designate ProxySQL rules config] **********",
                                "Tuesday 24 February 2026  00:51:30 +0000 (0:00:01.571)       0:01:55.128 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [include_role : etcd] *****************************************************",
                                "Tuesday 24 February 2026  00:51:32 +0000 (0:00:01.903)       0:01:57.032 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [include_role : glance] ***************************************************",
                                "Tuesday 24 February 2026  00:51:32 +0000 (0:00:00.290)       0:01:57.323 ****** ",
                                "\u001b[0;36mincluded: glance for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [haproxy-config : Copying over glance haproxy config] *********************",
                                "Tuesday 24 February 2026  00:51:33 +0000 (0:00:00.799)       0:01:58.122 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'glance-api', 'value': {'container_name': 'glance_api', 'group': 'glance-api', 'host_in_groups': True, 'enabled': True, 'image': 'registry.osism.tech/kolla/glance-api:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.10,192.168.16.9'}, 'privileged': True, 'volumes': ['/etc/kolla/glance-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'glance:/var/lib/glance/', '', 'kolla_logs:/var/log/kolla/', '', '', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9292'], 'timeout': '30'}, 'haproxy': {'glance_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}, 'glance_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'glance-tls-proxy', 'value': {'container_name': 'glance_tls_proxy', 'group': 'glance-api', 'host_in_groups': True, 'enabled': 'no', 'image': 'registry.osism.tech/kolla/glance-tls-proxy:2024.2', 'volumes': ['/etc/kolla/glance-tls-proxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl -u openstack:password 192.168.16.10:9293'], 'timeout': '30'}, 'haproxy': {'glance_tls_proxy': {'enabled': False, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', ''], 'tls_backend': 'yes'}, 'glance_tls_proxy_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', ''], 'tls_backend': 'yes'}}}}) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'glance-api', 'value': {'container_name': 'glance_api', 'group': 'glance-api', 'host_in_groups': True, 'enabled': True, 'image': 'registry.osism.tech/kolla/glance-api:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.12,192.168.16.9'}, 'privileged': True, 'volumes': ['/etc/kolla/glance-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'glance:/var/lib/glance/', '', 'kolla_logs:/var/log/kolla/', '', '', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9292'], 'timeout': '30'}, 'haproxy': {'glance_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}, 'glance_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'glance-tls-proxy', 'value': {'container_name': 'glance_tls_proxy', 'group': 'glance-api', 'host_in_groups': True, 'enabled': 'no', 'image': 'registry.osism.tech/kolla/glance-tls-proxy:2024.2', 'volumes': ['/etc/kolla/glance-tls-proxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl -u openstack:password 192.168.16.12:9293'], 'timeout': '30'}, 'haproxy': {'glance_tls_proxy': {'enabled': False, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', ''], 'tls_backend': 'yes'}, 'glance_tls_proxy_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', ''], 'tls_backend': 'yes'}}}}) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'glance-api', 'value': {'container_name': 'glance_api', 'group': 'glance-api', 'host_in_groups': True, 'enabled': True, 'image': 'registry.osism.tech/kolla/glance-api:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.11,192.168.16.9'}, 'privileged': True, 'volumes': ['/etc/kolla/glance-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'glance:/var/lib/glance/', '', 'kolla_logs:/var/log/kolla/', '', '', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9292'], 'timeout': '30'}, 'haproxy': {'glance_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}, 'glance_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'glance-tls-proxy', 'value': {'container_name': 'glance_tls_proxy', 'group': 'glance-api', 'host_in_groups': True, 'enabled': 'no', 'image': 'registry.osism.tech/kolla/glance-tls-proxy:2024.2', 'volumes': ['/etc/kolla/glance-tls-proxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl -u openstack:password 192.168.16.11:9293'], 'timeout': '30'}, 'haproxy': {'glance_tls_proxy': {'enabled': False, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', ''], 'tls_backend': 'yes'}, 'glance_tls_proxy_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', ''], 'tls_backend': 'yes'}}}}) \u001b[0m",
                                "",
                                "TASK [haproxy-config : Add configuration for glance when using single external frontend] ***",
                                "Tuesday 24 February 2026  00:51:37 +0000 (0:00:04.542)       0:02:02.664 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'glance-api', 'value': {'container_name': 'glance_api', 'group': 'glance-api', 'host_in_groups': True, 'enabled': True, 'image': 'registry.osism.tech/kolla/glance-api:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.10,192.168.16.9'}, 'privileged': True, 'volumes': ['/etc/kolla/glance-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'glance:/var/lib/glance/', '', 'kolla_logs:/var/log/kolla/', '', '', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9292'], 'timeout': '30'}, 'haproxy': {'glance_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}, 'glance_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'glance-tls-proxy', 'value': {'container_name': 'glance_tls_proxy', 'group': 'glance-api', 'host_in_groups': True, 'enabled': 'no', 'image': 'registry.osism.tech/kolla/glance-tls-proxy:2024.2', 'volumes': ['/etc/kolla/glance-tls-proxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl -u openstack:password 192.168.16.10:9293'], 'timeout': '30'}, 'haproxy': {'glance_tls_proxy': {'enabled': False, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', ''], 'tls_backend': 'yes'}, 'glance_tls_proxy_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', ''], 'tls_backend': 'yes'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'glance-api', 'value': {'container_name': 'glance_api', 'group': 'glance-api', 'host_in_groups': True, 'enabled': True, 'image': 'registry.osism.tech/kolla/glance-api:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.12,192.168.16.9'}, 'privileged': True, 'volumes': ['/etc/kolla/glance-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'glance:/var/lib/glance/', '', 'kolla_logs:/var/log/kolla/', '', '', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9292'], 'timeout': '30'}, 'haproxy': {'glance_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}, 'glance_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'glance-tls-proxy', 'value': {'container_name': 'glance_tls_proxy', 'group': 'glance-api', 'host_in_groups': True, 'enabled': 'no', 'image': 'registry.osism.tech/kolla/glance-tls-proxy:2024.2', 'volumes': ['/etc/kolla/glance-tls-proxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl -u openstack:password 192.168.16.12:9293'], 'timeout': '30'}, 'haproxy': {'glance_tls_proxy': {'enabled': False, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', ''], 'tls_backend': 'yes'}, 'glance_tls_proxy_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', ''], 'tls_backend': 'yes'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'glance-api', 'value': {'container_name': 'glance_api', 'group': 'glance-api', 'host_in_groups': True, 'enabled': True, 'image': 'registry.osism.tech/kolla/glance-api:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.11,192.168.16.9'}, 'privileged': True, 'volumes': ['/etc/kolla/glance-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'glance:/var/lib/glance/', '', 'kolla_logs:/var/log/kolla/', '', '', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9292'], 'timeout': '30'}, 'haproxy': {'glance_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}, 'glance_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'glance-tls-proxy', 'value': {'container_name': 'glance_tls_proxy', 'group': 'glance-api', 'host_in_groups': True, 'enabled': 'no', 'image': 'registry.osism.tech/kolla/glance-tls-proxy:2024.2', 'volumes': ['/etc/kolla/glance-tls-proxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl -u openstack:password 192.168.16.11:9293'], 'timeout': '30'}, 'haproxy': {'glance_tls_proxy': {'enabled': False, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', ''], 'tls_backend': 'yes'}, 'glance_tls_proxy_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5 ssl verify required ca-file ca-certificates.crt', ''], 'tls_backend': 'yes'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [haproxy-config : Configuring firewall for glance] ************************",
                                "Tuesday 24 February 2026  00:51:40 +0000 (0:00:02.663)       0:02:05.327 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'glance_api', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'glance_api_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'glance_api', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'glance_api_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'glance_api', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'glance_api_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [proxysql-config : Copying over glance ProxySQL users config] *************",
                                "Tuesday 24 February 2026  00:51:43 +0000 (0:00:02.677)       0:02:08.005 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [proxysql-config : Copying over glance ProxySQL rules config] *************",
                                "Tuesday 24 February 2026  00:51:44 +0000 (0:00:01.305)       0:02:09.311 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [include_role : gnocchi] **************************************************",
                                "Tuesday 24 February 2026  00:51:46 +0000 (0:00:01.919)       0:02:11.230 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [include_role : grafana] **************************************************",
                                "Tuesday 24 February 2026  00:51:46 +0000 (0:00:00.431)       0:02:11.661 ****** ",
                                "\u001b[0;36mincluded: grafana for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [haproxy-config : Copying over grafana haproxy config] ********************",
                                "Tuesday 24 February 2026  00:51:47 +0000 (0:00:00.834)       0:02:12.496 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'grafana', 'value': {'container_name': 'grafana', 'group': 'grafana', 'enabled': True, 'image': 'registry.osism.tech/kolla/grafana:2024.2', 'volumes': ['/etc/kolla/grafana/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'grafana_server': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000'}, 'grafana_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'grafana', 'value': {'container_name': 'grafana', 'group': 'grafana', 'enabled': True, 'image': 'registry.osism.tech/kolla/grafana:2024.2', 'volumes': ['/etc/kolla/grafana/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'grafana_server': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000'}, 'grafana_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'grafana', 'value': {'container_name': 'grafana', 'group': 'grafana', 'enabled': True, 'image': 'registry.osism.tech/kolla/grafana:2024.2', 'volumes': ['/etc/kolla/grafana/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'grafana_server': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000'}, 'grafana_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000'}}}})\u001b[0m",
                                "",
                                "TASK [haproxy-config : Add configuration for grafana when using single external frontend] ***",
                                "Tuesday 24 February 2026  00:51:51 +0000 (0:00:04.209)       0:02:16.705 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'grafana', 'value': {'container_name': 'grafana', 'group': 'grafana', 'enabled': True, 'image': 'registry.osism.tech/kolla/grafana:2024.2', 'volumes': ['/etc/kolla/grafana/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'grafana_server': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000'}, 'grafana_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'grafana', 'value': {'container_name': 'grafana', 'group': 'grafana', 'enabled': True, 'image': 'registry.osism.tech/kolla/grafana:2024.2', 'volumes': ['/etc/kolla/grafana/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'grafana_server': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000'}, 'grafana_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'grafana', 'value': {'container_name': 'grafana', 'group': 'grafana', 'enabled': True, 'image': 'registry.osism.tech/kolla/grafana:2024.2', 'volumes': ['/etc/kolla/grafana/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'grafana_server': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000'}, 'grafana_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [haproxy-config : Configuring firewall for grafana] ***********************",
                                "Tuesday 24 February 2026  00:51:52 +0000 (0:00:00.588)       0:02:17.294 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'grafana_server', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'grafana_server_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'grafana_server', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'grafana_server_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'grafana_server', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'grafana_server_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [proxysql-config : Copying over grafana ProxySQL users config] ************",
                                "Tuesday 24 February 2026  00:51:52 +0000 (0:00:00.578)       0:02:17.873 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [proxysql-config : Copying over grafana ProxySQL rules config] ************",
                                "Tuesday 24 February 2026  00:51:54 +0000 (0:00:01.509)       0:02:19.382 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [include_role : heat] *****************************************************",
                                "Tuesday 24 February 2026  00:51:56 +0000 (0:00:01.975)       0:02:21.358 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [include_role : horizon] **************************************************",
                                "Tuesday 24 February 2026  00:51:56 +0000 (0:00:00.511)       0:02:21.869 ****** ",
                                "\u001b[0;36mincluded: horizon for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [haproxy-config : Copying over horizon haproxy config] ********************",
                                "Tuesday 24 February 2026  00:51:57 +0000 (0:00:00.886)       0:02:22.756 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'horizon', 'value': {'container_name': 'horizon', 'group': 'horizon', 'enabled': True, 'image': 'registry.osism.tech/kolla/horizon:2024.2', 'environment': {'ENABLE_BLAZAR': 'no', 'ENABLE_CLOUDKITTY': 'no', 'ENABLE_DESIGNATE': 'yes', 'ENABLE_FWAAS': 'no', 'ENABLE_HEAT': 'no', 'ENABLE_IRONIC': 'no', 'ENABLE_MAGNUM': 'yes', 'ENABLE_MANILA': 'yes', 'ENABLE_MASAKARI': 'no', 'ENABLE_MISTRAL': 'no', 'ENABLE_NEUTRON_VPNAAS': 'no', 'ENABLE_OCTAVIA': 'yes', 'ENABLE_TACKER': 'no', 'ENABLE_TROVE': 'no', 'ENABLE_WATCHER': 'no', 'ENABLE_ZUN': 'no', 'FORCE_GENERATE': 'no'}, 'volumes': ['/etc/kolla/horizon/:/var/lib/kolla/config_files/:ro', '', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:80'], 'timeout': '30'}, 'haproxy': {'horizon': {'enabled': True, 'mode': 'http', 'external': False, 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_redirect': {'enabled': True, 'mode': 'redirect', 'external': False, 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'horizon_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_external_redirect': {'enabled': True, 'mode': 'redirect', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'acme_client': {'enabled': True, 'with_frontend': False, 'custom_member_list': []}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'horizon', 'value': {'container_name': 'horizon', 'group': 'horizon', 'enabled': True, 'image': 'registry.osism.tech/kolla/horizon:2024.2', 'environment': {'ENABLE_BLAZAR': 'no', 'ENABLE_CLOUDKITTY': 'no', 'ENABLE_DESIGNATE': 'yes', 'ENABLE_FWAAS': 'no', 'ENABLE_HEAT': 'no', 'ENABLE_IRONIC': 'no', 'ENABLE_MAGNUM': 'yes', 'ENABLE_MANILA': 'yes', 'ENABLE_MASAKARI': 'no', 'ENABLE_MISTRAL': 'no', 'ENABLE_NEUTRON_VPNAAS': 'no', 'ENABLE_OCTAVIA': 'yes', 'ENABLE_TACKER': 'no', 'ENABLE_TROVE': 'no', 'ENABLE_WATCHER': 'no', 'ENABLE_ZUN': 'no', 'FORCE_GENERATE': 'no'}, 'volumes': ['/etc/kolla/horizon/:/var/lib/kolla/config_files/:ro', '', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:80'], 'timeout': '30'}, 'haproxy': {'horizon': {'enabled': True, 'mode': 'http', 'external': False, 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_redirect': {'enabled': True, 'mode': 'redirect', 'external': False, 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'horizon_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_external_redirect': {'enabled': True, 'mode': 'redirect', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'acme_client': {'enabled': True, 'with_frontend': False, 'custom_member_list': []}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'horizon', 'value': {'container_name': 'horizon', 'group': 'horizon', 'enabled': True, 'image': 'registry.osism.tech/kolla/horizon:2024.2', 'environment': {'ENABLE_BLAZAR': 'no', 'ENABLE_CLOUDKITTY': 'no', 'ENABLE_DESIGNATE': 'yes', 'ENABLE_FWAAS': 'no', 'ENABLE_HEAT': 'no', 'ENABLE_IRONIC': 'no', 'ENABLE_MAGNUM': 'yes', 'ENABLE_MANILA': 'yes', 'ENABLE_MASAKARI': 'no', 'ENABLE_MISTRAL': 'no', 'ENABLE_NEUTRON_VPNAAS': 'no', 'ENABLE_OCTAVIA': 'yes', 'ENABLE_TACKER': 'no', 'ENABLE_TROVE': 'no', 'ENABLE_WATCHER': 'no', 'ENABLE_ZUN': 'no', 'FORCE_GENERATE': 'no'}, 'volumes': ['/etc/kolla/horizon/:/var/lib/kolla/config_files/:ro', '', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:80'], 'timeout': '30'}, 'haproxy': {'horizon': {'enabled': True, 'mode': 'http', 'external': False, 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_redirect': {'enabled': True, 'mode': 'redirect', 'external': False, 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'horizon_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_external_redirect': {'enabled': True, 'mode': 'redirect', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'acme_client': {'enabled': True, 'with_frontend': False, 'custom_member_list': []}}}})\u001b[0m",
                                "",
                                "TASK [haproxy-config : Add configuration for horizon when using single external frontend] ***",
                                "Tuesday 24 February 2026  00:52:01 +0000 (0:00:04.063)       0:02:26.820 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'horizon', 'value': {'container_name': 'horizon', 'group': 'horizon', 'enabled': True, 'image': 'registry.osism.tech/kolla/horizon:2024.2', 'environment': {'ENABLE_BLAZAR': 'no', 'ENABLE_CLOUDKITTY': 'no', 'ENABLE_DESIGNATE': 'yes', 'ENABLE_FWAAS': 'no', 'ENABLE_HEAT': 'no', 'ENABLE_IRONIC': 'no', 'ENABLE_MAGNUM': 'yes', 'ENABLE_MANILA': 'yes', 'ENABLE_MASAKARI': 'no', 'ENABLE_MISTRAL': 'no', 'ENABLE_NEUTRON_VPNAAS': 'no', 'ENABLE_OCTAVIA': 'yes', 'ENABLE_TACKER': 'no', 'ENABLE_TROVE': 'no', 'ENABLE_WATCHER': 'no', 'ENABLE_ZUN': 'no', 'FORCE_GENERATE': 'no'}, 'volumes': ['/etc/kolla/horizon/:/var/lib/kolla/config_files/:ro', '', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:80'], 'timeout': '30'}, 'haproxy': {'horizon': {'enabled': True, 'mode': 'http', 'external': False, 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_redirect': {'enabled': True, 'mode': 'redirect', 'external': False, 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'horizon_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_external_redirect': {'enabled': True, 'mode': 'redirect', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'acme_client': {'enabled': True, 'with_frontend': False, 'custom_member_list': []}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'horizon', 'value': {'container_name': 'horizon', 'group': 'horizon', 'enabled': True, 'image': 'registry.osism.tech/kolla/horizon:2024.2', 'environment': {'ENABLE_BLAZAR': 'no', 'ENABLE_CLOUDKITTY': 'no', 'ENABLE_DESIGNATE': 'yes', 'ENABLE_FWAAS': 'no', 'ENABLE_HEAT': 'no', 'ENABLE_IRONIC': 'no', 'ENABLE_MAGNUM': 'yes', 'ENABLE_MANILA': 'yes', 'ENABLE_MASAKARI': 'no', 'ENABLE_MISTRAL': 'no', 'ENABLE_NEUTRON_VPNAAS': 'no', 'ENABLE_OCTAVIA': 'yes', 'ENABLE_TACKER': 'no', 'ENABLE_TROVE': 'no', 'ENABLE_WATCHER': 'no', 'ENABLE_ZUN': 'no', 'FORCE_GENERATE': 'no'}, 'volumes': ['/etc/kolla/horizon/:/var/lib/kolla/config_files/:ro', '', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:80'], 'timeout': '30'}, 'haproxy': {'horizon': {'enabled': True, 'mode': 'http', 'external': False, 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_redirect': {'enabled': True, 'mode': 'redirect', 'external': False, 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'horizon_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_external_redirect': {'enabled': True, 'mode': 'redirect', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'acme_client': {'enabled': True, 'with_frontend': False, 'custom_member_list': []}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'horizon', 'value': {'container_name': 'horizon', 'group': 'horizon', 'enabled': True, 'image': 'registry.osism.tech/kolla/horizon:2024.2', 'environment': {'ENABLE_BLAZAR': 'no', 'ENABLE_CLOUDKITTY': 'no', 'ENABLE_DESIGNATE': 'yes', 'ENABLE_FWAAS': 'no', 'ENABLE_HEAT': 'no', 'ENABLE_IRONIC': 'no', 'ENABLE_MAGNUM': 'yes', 'ENABLE_MANILA': 'yes', 'ENABLE_MASAKARI': 'no', 'ENABLE_MISTRAL': 'no', 'ENABLE_NEUTRON_VPNAAS': 'no', 'ENABLE_OCTAVIA': 'yes', 'ENABLE_TACKER': 'no', 'ENABLE_TROVE': 'no', 'ENABLE_WATCHER': 'no', 'ENABLE_ZUN': 'no', 'FORCE_GENERATE': 'no'}, 'volumes': ['/etc/kolla/horizon/:/var/lib/kolla/config_files/:ro', '', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:80'], 'timeout': '30'}, 'haproxy': {'horizon': {'enabled': True, 'mode': 'http', 'external': False, 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_redirect': {'enabled': True, 'mode': 'redirect', 'external': False, 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'horizon_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_external_redirect': {'enabled': True, 'mode': 'redirect', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'acme_client': {'enabled': True, 'with_frontend': False, 'custom_member_list': []}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [haproxy-config : Configuring firewall for horizon] ***********************",
                                "Tuesday 24 February 2026  00:52:03 +0000 (0:00:01.431)       0:02:28.251 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'horizon', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'horizon_redirect', 'value': {'enabled': True, 'mode': 'redirect', 'external': False, 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'horizon_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'horizon_external_redirect', 'value': {'enabled': True, 'mode': 'redirect', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'acme_client', 'value': {'enabled': True, 'with_frontend': False, 'custom_member_list': []}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'horizon', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'horizon_redirect', 'value': {'enabled': True, 'mode': 'redirect', 'external': False, 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'horizon_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'horizon_external_redirect', 'value': {'enabled': True, 'mode': 'redirect', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'acme_client', 'value': {'enabled': True, 'with_frontend': False, 'custom_member_list': []}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'horizon', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'horizon_redirect', 'value': {'enabled': True, 'mode': 'redirect', 'external': False, 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'horizon_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'horizon_external_redirect', 'value': {'enabled': True, 'mode': 'redirect', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'acme_client', 'value': {'enabled': True, 'with_frontend': False, 'custom_member_list': []}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [proxysql-config : Copying over horizon ProxySQL users config] ************",
                                "Tuesday 24 February 2026  00:52:05 +0000 (0:00:01.807)       0:02:30.059 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [proxysql-config : Copying over horizon ProxySQL rules config] ************",
                                "Tuesday 24 February 2026  00:52:06 +0000 (0:00:01.524)       0:02:31.583 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [include_role : influxdb] *************************************************",
                                "Tuesday 24 February 2026  00:52:09 +0000 (0:00:02.471)       0:02:34.054 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [include_role : ironic] ***************************************************",
                                "Tuesday 24 February 2026  00:52:09 +0000 (0:00:00.343)       0:02:34.398 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [include_role : keystone] *************************************************",
                                "Tuesday 24 February 2026  00:52:09 +0000 (0:00:00.478)       0:02:34.876 ****** ",
                                "\u001b[0;36mincluded: keystone for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [haproxy-config : Copying over keystone haproxy config] *******************",
                                "Tuesday 24 February 2026  00:52:10 +0000 (0:00:00.792)       0:02:35.669 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone:2024.2', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-ssh:2024.2', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-fernet:2024.2', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone:2024.2', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-ssh:2024.2', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-fernet:2024.2', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone:2024.2', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-ssh:2024.2', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-fernet:2024.2', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}}) \u001b[0m",
                                "",
                                "TASK [haproxy-config : Add configuration for keystone when using single external frontend] ***",
                                "Tuesday 24 February 2026  00:52:15 +0000 (0:00:04.897)       0:02:40.566 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone:2024.2', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-ssh:2024.2', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-fernet:2024.2', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone:2024.2', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-ssh:2024.2', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-fernet:2024.2', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone:2024.2', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-ssh:2024.2', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-fernet:2024.2', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [haproxy-config : Configuring firewall for keystone] **********************",
                                "Tuesday 24 February 2026  00:52:16 +0000 (0:00:01.083)       0:02:41.650 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'keystone_internal', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'keystone_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'keystone_internal', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'keystone_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'keystone_internal', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'keystone_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [proxysql-config : Copying over keystone ProxySQL users config] ***********",
                                "Tuesday 24 February 2026  00:52:17 +0000 (0:00:01.044)       0:02:42.695 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [proxysql-config : Copying over keystone ProxySQL rules config] ***********",
                                "Tuesday 24 February 2026  00:52:19 +0000 (0:00:01.621)       0:02:44.316 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [include_role : letsencrypt] **********************************************",
                                "Tuesday 24 February 2026  00:52:21 +0000 (0:00:01.915)       0:02:46.232 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [include_role : magnum] ***************************************************",
                                "Tuesday 24 February 2026  00:52:21 +0000 (0:00:00.510)       0:02:46.743 ****** ",
                                "\u001b[0;36mincluded: magnum for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [haproxy-config : Copying over magnum haproxy config] *********************",
                                "Tuesday 24 February 2026  00:52:22 +0000 (0:00:00.914)       0:02:47.658 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-api:2024.2', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-conductor:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.10,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-api:2024.2', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-conductor:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.11,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-api:2024.2', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-conductor:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.12,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "",
                                "TASK [haproxy-config : Add configuration for magnum when using single external frontend] ***",
                                "Tuesday 24 February 2026  00:52:26 +0000 (0:00:03.738)       0:02:51.396 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-api:2024.2', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-conductor:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.10,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-api:2024.2', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-conductor:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.11,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-api:2024.2', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-conductor:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.12,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [haproxy-config : Configuring firewall for magnum] ************************",
                                "Tuesday 24 February 2026  00:52:27 +0000 (0:00:00.740)       0:02:52.137 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'magnum_api', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'magnum_api_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'magnum_api', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'magnum_api_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'magnum_api', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'magnum_api_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [proxysql-config : Copying over magnum ProxySQL users config] *************",
                                "Tuesday 24 February 2026  00:52:28 +0000 (0:00:00.769)       0:02:52.907 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [proxysql-config : Copying over magnum ProxySQL rules config] *************",
                                "Tuesday 24 February 2026  00:52:29 +0000 (0:00:01.333)       0:02:54.241 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [include_role : manila] ***************************************************",
                                "Tuesday 24 February 2026  00:52:31 +0000 (0:00:02.157)       0:02:56.398 ****** ",
                                "\u001b[0;36mincluded: manila for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [haproxy-config : Copying over manila haproxy config] *********************",
                                "Tuesday 24 February 2026  00:52:32 +0000 (0:00:01.396)       0:02:57.794 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'manila-api', 'value': {'container_name': 'manila_api', 'group': 'manila-api', 'image': 'registry.osism.tech/kolla/manila-api:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/manila-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8786'], 'timeout': '30'}, 'haproxy': {'manila_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8786', 'listen_port': '8786'}, 'manila_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8786', 'listen_port': '8786'}}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'manila-scheduler', 'value': {'container_name': 'manila_scheduler', 'group': 'manila-scheduler', 'image': 'registry.osism.tech/kolla/manila-scheduler:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/manila-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'manila-share', 'value': {'container_name': 'manila_share', 'group': 'manila-share', 'image': 'registry.osism.tech/kolla/manila-share:2024.2', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/manila-share/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run:/run:shared', 'kolla_logs:/var/log/kolla/', '', '/lib/modules:/lib/modules:ro', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-share 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'manila-data', 'value': {'container_name': 'manila_data', 'group': 'manila-data', 'image': 'registry.osism.tech/kolla/manila-data:2024.2', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/manila-data/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run:/run:shared', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-data 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'manila-api', 'value': {'container_name': 'manila_api', 'group': 'manila-api', 'image': 'registry.osism.tech/kolla/manila-api:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/manila-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8786'], 'timeout': '30'}, 'haproxy': {'manila_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8786', 'listen_port': '8786'}, 'manila_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8786', 'listen_port': '8786'}}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'manila-scheduler', 'value': {'container_name': 'manila_scheduler', 'group': 'manila-scheduler', 'image': 'registry.osism.tech/kolla/manila-scheduler:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/manila-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'manila-share', 'value': {'container_name': 'manila_share', 'group': 'manila-share', 'image': 'registry.osism.tech/kolla/manila-share:2024.2', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/manila-share/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run:/run:shared', 'kolla_logs:/var/log/kolla/', '', '/lib/modules:/lib/modules:ro', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-share 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'manila-api', 'value': {'container_name': 'manila_api', 'group': 'manila-api', 'image': 'registry.osism.tech/kolla/manila-api:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/manila-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8786'], 'timeout': '30'}, 'haproxy': {'manila_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8786', 'listen_port': '8786'}, 'manila_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8786', 'listen_port': '8786'}}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'manila-data', 'value': {'container_name': 'manila_data', 'group': 'manila-data', 'image': 'registry.osism.tech/kolla/manila-data:2024.2', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/manila-data/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run:/run:shared', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-data 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'manila-scheduler', 'value': {'container_name': 'manila_scheduler', 'group': 'manila-scheduler', 'image': 'registry.osism.tech/kolla/manila-scheduler:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/manila-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'manila-share', 'value': {'container_name': 'manila_share', 'group': 'manila-share', 'image': 'registry.osism.tech/kolla/manila-share:2024.2', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/manila-share/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run:/run:shared', 'kolla_logs:/var/log/kolla/', '', '/lib/modules:/lib/modules:ro', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-share 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'manila-data', 'value': {'container_name': 'manila_data', 'group': 'manila-data', 'image': 'registry.osism.tech/kolla/manila-data:2024.2', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/manila-data/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run:/run:shared', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-data 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "",
                                "TASK [haproxy-config : Add configuration for manila when using single external frontend] ***",
                                "Tuesday 24 February 2026  00:52:37 +0000 (0:00:04.339)       0:03:02.134 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'manila-api', 'value': {'container_name': 'manila_api', 'group': 'manila-api', 'image': 'registry.osism.tech/kolla/manila-api:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/manila-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8786'], 'timeout': '30'}, 'haproxy': {'manila_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8786', 'listen_port': '8786'}, 'manila_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8786', 'listen_port': '8786'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'manila-scheduler', 'value': {'container_name': 'manila_scheduler', 'group': 'manila-scheduler', 'image': 'registry.osism.tech/kolla/manila-scheduler:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/manila-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'manila-share', 'value': {'container_name': 'manila_share', 'group': 'manila-share', 'image': 'registry.osism.tech/kolla/manila-share:2024.2', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/manila-share/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run:/run:shared', 'kolla_logs:/var/log/kolla/', '', '/lib/modules:/lib/modules:ro', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-share 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'manila-data', 'value': {'container_name': 'manila_data', 'group': 'manila-data', 'image': 'registry.osism.tech/kolla/manila-data:2024.2', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/manila-data/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run:/run:shared', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-data 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'manila-api', 'value': {'container_name': 'manila_api', 'group': 'manila-api', 'image': 'registry.osism.tech/kolla/manila-api:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/manila-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8786'], 'timeout': '30'}, 'haproxy': {'manila_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8786', 'listen_port': '8786'}, 'manila_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8786', 'listen_port': '8786'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'manila-scheduler', 'value': {'container_name': 'manila_scheduler', 'group': 'manila-scheduler', 'image': 'registry.osism.tech/kolla/manila-scheduler:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/manila-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'manila-share', 'value': {'container_name': 'manila_share', 'group': 'manila-share', 'image': 'registry.osism.tech/kolla/manila-share:2024.2', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/manila-share/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run:/run:shared', 'kolla_logs:/var/log/kolla/', '', '/lib/modules:/lib/modules:ro', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-share 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'manila-data', 'value': {'container_name': 'manila_data', 'group': 'manila-data', 'image': 'registry.osism.tech/kolla/manila-data:2024.2', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/manila-data/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run:/run:shared', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-data 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'manila-api', 'value': {'container_name': 'manila_api', 'group': 'manila-api', 'image': 'registry.osism.tech/kolla/manila-api:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/manila-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8786'], 'timeout': '30'}, 'haproxy': {'manila_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8786', 'listen_port': '8786'}, 'manila_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8786', 'listen_port': '8786'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'manila-scheduler', 'value': {'container_name': 'manila_scheduler', 'group': 'manila-scheduler', 'image': 'registry.osism.tech/kolla/manila-scheduler:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/manila-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'manila-share', 'value': {'container_name': 'manila_share', 'group': 'manila-share', 'image': 'registry.osism.tech/kolla/manila-share:2024.2', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/manila-share/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run:/run:shared', 'kolla_logs:/var/log/kolla/', '', '/lib/modules:/lib/modules:ro', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-share 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'manila-data', 'value': {'container_name': 'manila_data', 'group': 'manila-data', 'image': 'registry.osism.tech/kolla/manila-data:2024.2', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/manila-data/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/run:/run:shared', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port manila-data 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [haproxy-config : Configuring firewall for manila] ************************",
                                "Tuesday 24 February 2026  00:52:37 +0000 (0:00:00.727)       0:03:02.861 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'manila_api', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8786', 'listen_port': '8786'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'manila_api_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8786', 'listen_port': '8786'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'manila_api', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8786', 'listen_port': '8786'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'manila_api_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8786', 'listen_port': '8786'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'manila_api', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8786', 'listen_port': '8786'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'manila_api_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8786', 'listen_port': '8786'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [proxysql-config : Copying over manila ProxySQL users config] *************",
                                "Tuesday 24 February 2026  00:52:39 +0000 (0:00:01.397)       0:03:04.259 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [proxysql-config : Copying over manila ProxySQL rules config] *************",
                                "Tuesday 24 February 2026  00:52:40 +0000 (0:00:01.385)       0:03:05.645 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [include_role : mariadb] **************************************************",
                                "Tuesday 24 February 2026  00:52:42 +0000 (0:00:01.897)       0:03:07.542 ****** ",
                                "\u001b[0;36mincluded: mariadb for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [mariadb : Ensure mysql monitor user exist] *******************************",
                                "Tuesday 24 February 2026  00:52:43 +0000 (0:00:01.170)       0:03:08.713 ****** ",
                                "\u001b[0;32mok: [testbed-node-0] => (item=testbed-node-0)\u001b[0m",
                                "",
                                "TASK [haproxy-config : Copying over mariadb haproxy config] ********************",
                                "Tuesday 24 February 2026  00:52:46 +0000 (0:00:03.129)       0:03:11.843 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/mariadb-server:2024.2', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.10', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'mariadb-clustercheck', 'value': {'container_name': 'mariadb_clustercheck', 'group': 'mariadb_shard_0', 'enabled': False, 'image': 'registry.osism.tech/kolla/mariadb-clustercheck:2024.2', 'volumes': ['/etc/kolla/mariadb-clustercheck/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.10', 'AVAILABLE_WHEN_DONOR': '1'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/mariadb-server:2024.2', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.11', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/mariadb-server:2024.2', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.12', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'mariadb-clustercheck', 'value': {'container_name': 'mariadb_clustercheck', 'group': 'mariadb_shard_0', 'enabled': False, 'image': 'registry.osism.tech/kolla/mariadb-clustercheck:2024.2', 'volumes': ['/etc/kolla/mariadb-clustercheck/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.11', 'AVAILABLE_WHEN_DONOR': '1'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'mariadb-clustercheck', 'value': {'container_name': 'mariadb_clustercheck', 'group': 'mariadb_shard_0', 'enabled': False, 'image': 'registry.osism.tech/kolla/mariadb-clustercheck:2024.2', 'volumes': ['/etc/kolla/mariadb-clustercheck/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.12', 'AVAILABLE_WHEN_DONOR': '1'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [haproxy-config : Add configuration for mariadb when using single external frontend] ***",
                                "Tuesday 24 February 2026  00:52:49 +0000 (0:00:02.346)       0:03:14.190 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/mariadb-server:2024.2', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.10', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'mariadb-clustercheck', 'value': {'container_name': 'mariadb_clustercheck', 'group': 'mariadb_shard_0', 'enabled': False, 'image': 'registry.osism.tech/kolla/mariadb-clustercheck:2024.2', 'volumes': ['/etc/kolla/mariadb-clustercheck/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.10', 'AVAILABLE_WHEN_DONOR': '1'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/mariadb-server:2024.2', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.11', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'mariadb-clustercheck', 'value': {'container_name': 'mariadb_clustercheck', 'group': 'mariadb_shard_0', 'enabled': False, 'image': 'registry.osism.tech/kolla/mariadb-clustercheck:2024.2', 'volumes': ['/etc/kolla/mariadb-clustercheck/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.11', 'AVAILABLE_WHEN_DONOR': '1'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/mariadb-server:2024.2', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.12', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'mariadb-clustercheck', 'value': {'container_name': 'mariadb_clustercheck', 'group': 'mariadb_shard_0', 'enabled': False, 'image': 'registry.osism.tech/kolla/mariadb-clustercheck:2024.2', 'volumes': ['/etc/kolla/mariadb-clustercheck/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.12', 'AVAILABLE_WHEN_DONOR': '1'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [haproxy-config : Configuring firewall for mariadb] ***********************",
                                "Tuesday 24 February 2026  00:52:51 +0000 (0:00:02.385)       0:03:16.575 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'mariadb', 'value': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'mariadb_external_lb', 'value': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'mariadb', 'value': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'mariadb_external_lb', 'value': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'mariadb', 'value': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'mariadb_external_lb', 'value': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [proxysql-config : Copying over mariadb ProxySQL users config] ************",
                                "Tuesday 24 February 2026  00:52:54 +0000 (0:00:02.897)       0:03:19.473 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [proxysql-config : Copying over mariadb ProxySQL rules config] ************",
                                "Tuesday 24 February 2026  00:52:56 +0000 (0:00:01.952)       0:03:21.426 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [include_role : masakari] *************************************************",
                                "Tuesday 24 February 2026  00:52:58 +0000 (0:00:01.536)       0:03:22.962 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [include_role : memcached] ************************************************",
                                "Tuesday 24 February 2026  00:52:58 +0000 (0:00:00.350)       0:03:23.313 ****** ",
                                "\u001b[0;36mincluded: memcached for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [haproxy-config : Copying over memcached haproxy config] ******************",
                                "Tuesday 24 February 2026  00:52:59 +0000 (0:00:01.433)       0:03:24.747 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'memcached', 'value': {'container_name': 'memcached', 'image': 'registry.osism.tech/kolla/memcached:2024.2', 'enabled': True, 'group': 'memcached', 'volumes': ['/etc/kolla/memcached/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen memcached 11211'], 'timeout': '30'}, 'haproxy': {'memcached': {'enabled': False, 'mode': 'tcp', 'port': '11211', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'active_passive': True}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'memcached', 'value': {'container_name': 'memcached', 'image': 'registry.osism.tech/kolla/memcached:2024.2', 'enabled': True, 'group': 'memcached', 'volumes': ['/etc/kolla/memcached/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen memcached 11211'], 'timeout': '30'}, 'haproxy': {'memcached': {'enabled': False, 'mode': 'tcp', 'port': '11211', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'active_passive': True}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'memcached', 'value': {'container_name': 'memcached', 'image': 'registry.osism.tech/kolla/memcached:2024.2', 'enabled': True, 'group': 'memcached', 'volumes': ['/etc/kolla/memcached/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen memcached 11211'], 'timeout': '30'}, 'haproxy': {'memcached': {'enabled': False, 'mode': 'tcp', 'port': '11211', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'active_passive': True}}}})\u001b[0m",
                                "",
                                "TASK [haproxy-config : Add configuration for memcached when using single external frontend] ***",
                                "Tuesday 24 February 2026  00:53:01 +0000 (0:00:01.566)       0:03:26.313 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'memcached', 'value': {'container_name': 'memcached', 'image': 'registry.osism.tech/kolla/memcached:2024.2', 'enabled': True, 'group': 'memcached', 'volumes': ['/etc/kolla/memcached/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen memcached 11211'], 'timeout': '30'}, 'haproxy': {'memcached': {'enabled': False, 'mode': 'tcp', 'port': '11211', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'active_passive': True}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'memcached', 'value': {'container_name': 'memcached', 'image': 'registry.osism.tech/kolla/memcached:2024.2', 'enabled': True, 'group': 'memcached', 'volumes': ['/etc/kolla/memcached/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen memcached 11211'], 'timeout': '30'}, 'haproxy': {'memcached': {'enabled': False, 'mode': 'tcp', 'port': '11211', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'active_passive': True}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'memcached', 'value': {'container_name': 'memcached', 'image': 'registry.osism.tech/kolla/memcached:2024.2', 'enabled': True, 'group': 'memcached', 'volumes': ['/etc/kolla/memcached/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen memcached 11211'], 'timeout': '30'}, 'haproxy': {'memcached': {'enabled': False, 'mode': 'tcp', 'port': '11211', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'active_passive': True}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [haproxy-config : Configuring firewall for memcached] *********************",
                                "Tuesday 24 February 2026  00:53:01 +0000 (0:00:00.398)       0:03:26.711 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'memcached', 'value': {'enabled': False, 'mode': 'tcp', 'port': '11211', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'active_passive': True}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'memcached', 'value': {'enabled': False, 'mode': 'tcp', 'port': '11211', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'active_passive': True}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'memcached', 'value': {'enabled': False, 'mode': 'tcp', 'port': '11211', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'active_passive': True}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [proxysql-config : Copying over memcached ProxySQL users config] **********",
                                "Tuesday 24 February 2026  00:53:02 +0000 (0:00:00.792)       0:03:27.504 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [proxysql-config : Copying over memcached ProxySQL rules config] **********",
                                "Tuesday 24 February 2026  00:53:03 +0000 (0:00:00.440)       0:03:27.944 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [include_role : mistral] **************************************************",
                                "Tuesday 24 February 2026  00:53:04 +0000 (0:00:01.288)       0:03:29.232 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [include_role : neutron] **************************************************",
                                "Tuesday 24 February 2026  00:53:04 +0000 (0:00:00.329)       0:03:29.562 ****** ",
                                "\u001b[0;36mincluded: neutron for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [haproxy-config : Copying over neutron haproxy config] ********************",
                                "Tuesday 24 February 2026  00:53:06 +0000 (0:00:01.453)       0:03:31.016 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2024.2', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-openvswitch-agent', 'value': {'container_name': 'neutron_openvswitch_agent', 'image': 'registry.osism.tech/kolla/neutron-openvswitch-agent:2024.2', 'enabled': False, 'privileged': True, 'environment': {'KOLLA_LEGACY_IPTABLES': 'false'}, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-openvswitch-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-openvswitch-agent 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-linuxbridge-agent', 'value': {'container_name': 'neutron_linuxbridge_agent', 'image': 'registry.osism.tech/kolla/neutron-linuxbridge-agent:2024.2', 'privileged': True, 'enabled': False, 'environment': {'KOLLA_LEGACY_IPTABLES': 'false'}, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-linuxbridge-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-linuxbridge-agent 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-dhcp-agent', 'value': {'container_name': 'neutron_dhcp_agent', 'image': 'registry.osism.tech/kolla/neutron-dhcp-agent:2024.2', 'privileged': True, 'enabled': False, 'group': 'neutron-dhcp-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-dhcp-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-dhcp-agent 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-l3-agent', 'value': {'container_name': 'neutron_l3_agent', 'image': 'registry.osism.tech/kolla/neutron-l3-agent:2024.2', 'privileged': True, 'enabled': False, 'environment': {'KOLLA_LEGACY_IPTABLES': 'false'}, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-l3-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', \"healthcheck_port 'neutron-l3-agent ' 5672\"], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-sriov-agent', 'value': {'container_name': 'neutron_sriov_agent', 'image': 'registry.osism.tech/kolla/neutron-sriov-agent:2024.2', 'privileged': True, 'enabled': False, 'host_in_groups': False, 'volumes': ['/etc/kolla/neutron-sriov-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-sriov-nic-agent 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-mlnx-agent', 'value': {'container_name': 'neutron_mlnx_agent', 'image': 'registry.osism.tech/kolla/neutron-mlnx-agent:2024.2', 'enabled': False, 'host_in_groups': False, 'volumes': ['/etc/kolla/neutron-mlnx-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2024.2', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-eswitchd', 'value': {'container_name': 'neutron_eswitchd', 'image': 'registry.osism.tech/kolla/neutron-eswitchd:2024.2', 'privileged': True, 'enabled': False, 'host_in_groups': False, 'volumes': ['/etc/kolla/neutron-eswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/run/libvirt:/run/libvirt:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-metadata-agent', 'value': {'container_name': 'neutron_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'privileged': True, 'enabled': False, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-metadata-agent 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-openvswitch-agent', 'value': {'container_name': 'neutron_openvswitch_agent', 'image': 'registry.osism.tech/kolla/neutron-openvswitch-agent:2024.2', 'enabled': False, 'privileged': True, 'environment': {'KOLLA_LEGACY_IPTABLES': 'false'}, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-openvswitch-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-openvswitch-agent 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'privileged': True, 'enabled': True, 'host_in_groups': False, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-linuxbridge-agent', 'value': {'container_name': 'neutron_linuxbridge_agent', 'image': 'registry.osism.tech/kolla/neutron-linuxbridge-agent:2024.2', 'privileged': True, 'enabled': False, 'environment': {'KOLLA_LEGACY_IPTABLES': 'false'}, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-linuxbridge-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-linuxbridge-agent 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-bgp-dragent', 'value': {'container_name': 'neutron_bgp_dragent', 'image': 'registry.osism.tech/kolla/neutron-bgp-dragent:2024.2', 'privileged': True, 'enabled': False, 'group': 'neutron-bgp-dragent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-bgp-dragent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-bgp-dragent 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-dhcp-agent', 'value': {'container_name': 'neutron_dhcp_agent', 'image': 'registry.osism.tech/kolla/neutron-dhcp-agent:2024.2', 'privileged': True, 'enabled': False, 'group': 'neutron-dhcp-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-dhcp-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-dhcp-agent 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-infoblox-ipam-agent', 'value': {'container_name': 'neutron_infoblox_ipam_agent', 'image': 'registry.osism.tech/kolla/neutron-infoblox-ipam-agent:2024.2', 'privileged': True, 'enabled': False, 'group': 'neutron-infoblox-ipam-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-infoblox-ipam-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-l3-agent', 'value': {'container_name': 'neutron_l3_agent', 'image': 'registry.osism.tech/kolla/neutron-l3-agent:2024.2', 'privileged': True, 'enabled': False, 'environment': {'KOLLA_LEGACY_IPTABLES': 'false'}, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-l3-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', \"healthcheck_port 'neutron-l3-agent ' 5672\"], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-metering-agent', 'value': {'container_name': 'neutron_metering_agent', 'image': 'registry.osism.tech/kolla/neutron-metering-agent:2024.2', 'privileged': True, 'enabled': False, 'group': 'neutron-metering-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-metering-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2024.2', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'ironic-neutron-agent', 'value': {'container_name': 'ironic_neutron_agent', 'image': 'registry.osism.tech/kolla/ironic-neutron-agent:2024.2', 'privileged': False, 'enabled': False, 'group': 'ironic-neutron-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/ironic-neutron-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port ironic-neutron-agent 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-sriov-agent', 'value': {'container_name': 'neutron_sriov_agent', 'image': 'registry.osism.tech/kolla/neutron-sriov-agent:2024.2', 'privileged': True, 'enabled': False, 'host_in_groups': False, 'volumes': ['/etc/kolla/neutron-sriov-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-sriov-nic-agent 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-openvswitch-agent', 'value': {'container_name': 'neutron_openvswitch_agent', 'image': 'registry.osism.tech/kolla/neutron-openvswitch-agent:2024.2', 'enabled': False, 'privileged': True, 'environment': {'KOLLA_LEGACY_IPTABLES': 'false'}, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-openvswitch-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-openvswitch-agent 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-tls-proxy', 'value': {'container_name': 'neutron_tls_proxy', 'group': 'neutron-server', 'host_in_groups': True, 'enabled': 'no', 'image': 'registry.osism.tech/kolla/neutron-tls-proxy:2024.2', 'volumes': ['/etc/kolla/neutron-tls-proxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl -u openstack:password 192.168.16.10:9697'], 'timeout': '30'}, 'haproxy': {'neutron_tls_proxy': {'enabled': False, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696', 'tls_backend': 'yes'}, 'neutron_tls_proxy_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696', 'tls_backend': 'yes'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-linuxbridge-agent', 'value': {'container_name': 'neutron_linuxbridge_agent', 'image': 'registry.osism.tech/kolla/neutron-linuxbridge-agent:2024.2', 'privileged': True, 'enabled': False, 'environment': {'KOLLA_LEGACY_IPTABLES': 'false'}, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-linuxbridge-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-linuxbridge-agent 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-ovn-agent', 'value': {'container_name': 'neutron_ovn_agent', 'group': 'neutron-ovn-agent', 'host_in_groups': False, 'enabled': False, 'image': 'registry.osism.tech/dockerhub/kolla/neutron-ovn-agent:2024.2', 'volumes': ['/etc/kolla/neutron-ovn-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-agent 6640'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-dhcp-agent', 'value': {'container_name': 'neutron_dhcp_agent', 'image': 'registry.osism.tech/kolla/neutron-dhcp-agent:2024.2', 'privileged': True, 'enabled': False, 'group': 'neutron-dhcp-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-dhcp-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-dhcp-agent 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-l3-agent', 'value': {'container_name': 'neutron_l3_agent', 'image': 'registry.osism.tech/kolla/neutron-l3-agent:2024.2', 'privileged': True, 'enabled': False, 'environment': {'KOLLA_LEGACY_IPTABLES': 'false'}, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-l3-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', \"healthcheck_port 'neutron-l3-agent ' 5672\"], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-sriov-agent', 'value': {'container_name': 'neutron_sriov_agent', 'image': 'registry.osism.tech/kolla/neutron-sriov-agent:2024.2', 'privileged': True, 'enabled': False, 'host_in_groups': False, 'volumes': ['/etc/kolla/neutron-sriov-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-sriov-nic-agent 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-mlnx-agent', 'value': {'container_name': 'neutron_mlnx_agent', 'image': 'registry.osism.tech/kolla/neutron-mlnx-agent:2024.2', 'enabled': False, 'host_in_groups': False, 'volumes': ['/etc/kolla/neutron-mlnx-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-eswitchd', 'value': {'container_name': 'neutron_eswitchd', 'image': 'registry.osism.tech/kolla/neutron-eswitchd:2024.2', 'privileged': True, 'enabled': False, 'host_in_groups': False, 'volumes': ['/etc/kolla/neutron-eswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/run/libvirt:/run/libvirt:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-mlnx-agent', 'value': {'container_name': 'neutron_mlnx_agent', 'image': 'registry.osism.tech/kolla/neutron-mlnx-agent:2024.2', 'enabled': False, 'host_in_groups': False, 'volumes': ['/etc/kolla/neutron-mlnx-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-metadata-agent', 'value': {'container_name': 'neutron_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'privileged': True, 'enabled': False, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-metadata-agent 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-eswitchd', 'value': {'container_name': 'neutron_eswitchd', 'image': 'registry.osism.tech/kolla/neutron-eswitchd:2024.2', 'privileged': True, 'enabled': False, 'host_in_groups': False, 'volumes': ['/etc/kolla/neutron-eswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/run/libvirt:/run/libvirt:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'privileged': True, 'enabled': True, 'host_in_groups': False, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-bgp-dragent', 'value': {'container_name': 'neutron_bgp_dragent', 'image': 'registry.osism.tech/kolla/neutron-bgp-dragent:2024.2', 'privileged': True, 'enabled': False, 'group': 'neutron-bgp-dragent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-bgp-dragent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-bgp-dragent 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-metadata-agent', 'value': {'container_name': 'neutron_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'privileged': True, 'enabled': False, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-metadata-agent 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-infoblox-ipam-agent', 'value': {'container_name': 'neutron_infoblox_ipam_agent', 'image': 'registry.osism.tech/kolla/neutron-infoblox-ipam-agent:2024.2', 'privileged': True, 'enabled': False, 'group': 'neutron-infoblox-ipam-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-infoblox-ipam-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'privileged': True, 'enabled': True, 'host_in_groups': False, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-metering-agent', 'value': {'container_name': 'neutron_metering_agent', 'image': 'registry.osism.tech/kolla/neutron-metering-agent:2024.2', 'privileged': True, 'enabled': False, 'group': 'neutron-metering-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-metering-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-bgp-dragent', 'value': {'container_name': 'neutron_bgp_dragent', 'image': 'registry.osism.tech/kolla/neutron-bgp-dragent:2024.2', 'privileged': True, 'enabled': False, 'group': 'neutron-bgp-dragent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-bgp-dragent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-bgp-dragent 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'ironic-neutron-agent', 'value': {'container_name': 'ironic_neutron_agent', 'image': 'registry.osism.tech/kolla/ironic-neutron-agent:2024.2', 'privileged': False, 'enabled': False, 'group': 'ironic-neutron-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/ironic-neutron-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port ironic-neutron-agent 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-infoblox-ipam-agent', 'value': {'container_name': 'neutron_infoblox_ipam_agent', 'image': 'registry.osism.tech/kolla/neutron-infoblox-ipam-agent:2024.2', 'privileged': True, 'enabled': False, 'group': 'neutron-infoblox-ipam-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-infoblox-ipam-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-tls-proxy', 'value': {'container_name': 'neutron_tls_proxy', 'group': 'neutron-server', 'host_in_groups': True, 'enabled': 'no', 'image': 'registry.osism.tech/kolla/neutron-tls-proxy:2024.2', 'volumes': ['/etc/kolla/neutron-tls-proxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl -u openstack:password 192.168.16.12:9697'], 'timeout': '30'}, 'haproxy': {'neutron_tls_proxy': {'enabled': False, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696', 'tls_backend': 'yes'}, 'neutron_tls_proxy_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696', 'tls_backend': 'yes'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-metering-agent', 'value': {'container_name': 'neutron_metering_agent', 'image': 'registry.osism.tech/kolla/neutron-metering-agent:2024.2', 'privileged': True, 'enabled': False, 'group': 'neutron-metering-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-metering-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-ovn-agent', 'value': {'container_name': 'neutron_ovn_agent', 'group': 'neutron-ovn-agent', 'host_in_groups': False, 'enabled': False, 'image': 'registry.osism.tech/dockerhub/kolla/neutron-ovn-agent:2024.2', 'volumes': ['/etc/kolla/neutron-ovn-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-agent 6640'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'ironic-neutron-agent', 'value': {'container_name': 'ironic_neutron_agent', 'image': 'registry.osism.tech/kolla/ironic-neutron-agent:2024.2', 'privileged': False, 'enabled': False, 'group': 'ironic-neutron-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/ironic-neutron-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port ironic-neutron-agent 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-tls-proxy', 'value': {'container_name': 'neutron_tls_proxy', 'group': 'neutron-server', 'host_in_groups': True, 'enabled': 'no', 'image': 'registry.osism.tech/kolla/neutron-tls-proxy:2024.2', 'volumes': ['/etc/kolla/neutron-tls-proxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl -u openstack:password 192.168.16.11:9697'], 'timeout': '30'}, 'haproxy': {'neutron_tls_proxy': {'enabled': False, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696', 'tls_backend': 'yes'}, 'neutron_tls_proxy_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696', 'tls_backend': 'yes'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-ovn-agent', 'value': {'container_name': 'neutron_ovn_agent', 'group': 'neutron-ovn-agent', 'host_in_groups': False, 'enabled': False, 'image': 'registry.osism.tech/dockerhub/kolla/neutron-ovn-agent:2024.2', 'volumes': ['/etc/kolla/neutron-ovn-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-agent 6640'], 'timeout': '30'}}}) \u001b[0m",
                                "",
                                "TASK [haproxy-config : Add configuration for neutron when using single external frontend] ***",
                                "Tuesday 24 February 2026  00:53:10 +0000 (0:00:04.136)       0:03:35.152 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2024.2', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-openvswitch-agent', 'value': {'container_name': 'neutron_openvswitch_agent', 'image': 'registry.osism.tech/kolla/neutron-openvswitch-agent:2024.2', 'enabled': False, 'privileged': True, 'environment': {'KOLLA_LEGACY_IPTABLES': 'false'}, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-openvswitch-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-openvswitch-agent 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-linuxbridge-agent', 'value': {'container_name': 'neutron_linuxbridge_agent', 'image': 'registry.osism.tech/kolla/neutron-linuxbridge-agent:2024.2', 'privileged': True, 'enabled': False, 'environment': {'KOLLA_LEGACY_IPTABLES': 'false'}, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-linuxbridge-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-linuxbridge-agent 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-dhcp-agent', 'value': {'container_name': 'neutron_dhcp_agent', 'image': 'registry.osism.tech/kolla/neutron-dhcp-agent:2024.2', 'privileged': True, 'enabled': False, 'group': 'neutron-dhcp-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-dhcp-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-dhcp-agent 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-l3-agent', 'value': {'container_name': 'neutron_l3_agent', 'image': 'registry.osism.tech/kolla/neutron-l3-agent:2024.2', 'privileged': True, 'enabled': False, 'environment': {'KOLLA_LEGACY_IPTABLES': 'false'}, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-l3-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', \"healthcheck_port 'neutron-l3-agent ' 5672\"], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-sriov-agent', 'value': {'container_name': 'neutron_sriov_agent', 'image': 'registry.osism.tech/kolla/neutron-sriov-agent:2024.2', 'privileged': True, 'enabled': False, 'host_in_groups': False, 'volumes': ['/etc/kolla/neutron-sriov-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-sriov-nic-agent 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-mlnx-agent', 'value': {'container_name': 'neutron_mlnx_agent', 'image': 'registry.osism.tech/kolla/neutron-mlnx-agent:2024.2', 'enabled': False, 'host_in_groups': False, 'volumes': ['/etc/kolla/neutron-mlnx-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2024.2', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-eswitchd', 'value': {'container_name': 'neutron_eswitchd', 'image': 'registry.osism.tech/kolla/neutron-eswitchd:2024.2', 'privileged': True, 'enabled': False, 'host_in_groups': False, 'volumes': ['/etc/kolla/neutron-eswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/run/libvirt:/run/libvirt:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-openvswitch-agent', 'value': {'container_name': 'neutron_openvswitch_agent', 'image': 'registry.osism.tech/kolla/neutron-openvswitch-agent:2024.2', 'enabled': False, 'privileged': True, 'environment': {'KOLLA_LEGACY_IPTABLES': 'false'}, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-openvswitch-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-openvswitch-agent 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-metadata-agent', 'value': {'container_name': 'neutron_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'privileged': True, 'enabled': False, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-metadata-agent 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-linuxbridge-agent', 'value': {'container_name': 'neutron_linuxbridge_agent', 'image': 'registry.osism.tech/kolla/neutron-linuxbridge-agent:2024.2', 'privileged': True, 'enabled': False, 'environment': {'KOLLA_LEGACY_IPTABLES': 'false'}, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-linuxbridge-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-linuxbridge-agent 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'privileged': True, 'enabled': True, 'host_in_groups': False, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-dhcp-agent', 'value': {'container_name': 'neutron_dhcp_agent', 'image': 'registry.osism.tech/kolla/neutron-dhcp-agent:2024.2', 'privileged': True, 'enabled': False, 'group': 'neutron-dhcp-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-dhcp-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-dhcp-agent 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-bgp-dragent', 'value': {'container_name': 'neutron_bgp_dragent', 'image': 'registry.osism.tech/kolla/neutron-bgp-dragent:2024.2', 'privileged': True, 'enabled': False, 'group': 'neutron-bgp-dragent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-bgp-dragent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-bgp-dragent 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-infoblox-ipam-agent', 'value': {'container_name': 'neutron_infoblox_ipam_agent', 'image': 'registry.osism.tech/kolla/neutron-infoblox-ipam-agent:2024.2', 'privileged': True, 'enabled': False, 'group': 'neutron-infoblox-ipam-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-infoblox-ipam-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-l3-agent', 'value': {'container_name': 'neutron_l3_agent', 'image': 'registry.osism.tech/kolla/neutron-l3-agent:2024.2', 'privileged': True, 'enabled': False, 'environment': {'KOLLA_LEGACY_IPTABLES': 'false'}, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-l3-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', \"healthcheck_port 'neutron-l3-agent ' 5672\"], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2024.2', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-metering-agent', 'value': {'container_name': 'neutron_metering_agent', 'image': 'registry.osism.tech/kolla/neutron-metering-agent:2024.2', 'privileged': True, 'enabled': False, 'group': 'neutron-metering-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-metering-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-sriov-agent', 'value': {'container_name': 'neutron_sriov_agent', 'image': 'registry.osism.tech/kolla/neutron-sriov-agent:2024.2', 'privileged': True, 'enabled': False, 'host_in_groups': False, 'volumes': ['/etc/kolla/neutron-sriov-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-sriov-nic-agent 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-openvswitch-agent', 'value': {'container_name': 'neutron_openvswitch_agent', 'image': 'registry.osism.tech/kolla/neutron-openvswitch-agent:2024.2', 'enabled': False, 'privileged': True, 'environment': {'KOLLA_LEGACY_IPTABLES': 'false'}, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-openvswitch-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run/openvswitch:/run/openvswitch:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-openvswitch-agent 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'ironic-neutron-agent', 'value': {'container_name': 'ironic_neutron_agent', 'image': 'registry.osism.tech/kolla/ironic-neutron-agent:2024.2', 'privileged': False, 'enabled': False, 'group': 'ironic-neutron-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/ironic-neutron-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port ironic-neutron-agent 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-linuxbridge-agent', 'value': {'container_name': 'neutron_linuxbridge_agent', 'image': 'registry.osism.tech/kolla/neutron-linuxbridge-agent:2024.2', 'privileged': True, 'enabled': False, 'environment': {'KOLLA_LEGACY_IPTABLES': 'false'}, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-linuxbridge-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-linuxbridge-agent 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-mlnx-agent', 'value': {'container_name': 'neutron_mlnx_agent', 'image': 'registry.osism.tech/kolla/neutron-mlnx-agent:2024.2', 'enabled': False, 'host_in_groups': False, 'volumes': ['/etc/kolla/neutron-mlnx-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-tls-proxy', 'value': {'container_name': 'neutron_tls_proxy', 'group': 'neutron-server', 'host_in_groups': True, 'enabled': 'no', 'image': 'registry.osism.tech/kolla/neutron-tls-proxy:2024.2', 'volumes': ['/etc/kolla/neutron-tls-proxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl -u openstack:password 192.168.16.10:9697'], 'timeout': '30'}, 'haproxy': {'neutron_tls_proxy': {'enabled': False, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696', 'tls_backend': 'yes'}, 'neutron_tls_proxy_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696', 'tls_backend': 'yes'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-dhcp-agent', 'value': {'container_name': 'neutron_dhcp_agent', 'image': 'registry.osism.tech/kolla/neutron-dhcp-agent:2024.2', 'privileged': True, 'enabled': False, 'group': 'neutron-dhcp-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-dhcp-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-dhcp-agent 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-eswitchd', 'value': {'container_name': 'neutron_eswitchd', 'image': 'registry.osism.tech/kolla/neutron-eswitchd:2024.2', 'privileged': True, 'enabled': False, 'host_in_groups': False, 'volumes': ['/etc/kolla/neutron-eswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/run/libvirt:/run/libvirt:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-ovn-agent', 'value': {'container_name': 'neutron_ovn_agent', 'group': 'neutron-ovn-agent', 'host_in_groups': False, 'enabled': False, 'image': 'registry.osism.tech/dockerhub/kolla/neutron-ovn-agent:2024.2', 'volumes': ['/etc/kolla/neutron-ovn-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-agent 6640'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-l3-agent', 'value': {'container_name': 'neutron_l3_agent', 'image': 'registry.osism.tech/kolla/neutron-l3-agent:2024.2', 'privileged': True, 'enabled': False, 'environment': {'KOLLA_LEGACY_IPTABLES': 'false'}, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-l3-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', \"healthcheck_port 'neutron-l3-agent ' 5672\"], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-metadata-agent', 'value': {'container_name': 'neutron_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'privileged': True, 'enabled': False, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-metadata-agent 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-sriov-agent', 'value': {'container_name': 'neutron_sriov_agent', 'image': 'registry.osism.tech/kolla/neutron-sriov-agent:2024.2', 'privileged': True, 'enabled': False, 'host_in_groups': False, 'volumes': ['/etc/kolla/neutron-sriov-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-sriov-nic-agent 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'privileged': True, 'enabled': True, 'host_in_groups': False, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-mlnx-agent', 'value': {'container_name': 'neutron_mlnx_agent', 'image': 'registry.osism.tech/kolla/neutron-mlnx-agent:2024.2', 'enabled': False, 'host_in_groups': False, 'volumes': ['/etc/kolla/neutron-mlnx-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-bgp-dragent', 'value': {'container_name': 'neutron_bgp_dragent', 'image': 'registry.osism.tech/kolla/neutron-bgp-dragent:2024.2', 'privileged': True, 'enabled': False, 'group': 'neutron-bgp-dragent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-bgp-dragent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-bgp-dragent 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-eswitchd', 'value': {'container_name': 'neutron_eswitchd', 'image': 'registry.osism.tech/kolla/neutron-eswitchd:2024.2', 'privileged': True, 'enabled': False, 'host_in_groups': False, 'volumes': ['/etc/kolla/neutron-eswitchd/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/run/libvirt:/run/libvirt:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-infoblox-ipam-agent', 'value': {'container_name': 'neutron_infoblox_ipam_agent', 'image': 'registry.osism.tech/kolla/neutron-infoblox-ipam-agent:2024.2', 'privileged': True, 'enabled': False, 'group': 'neutron-infoblox-ipam-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-infoblox-ipam-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-metadata-agent', 'value': {'container_name': 'neutron_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'privileged': True, 'enabled': False, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-metadata-agent 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-metering-agent', 'value': {'container_name': 'neutron_metering_agent', 'image': 'registry.osism.tech/kolla/neutron-metering-agent:2024.2', 'privileged': True, 'enabled': False, 'group': 'neutron-metering-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-metering-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'ironic-neutron-agent', 'value': {'container_name': 'ironic_neutron_agent', 'image': 'registry.osism.tech/kolla/ironic-neutron-agent:2024.2', 'privileged': False, 'enabled': False, 'group': 'ironic-neutron-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/ironic-neutron-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port ironic-neutron-agent 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'privileged': True, 'enabled': True, 'host_in_groups': False, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-tls-proxy', 'value': {'container_name': 'neutron_tls_proxy', 'group': 'neutron-server', 'host_in_groups': True, 'enabled': 'no', 'image': 'registry.osism.tech/kolla/neutron-tls-proxy:2024.2', 'volumes': ['/etc/kolla/neutron-tls-proxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl -u openstack:password 192.168.16.11:9697'], 'timeout': '30'}, 'haproxy': {'neutron_tls_proxy': {'enabled': False, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696', 'tls_backend': 'yes'}, 'neutron_tls_proxy_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696', 'tls_backend': 'yes'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-bgp-dragent', 'value': {'container_name': 'neutron_bgp_dragent', 'image': 'registry.osism.tech/kolla/neutron-bgp-dragent:2024.2', 'privileged': True, 'enabled': False, 'group': 'neutron-bgp-dragent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-bgp-dragent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-bgp-dragent 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-ovn-agent', 'value': {'container_name': 'neutron_ovn_agent', 'group': 'neutron-ovn-agent', 'host_in_groups': False, 'enabled': False, 'image': 'registry.osism.tech/dockerhub/kolla/neutron-ovn-agent:2024.2', 'volumes': ['/etc/kolla/neutron-ovn-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-agent 6640'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-infoblox-ipam-agent', 'value': {'container_name': 'neutron_infoblox_ipam_agent', 'image': 'registry.osism.tech/kolla/neutron-infoblox-ipam-agent:2024.2', 'privileged': True, 'enabled': False, 'group': 'neutron-infoblox-ipam-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-infoblox-ipam-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-metering-agent', 'value': {'container_name': 'neutron_metering_agent', 'image': 'registry.osism.tech/kolla/neutron-metering-agent:2024.2', 'privileged': True, 'enabled': False, 'group': 'neutron-metering-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-metering-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'ironic-neutron-agent', 'value': {'container_name': 'ironic_neutron_agent', 'image': 'registry.osism.tech/kolla/ironic-neutron-agent:2024.2', 'privileged': False, 'enabled': False, 'group': 'ironic-neutron-agent', 'host_in_groups': True, 'volumes': ['/etc/kolla/ironic-neutron-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port ironic-neutron-agent 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-tls-proxy', 'value': {'container_name': 'neutron_tls_proxy', 'group': 'neutron-server', 'host_in_groups': True, 'enabled': 'no', 'image': 'registry.osism.tech/kolla/neutron-tls-proxy:2024.2', 'volumes': ['/etc/kolla/neutron-tls-proxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl -u openstack:password 192.168.16.12:9697'], 'timeout': '30'}, 'haproxy': {'neutron_tls_proxy': {'enabled': False, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696', 'tls_backend': 'yes'}, 'neutron_tls_proxy_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696', 'tls_backend': 'yes'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-ovn-agent', 'value': {'container_name': 'neutron_ovn_agent', 'group': 'neutron-ovn-agent', 'host_in_groups': False, 'enabled': False, 'image': 'registry.osism.tech/dockerhub/kolla/neutron-ovn-agent:2024.2', 'volumes': ['/etc/kolla/neutron-ovn-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-agent 6640'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [haproxy-config : Configuring firewall for neutron] ***********************",
                                "Tuesday 24 February 2026  00:53:11 +0000 (0:00:01.325)       0:03:36.478 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron_server', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron_server_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron_server', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron_server_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron_server', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron_server_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [proxysql-config : Copying over neutron ProxySQL users config] ************",
                                "Tuesday 24 February 2026  00:53:13 +0000 (0:00:01.750)       0:03:38.228 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [proxysql-config : Copying over neutron ProxySQL rules config] ************",
                                "Tuesday 24 February 2026  00:53:14 +0000 (0:00:01.397)       0:03:39.626 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [include_role : placement] ************************************************",
                                "Tuesday 24 February 2026  00:53:16 +0000 (0:00:02.086)       0:03:41.712 ****** ",
                                "\u001b[0;36mincluded: placement for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [haproxy-config : Copying over placement haproxy config] ******************",
                                "Tuesday 24 February 2026  00:53:18 +0000 (0:00:01.206)       0:03:42.919 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/placement-api:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8780'], 'timeout': '30'}, 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/placement-api:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8780'], 'timeout': '30'}, 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/placement-api:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8780'], 'timeout': '30'}, 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}}})\u001b[0m",
                                "",
                                "TASK [haproxy-config : Add configuration for placement when using single external frontend] ***",
                                "Tuesday 24 February 2026  00:53:21 +0000 (0:00:03.743)       0:03:46.663 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/placement-api:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8780'], 'timeout': '30'}, 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/placement-api:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8780'], 'timeout': '30'}, 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/placement-api:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8780'], 'timeout': '30'}, 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [haproxy-config : Configuring firewall for placement] *********************",
                                "Tuesday 24 February 2026  00:53:22 +0000 (0:00:00.551)       0:03:47.215 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'placement_api', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'placement_api_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'placement_api', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'placement_api_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'placement_api', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'placement_api_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [proxysql-config : Copying over placement ProxySQL users config] **********",
                                "Tuesday 24 February 2026  00:53:22 +0000 (0:00:00.682)       0:03:47.897 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [proxysql-config : Copying over placement ProxySQL rules config] **********",
                                "Tuesday 24 February 2026  00:53:24 +0000 (0:00:01.743)       0:03:49.640 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [include_role : nova] *****************************************************",
                                "Tuesday 24 February 2026  00:53:26 +0000 (0:00:01.873)       0:03:51.513 ****** ",
                                "\u001b[0;36mincluded: nova for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [haproxy-config : Copying over nova haproxy config] ***********************",
                                "Tuesday 24 February 2026  00:53:28 +0000 (0:00:01.578)       0:03:53.092 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/nova-api:2024.2', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8774 '], 'timeout': '30'}, 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/nova-scheduler:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova-super-conductor', 'value': {'container_name': 'nova_super_conductor', 'group': 'nova-super-conductor', 'enabled': 'no', 'image': 'registry.osism.tech/kolla/nova-super-conductor:2024.2', 'volumes': ['/etc/kolla/nova-super-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/nova-api:2024.2', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8774 '], 'timeout': '30'}, 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/nova-scheduler:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova-super-conductor', 'value': {'container_name': 'nova_super_conductor', 'group': 'nova-super-conductor', 'enabled': 'no', 'image': 'registry.osism.tech/kolla/nova-super-conductor:2024.2', 'volumes': ['/etc/kolla/nova-super-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/nova-api:2024.2', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8774 '], 'timeout': '30'}, 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/nova-scheduler:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova-super-conductor', 'value': {'container_name': 'nova_super_conductor', 'group': 'nova-super-conductor', 'enabled': 'no', 'image': 'registry.osism.tech/kolla/nova-super-conductor:2024.2', 'volumes': ['/etc/kolla/nova-super-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "",
                                "TASK [haproxy-config : Add configuration for nova when using single external frontend] ***",
                                "Tuesday 24 February 2026  00:53:32 +0000 (0:00:04.504)       0:03:57.596 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/nova-api:2024.2', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8774 '], 'timeout': '30'}, 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/nova-scheduler:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova-super-conductor', 'value': {'container_name': 'nova_super_conductor', 'group': 'nova-super-conductor', 'enabled': 'no', 'image': 'registry.osism.tech/kolla/nova-super-conductor:2024.2', 'volumes': ['/etc/kolla/nova-super-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/nova-api:2024.2', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8774 '], 'timeout': '30'}, 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/nova-scheduler:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova-super-conductor', 'value': {'container_name': 'nova_super_conductor', 'group': 'nova-super-conductor', 'enabled': 'no', 'image': 'registry.osism.tech/kolla/nova-super-conductor:2024.2', 'volumes': ['/etc/kolla/nova-super-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/nova-api:2024.2', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8774 '], 'timeout': '30'}, 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/nova-scheduler:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova-super-conductor', 'value': {'container_name': 'nova_super_conductor', 'group': 'nova-super-conductor', 'enabled': 'no', 'image': 'registry.osism.tech/kolla/nova-super-conductor:2024.2', 'volumes': ['/etc/kolla/nova-super-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [haproxy-config : Configuring firewall for nova] **************************",
                                "Tuesday 24 February 2026  00:53:33 +0000 (0:00:01.282)       0:03:58.879 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova_api', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova_api_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova_metadata', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova_metadata_external', 'value': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova_api', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova_api_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova_metadata', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova_metadata_external', 'value': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova_api', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova_api_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova_metadata', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova_metadata_external', 'value': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [proxysql-config : Copying over nova ProxySQL users config] ***************",
                                "Tuesday 24 February 2026  00:53:35 +0000 (0:00:01.082)       0:03:59.961 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [proxysql-config : Copying over nova ProxySQL rules config] ***************",
                                "Tuesday 24 February 2026  00:53:36 +0000 (0:00:01.694)       0:04:01.656 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [include_role : nova-cell] ************************************************",
                                "Tuesday 24 February 2026  00:53:39 +0000 (0:00:02.421)       0:04:04.078 ****** ",
                                "\u001b[0;36mincluded: nova-cell for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [nova-cell : Configure loadbalancer for nova-novncproxy] ******************",
                                "Tuesday 24 February 2026  00:53:40 +0000 (0:00:01.717)       0:04:05.795 ****** ",
                                "\u001b[0;36mincluded: /ansible/roles/nova-cell/tasks/cell_proxy_loadbalancer.yml for testbed-node-0, testbed-node-1, testbed-node-2 => (item=nova-novncproxy)\u001b[0m",
                                "",
                                "TASK [haproxy-config : Copying over nova-cell:nova-novncproxy haproxy config] ***",
                                "Tuesday 24 February 2026  00:53:41 +0000 (0:00:00.899)       0:04:06.695 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'nova-novncproxy', 'value': {'group': 'nova-novncproxy', 'enabled': True, 'haproxy': {'nova_novncproxy': {'enabled': True, 'mode': 'http', 'external': False, 'port': '6080', 'listen_port': '6080', 'backend_http_extra': ['timeout tunnel 1h']}, 'nova_novncproxy_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6080', 'listen_port': '6080', 'backend_http_extra': ['timeout tunnel 1h']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'nova-novncproxy', 'value': {'group': 'nova-novncproxy', 'enabled': True, 'haproxy': {'nova_novncproxy': {'enabled': True, 'mode': 'http', 'external': False, 'port': '6080', 'listen_port': '6080', 'backend_http_extra': ['timeout tunnel 1h']}, 'nova_novncproxy_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6080', 'listen_port': '6080', 'backend_http_extra': ['timeout tunnel 1h']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'nova-novncproxy', 'value': {'group': 'nova-novncproxy', 'enabled': True, 'haproxy': {'nova_novncproxy': {'enabled': True, 'mode': 'http', 'external': False, 'port': '6080', 'listen_port': '6080', 'backend_http_extra': ['timeout tunnel 1h']}, 'nova_novncproxy_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6080', 'listen_port': '6080', 'backend_http_extra': ['timeout tunnel 1h']}}}})\u001b[0m",
                                "",
                                "TASK [haproxy-config : Add configuration for nova-cell:nova-novncproxy when using single external frontend] ***",
                                "Tuesday 24 February 2026  00:53:46 +0000 (0:00:04.791)       0:04:11.487 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova-novncproxy', 'value': {'group': 'nova-novncproxy', 'enabled': True, 'haproxy': {'nova_novncproxy': {'enabled': True, 'mode': 'http', 'external': False, 'port': '6080', 'listen_port': '6080', 'backend_http_extra': ['timeout tunnel 1h']}, 'nova_novncproxy_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6080', 'listen_port': '6080', 'backend_http_extra': ['timeout tunnel 1h']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova-novncproxy', 'value': {'group': 'nova-novncproxy', 'enabled': True, 'haproxy': {'nova_novncproxy': {'enabled': True, 'mode': 'http', 'external': False, 'port': '6080', 'listen_port': '6080', 'backend_http_extra': ['timeout tunnel 1h']}, 'nova_novncproxy_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6080', 'listen_port': '6080', 'backend_http_extra': ['timeout tunnel 1h']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova-novncproxy', 'value': {'group': 'nova-novncproxy', 'enabled': True, 'haproxy': {'nova_novncproxy': {'enabled': True, 'mode': 'http', 'external': False, 'port': '6080', 'listen_port': '6080', 'backend_http_extra': ['timeout tunnel 1h']}, 'nova_novncproxy_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6080', 'listen_port': '6080', 'backend_http_extra': ['timeout tunnel 1h']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [haproxy-config : Configuring firewall for nova-cell:nova-novncproxy] *****",
                                "Tuesday 24 February 2026  00:53:47 +0000 (0:00:01.089)       0:04:12.576 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova_novncproxy', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '6080', 'listen_port': '6080', 'backend_http_extra': ['timeout tunnel 1h']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova_novncproxy_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6080', 'listen_port': '6080', 'backend_http_extra': ['timeout tunnel 1h']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova_novncproxy', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '6080', 'listen_port': '6080', 'backend_http_extra': ['timeout tunnel 1h']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova_novncproxy_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6080', 'listen_port': '6080', 'backend_http_extra': ['timeout tunnel 1h']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova_novncproxy', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '6080', 'listen_port': '6080', 'backend_http_extra': ['timeout tunnel 1h']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova_novncproxy_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6080', 'listen_port': '6080', 'backend_http_extra': ['timeout tunnel 1h']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [proxysql-config : Copying over nova-cell ProxySQL users config] **********",
                                "Tuesday 24 February 2026  00:53:49 +0000 (0:00:01.729)       0:04:14.306 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [proxysql-config : Copying over nova-cell ProxySQL rules config] **********",
                                "Tuesday 24 February 2026  00:53:52 +0000 (0:00:02.960)       0:04:17.267 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [nova-cell : Configure loadbalancer for nova-spicehtml5proxy] *************",
                                "Tuesday 24 February 2026  00:53:55 +0000 (0:00:03.438)       0:04:20.705 ****** ",
                                "\u001b[0;36mincluded: /ansible/roles/nova-cell/tasks/cell_proxy_loadbalancer.yml for testbed-node-0, testbed-node-1, testbed-node-2 => (item=nova-spicehtml5proxy)\u001b[0m",
                                "",
                                "TASK [haproxy-config : Copying over nova-cell:nova-spicehtml5proxy haproxy config] ***",
                                "Tuesday 24 February 2026  00:53:57 +0000 (0:00:01.461)       0:04:22.166 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova-spicehtml5proxy', 'value': {'group': 'nova-spicehtml5proxy', 'enabled': False, 'haproxy': {'nova_spicehtml5proxy': {'enabled': False, 'mode': 'http', 'external': False, 'port': '6082', 'listen_port': '6082', 'backend_http_extra': ['timeout tunnel 1h']}, 'nova_spicehtml5proxy_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6082', 'listen_port': '6082', 'backend_http_extra': ['timeout tunnel 1h']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova-spicehtml5proxy', 'value': {'group': 'nova-spicehtml5proxy', 'enabled': False, 'haproxy': {'nova_spicehtml5proxy': {'enabled': False, 'mode': 'http', 'external': False, 'port': '6082', 'listen_port': '6082', 'backend_http_extra': ['timeout tunnel 1h']}, 'nova_spicehtml5proxy_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6082', 'listen_port': '6082', 'backend_http_extra': ['timeout tunnel 1h']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova-spicehtml5proxy', 'value': {'group': 'nova-spicehtml5proxy', 'enabled': False, 'haproxy': {'nova_spicehtml5proxy': {'enabled': False, 'mode': 'http', 'external': False, 'port': '6082', 'listen_port': '6082', 'backend_http_extra': ['timeout tunnel 1h']}, 'nova_spicehtml5proxy_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6082', 'listen_port': '6082', 'backend_http_extra': ['timeout tunnel 1h']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [haproxy-config : Add configuration for nova-cell:nova-spicehtml5proxy when using single external frontend] ***",
                                "Tuesday 24 February 2026  00:53:58 +0000 (0:00:01.400)       0:04:23.567 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova-spicehtml5proxy', 'value': {'group': 'nova-spicehtml5proxy', 'enabled': False, 'haproxy': {'nova_spicehtml5proxy': {'enabled': False, 'mode': 'http', 'external': False, 'port': '6082', 'listen_port': '6082', 'backend_http_extra': ['timeout tunnel 1h']}, 'nova_spicehtml5proxy_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6082', 'listen_port': '6082', 'backend_http_extra': ['timeout tunnel 1h']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova-spicehtml5proxy', 'value': {'group': 'nova-spicehtml5proxy', 'enabled': False, 'haproxy': {'nova_spicehtml5proxy': {'enabled': False, 'mode': 'http', 'external': False, 'port': '6082', 'listen_port': '6082', 'backend_http_extra': ['timeout tunnel 1h']}, 'nova_spicehtml5proxy_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6082', 'listen_port': '6082', 'backend_http_extra': ['timeout tunnel 1h']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova-spicehtml5proxy', 'value': {'group': 'nova-spicehtml5proxy', 'enabled': False, 'haproxy': {'nova_spicehtml5proxy': {'enabled': False, 'mode': 'http', 'external': False, 'port': '6082', 'listen_port': '6082', 'backend_http_extra': ['timeout tunnel 1h']}, 'nova_spicehtml5proxy_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6082', 'listen_port': '6082', 'backend_http_extra': ['timeout tunnel 1h']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [haproxy-config : Configuring firewall for nova-cell:nova-spicehtml5proxy] ***",
                                "Tuesday 24 February 2026  00:54:00 +0000 (0:00:01.385)       0:04:24.952 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [proxysql-config : Copying over nova-cell ProxySQL users config] **********",
                                "Tuesday 24 February 2026  00:54:02 +0000 (0:00:02.008)       0:04:26.961 ****** ",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [proxysql-config : Copying over nova-cell ProxySQL rules config] **********",
                                "Tuesday 24 February 2026  00:54:04 +0000 (0:00:02.605)       0:04:29.567 ****** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [nova-cell : Configure loadbalancer for nova-serialproxy] *****************",
                                "Tuesday 24 February 2026  00:54:08 +0000 (0:00:03.427)       0:04:32.995 ****** ",
                                "\u001b[0;36mincluded: /ansible/roles/nova-cell/tasks/cell_proxy_loadbalancer.yml for testbed-node-0, testbed-node-1, testbed-node-2 => (item=nova-serialproxy)\u001b[0m",
                                "",
                                "TASK [haproxy-config : Copying over nova-cell:nova-serialproxy haproxy config] ***",
                                "Tuesday 24 February 2026  00:54:09 +0000 (0:00:00.953)       0:04:33.948 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova-serialproxy', 'value': {'group': 'nova-serialproxy', 'enabled': False, 'haproxy': {'nova_serialconsole_proxy': {'enabled': False, 'mode': 'http', 'external': False, 'port': '6083', 'listen_port': '6083', 'backend_http_extra': ['timeout tunnel 10m']}, 'nova_serialconsole_proxy_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6083', 'listen_port': '6083', 'backend_http_extra': ['timeout tunnel 10m']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova-serialproxy', 'value': {'group': 'nova-serialproxy', 'enabled': False, 'haproxy': {'nova_serialconsole_proxy': {'enabled': False, 'mode': 'http', 'external': False, 'port': '6083', 'listen_port': '6083', 'backend_http_extra': ['timeout tunnel 10m']}, 'nova_serialconsole_proxy_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6083', 'listen_port': '6083', 'backend_http_extra': ['timeout tunnel 10m']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova-serialproxy', 'value': {'group': 'nova-serialproxy', 'enabled': False, 'haproxy': {'nova_serialconsole_proxy': {'enabled': False, 'mode': 'http', 'external': False, 'port': '6083', 'listen_port': '6083', 'backend_http_extra': ['timeout tunnel 10m']}, 'nova_serialconsole_proxy_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6083', 'listen_port': '6083', 'backend_http_extra': ['timeout tunnel 10m']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [haproxy-config : Add configuration for nova-cell:nova-serialproxy when using single external frontend] ***",
                                "Tuesday 24 February 2026  00:54:10 +0000 (0:00:01.553)       0:04:35.502 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova-serialproxy', 'value': {'group': 'nova-serialproxy', 'enabled': False, 'haproxy': {'nova_serialconsole_proxy': {'enabled': False, 'mode': 'http', 'external': False, 'port': '6083', 'listen_port': '6083', 'backend_http_extra': ['timeout tunnel 10m']}, 'nova_serialconsole_proxy_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6083', 'listen_port': '6083', 'backend_http_extra': ['timeout tunnel 10m']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova-serialproxy', 'value': {'group': 'nova-serialproxy', 'enabled': False, 'haproxy': {'nova_serialconsole_proxy': {'enabled': False, 'mode': 'http', 'external': False, 'port': '6083', 'listen_port': '6083', 'backend_http_extra': ['timeout tunnel 10m']}, 'nova_serialconsole_proxy_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6083', 'listen_port': '6083', 'backend_http_extra': ['timeout tunnel 10m']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova-serialproxy', 'value': {'group': 'nova-serialproxy', 'enabled': False, 'haproxy': {'nova_serialconsole_proxy': {'enabled': False, 'mode': 'http', 'external': False, 'port': '6083', 'listen_port': '6083', 'backend_http_extra': ['timeout tunnel 10m']}, 'nova_serialconsole_proxy_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '6083', 'listen_port': '6083', 'backend_http_extra': ['timeout tunnel 10m']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [haproxy-config : Configuring firewall for nova-cell:nova-serialproxy] ****",
                                "Tuesday 24 February 2026  00:54:12 +0000 (0:00:01.852)       0:04:37.355 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [proxysql-config : Copying over nova-cell ProxySQL users config] **********",
                                "Tuesday 24 February 2026  00:54:14 +0000 (0:00:01.635)       0:04:38.990 ****** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [proxysql-config : Copying over nova-cell ProxySQL rules config] **********",
                                "Tuesday 24 February 2026  00:54:16 +0000 (0:00:02.429)       0:04:41.420 ****** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [include_role : octavia] **************************************************",
                                "Tuesday 24 February 2026  00:54:19 +0000 (0:00:03.447)       0:04:44.867 ****** ",
                                "\u001b[0;36mincluded: octavia for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [haproxy-config : Copying over octavia haproxy config] ********************",
                                "Tuesday 24 February 2026  00:54:21 +0000 (0:00:01.715)       0:04:46.583 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-api:2024.2', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-driver-agent:2024.2', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-health-manager:2024.2', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-housekeeping:2024.2', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-worker:2024.2', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-api:2024.2', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-driver-agent:2024.2', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-health-manager:2024.2', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-housekeeping:2024.2', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-worker:2024.2', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-api:2024.2', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-driver-agent:2024.2', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-health-manager:2024.2', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-housekeeping:2024.2', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-worker:2024.2', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "",
                                "TASK [haproxy-config : Add configuration for octavia when using single external frontend] ***",
                                "Tuesday 24 February 2026  00:54:25 +0000 (0:00:03.854)       0:04:50.437 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-api:2024.2', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-driver-agent:2024.2', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-health-manager:2024.2', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-housekeeping:2024.2', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-api:2024.2', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-worker:2024.2', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-driver-agent:2024.2', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-health-manager:2024.2', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-housekeeping:2024.2', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-worker:2024.2', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-api:2024.2', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-driver-agent:2024.2', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-health-manager:2024.2', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-housekeeping:2024.2', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-worker:2024.2', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [haproxy-config : Configuring firewall for octavia] ***********************",
                                "Tuesday 24 February 2026  00:54:26 +0000 (0:00:00.814)       0:04:51.252 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'octavia_api', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'octavia_api_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'octavia_api', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'octavia_api_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'octavia_api', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'octavia_api_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [proxysql-config : Copying over octavia ProxySQL users config] ************",
                                "Tuesday 24 February 2026  00:54:28 +0000 (0:00:01.860)       0:04:53.113 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [proxysql-config : Copying over octavia ProxySQL rules config] ************",
                                "Tuesday 24 February 2026  00:54:29 +0000 (0:00:01.495)       0:04:54.609 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [include_role : opensearch] ***********************************************",
                                "Tuesday 24 February 2026  00:54:32 +0000 (0:00:02.349)       0:04:56.958 ****** ",
                                "\u001b[0;36mincluded: opensearch for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [haproxy-config : Copying over opensearch haproxy config] *****************",
                                "Tuesday 24 February 2026  00:54:33 +0000 (0:00:01.715)       0:04:58.673 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/opensearch:2024.2', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/opensearch:2024.2', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/opensearch:2024.2', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/opensearch-dashboards:2024.2', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/opensearch-dashboards:2024.2', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/opensearch-dashboards:2024.2', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}}}})\u001b[0m",
                                "",
                                "TASK [haproxy-config : Add configuration for opensearch when using single external frontend] ***",
                                "Tuesday 24 February 2026  00:54:39 +0000 (0:00:06.124)       0:05:04.798 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/opensearch:2024.2', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/opensearch-dashboards:2024.2', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/opensearch:2024.2', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/opensearch-dashboards:2024.2', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/opensearch:2024.2', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/opensearch-dashboards:2024.2', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [haproxy-config : Configuring firewall for opensearch] ********************",
                                "Tuesday 24 February 2026  00:54:40 +0000 (0:00:00.757)       0:05:05.556 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'opensearch', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'opensearch-dashboards', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'opensearch_dashboards_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'opensearch', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'opensearch-dashboards', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'opensearch_dashboards_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'opensearch', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal']}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'opensearch-dashboards', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'opensearch_dashboards_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [proxysql-config : Copying over opensearch ProxySQL users config] *********",
                                "Tuesday 24 February 2026  00:54:41 +0000 (0:00:00.904)       0:05:06.461 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [proxysql-config : Copying over opensearch ProxySQL rules config] *********",
                                "Tuesday 24 February 2026  00:54:42 +0000 (0:00:00.935)       0:05:07.396 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [include_role : prometheus] ***********************************************",
                                "Tuesday 24 February 2026  00:54:43 +0000 (0:00:01.494)       0:05:08.890 ****** ",
                                "\u001b[0;36mincluded: prometheus for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [haproxy-config : Copying over prometheus haproxy config] *****************",
                                "Tuesday 24 February 2026  00:54:45 +0000 (0:00:01.528)       0:05:10.419 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'prometheus-server', 'value': {'container_name': 'prometheus_server', 'group': 'prometheus', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-v2-server:2024.2', 'volumes': ['/etc/kolla/prometheus-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'prometheus_v2:/var/lib/prometheus', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'prometheus_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9091', 'active_passive': True}, 'prometheus_server_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9091', 'listen_port': '9091', 'active_passive': True}}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2024.2', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-mysqld-exporter', 'value': {'container_name': 'prometheus_mysqld_exporter', 'group': 'prometheus-mysqld-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-mysqld-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-mysqld-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-memcached-exporter', 'value': {'container_name': 'prometheus_memcached_exporter', 'group': 'prometheus-memcached-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-memcached-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-memcached-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'prometheus-server', 'value': {'container_name': 'prometheus_server', 'group': 'prometheus', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-v2-server:2024.2', 'volumes': ['/etc/kolla/prometheus-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'prometheus_v2:/var/lib/prometheus', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'prometheus_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9091', 'active_passive': True}, 'prometheus_server_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9091', 'listen_port': '9091', 'active_passive': True}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'prometheus-server', 'value': {'container_name': 'prometheus_server', 'group': 'prometheus', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-v2-server:2024.2', 'volumes': ['/etc/kolla/prometheus-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'prometheus_v2:/var/lib/prometheus', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'prometheus_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9091', 'active_passive': True}, 'prometheus_server_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9091', 'listen_port': '9091', 'active_passive': True}}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2024.2', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2024.2', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2024.2', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-mysqld-exporter', 'value': {'container_name': 'prometheus_mysqld_exporter', 'group': 'prometheus-mysqld-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-mysqld-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-mysqld-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-memcached-exporter', 'value': {'container_name': 'prometheus_memcached_exporter', 'group': 'prometheus-memcached-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-memcached-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-memcached-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-mysqld-exporter', 'value': {'container_name': 'prometheus_mysqld_exporter', 'group': 'prometheus-mysqld-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-mysqld-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-mysqld-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2024.2', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-memcached-exporter', 'value': {'container_name': 'prometheus_memcached_exporter', 'group': 'prometheus-memcached-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-memcached-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-memcached-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2024.2', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'prometheus-alertmanager', 'value': {'container_name': 'prometheus_alertmanager', 'group': 'prometheus-alertmanager', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-alertmanager:2024.2', 'volumes': ['/etc/kolla/prometheus-alertmanager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'prometheus:/var/lib/prometheus'], 'dimensions': {}, 'haproxy': {'prometheus_alertmanager': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True}, 'prometheus_alertmanager_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9093', 'listen_port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True}}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-openstack-exporter', 'value': {'container_name': 'prometheus_openstack_exporter', 'group': 'prometheus-openstack-exporter', 'enabled': False, 'environment': {'OS_COMPUTE_API_VERSION': 'latest'}, 'image': 'registry.osism.tech/kolla/prometheus-openstack-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-openstack-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'prometheus_openstack_exporter': {'enabled': False, 'mode': 'http', 'external': False, 'port': '9198', 'backend_http_extra': ['timeout server 45s']}, 'prometheus_openstack_exporter_external': {'enabled': False, 'mode': 'http', 'external': True, 'port': '9198', 'backend_http_extra': ['timeout server 45s']}}}}) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'prometheus-alertmanager', 'value': {'container_name': 'prometheus_alertmanager', 'group': 'prometheus-alertmanager', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-alertmanager:2024.2', 'volumes': ['/etc/kolla/prometheus-alertmanager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'prometheus:/var/lib/prometheus'], 'dimensions': {}, 'haproxy': {'prometheus_alertmanager': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True}, 'prometheus_alertmanager_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9093', 'listen_port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True}}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-elasticsearch-exporter', 'value': {'container_name': 'prometheus_elasticsearch_exporter', 'group': 'prometheus-elasticsearch-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-elasticsearch-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-elasticsearch-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-blackbox-exporter', 'value': {'container_name': 'prometheus_blackbox_exporter', 'group': 'prometheus-blackbox-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-blackbox-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-blackbox-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-openstack-exporter', 'value': {'container_name': 'prometheus_openstack_exporter', 'group': 'prometheus-openstack-exporter', 'enabled': False, 'environment': {'OS_COMPUTE_API_VERSION': 'latest'}, 'image': 'registry.osism.tech/kolla/prometheus-openstack-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-openstack-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'prometheus_openstack_exporter': {'enabled': False, 'mode': 'http', 'external': False, 'port': '9198', 'backend_http_extra': ['timeout server 45s']}, 'prometheus_openstack_exporter_external': {'enabled': False, 'mode': 'http', 'external': True, 'port': '9198', 'backend_http_extra': ['timeout server 45s']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-libvirt-exporter', 'value': {'container_name': 'prometheus_libvirt_exporter', 'group': 'prometheus-libvirt-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-libvirt-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-libvirt-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run/libvirt:/run/libvirt:ro'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-elasticsearch-exporter', 'value': {'container_name': 'prometheus_elasticsearch_exporter', 'group': 'prometheus-elasticsearch-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-elasticsearch-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-elasticsearch-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-blackbox-exporter', 'value': {'container_name': 'prometheus_blackbox_exporter', 'group': 'prometheus-blackbox-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-blackbox-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-blackbox-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-libvirt-exporter', 'value': {'container_name': 'prometheus_libvirt_exporter', 'group': 'prometheus-libvirt-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-libvirt-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-libvirt-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run/libvirt:/run/libvirt:ro'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'prometheus-alertmanager', 'value': {'container_name': 'prometheus_alertmanager', 'group': 'prometheus-alertmanager', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-alertmanager:2024.2', 'volumes': ['/etc/kolla/prometheus-alertmanager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'prometheus:/var/lib/prometheus'], 'dimensions': {}, 'haproxy': {'prometheus_alertmanager': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True}, 'prometheus_alertmanager_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9093', 'listen_port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True}}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-openstack-exporter', 'value': {'container_name': 'prometheus_openstack_exporter', 'group': 'prometheus-openstack-exporter', 'enabled': False, 'environment': {'OS_COMPUTE_API_VERSION': 'latest'}, 'image': 'registry.osism.tech/kolla/prometheus-openstack-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-openstack-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'prometheus_openstack_exporter': {'enabled': False, 'mode': 'http', 'external': False, 'port': '9198', 'backend_http_extra': ['timeout server 45s']}, 'prometheus_openstack_exporter_external': {'enabled': False, 'mode': 'http', 'external': True, 'port': '9198', 'backend_http_extra': ['timeout server 45s']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-elasticsearch-exporter', 'value': {'container_name': 'prometheus_elasticsearch_exporter', 'group': 'prometheus-elasticsearch-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-elasticsearch-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-elasticsearch-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-blackbox-exporter', 'value': {'container_name': 'prometheus_blackbox_exporter', 'group': 'prometheus-blackbox-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-blackbox-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-blackbox-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-libvirt-exporter', 'value': {'container_name': 'prometheus_libvirt_exporter', 'group': 'prometheus-libvirt-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-libvirt-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-libvirt-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run/libvirt:/run/libvirt:ro'], 'dimensions': {}}}) \u001b[0m",
                                "",
                                "TASK [haproxy-config : Add configuration for prometheus when using single external frontend] ***",
                                "Tuesday 24 February 2026  00:54:50 +0000 (0:00:05.244)       0:05:15.664 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-server', 'value': {'container_name': 'prometheus_server', 'group': 'prometheus', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-v2-server:2024.2', 'volumes': ['/etc/kolla/prometheus-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'prometheus_v2:/var/lib/prometheus', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'prometheus_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9091', 'active_passive': True}, 'prometheus_server_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9091', 'listen_port': '9091', 'active_passive': True}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2024.2', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-mysqld-exporter', 'value': {'container_name': 'prometheus_mysqld_exporter', 'group': 'prometheus-mysqld-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-mysqld-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-mysqld-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-memcached-exporter', 'value': {'container_name': 'prometheus_memcached_exporter', 'group': 'prometheus-memcached-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-memcached-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-memcached-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2024.2', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-alertmanager', 'value': {'container_name': 'prometheus_alertmanager', 'group': 'prometheus-alertmanager', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-alertmanager:2024.2', 'volumes': ['/etc/kolla/prometheus-alertmanager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'prometheus:/var/lib/prometheus'], 'dimensions': {}, 'haproxy': {'prometheus_alertmanager': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True}, 'prometheus_alertmanager_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9093', 'listen_port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-openstack-exporter', 'value': {'container_name': 'prometheus_openstack_exporter', 'group': 'prometheus-openstack-exporter', 'enabled': False, 'environment': {'OS_COMPUTE_API_VERSION': 'latest'}, 'image': 'registry.osism.tech/kolla/prometheus-openstack-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-openstack-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'prometheus_openstack_exporter': {'enabled': False, 'mode': 'http', 'external': False, 'port': '9198', 'backend_http_extra': ['timeout server 45s']}, 'prometheus_openstack_exporter_external': {'enabled': False, 'mode': 'http', 'external': True, 'port': '9198', 'backend_http_extra': ['timeout server 45s']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-elasticsearch-exporter', 'value': {'container_name': 'prometheus_elasticsearch_exporter', 'group': 'prometheus-elasticsearch-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-elasticsearch-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-elasticsearch-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-blackbox-exporter', 'value': {'container_name': 'prometheus_blackbox_exporter', 'group': 'prometheus-blackbox-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-blackbox-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-blackbox-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-libvirt-exporter', 'value': {'container_name': 'prometheus_libvirt_exporter', 'group': 'prometheus-libvirt-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-libvirt-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-libvirt-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run/libvirt:/run/libvirt:ro'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-server', 'value': {'container_name': 'prometheus_server', 'group': 'prometheus', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-v2-server:2024.2', 'volumes': ['/etc/kolla/prometheus-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'prometheus_v2:/var/lib/prometheus', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'prometheus_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9091', 'active_passive': True}, 'prometheus_server_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9091', 'listen_port': '9091', 'active_passive': True}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2024.2', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-mysqld-exporter', 'value': {'container_name': 'prometheus_mysqld_exporter', 'group': 'prometheus-mysqld-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-mysqld-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-mysqld-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-memcached-exporter', 'value': {'container_name': 'prometheus_memcached_exporter', 'group': 'prometheus-memcached-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-memcached-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-memcached-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2024.2', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-alertmanager', 'value': {'container_name': 'prometheus_alertmanager', 'group': 'prometheus-alertmanager', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-alertmanager:2024.2', 'volumes': ['/etc/kolla/prometheus-alertmanager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'prometheus:/var/lib/prometheus'], 'dimensions': {}, 'haproxy': {'prometheus_alertmanager': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True}, 'prometheus_alertmanager_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9093', 'listen_port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-openstack-exporter', 'value': {'container_name': 'prometheus_openstack_exporter', 'group': 'prometheus-openstack-exporter', 'enabled': False, 'environment': {'OS_COMPUTE_API_VERSION': 'latest'}, 'image': 'registry.osism.tech/kolla/prometheus-openstack-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-openstack-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'prometheus_openstack_exporter': {'enabled': False, 'mode': 'http', 'external': False, 'port': '9198', 'backend_http_extra': ['timeout server 45s']}, 'prometheus_openstack_exporter_external': {'enabled': False, 'mode': 'http', 'external': True, 'port': '9198', 'backend_http_extra': ['timeout server 45s']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-elasticsearch-exporter', 'value': {'container_name': 'prometheus_elasticsearch_exporter', 'group': 'prometheus-elasticsearch-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-elasticsearch-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-elasticsearch-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-blackbox-exporter', 'value': {'container_name': 'prometheus_blackbox_exporter', 'group': 'prometheus-blackbox-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-blackbox-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-blackbox-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-libvirt-exporter', 'value': {'container_name': 'prometheus_libvirt_exporter', 'group': 'prometheus-libvirt-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-libvirt-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-libvirt-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run/libvirt:/run/libvirt:ro'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-server', 'value': {'container_name': 'prometheus_server', 'group': 'prometheus', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-v2-server:2024.2', 'volumes': ['/etc/kolla/prometheus-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'prometheus_v2:/var/lib/prometheus', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'prometheus_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9091', 'active_passive': True}, 'prometheus_server_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9091', 'listen_port': '9091', 'active_passive': True}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2024.2', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-mysqld-exporter', 'value': {'container_name': 'prometheus_mysqld_exporter', 'group': 'prometheus-mysqld-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-mysqld-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-mysqld-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-memcached-exporter', 'value': {'container_name': 'prometheus_memcached_exporter', 'group': 'prometheus-memcached-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-memcached-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-memcached-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2024.2', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-alertmanager', 'value': {'container_name': 'prometheus_alertmanager', 'group': 'prometheus-alertmanager', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-alertmanager:2024.2', 'volumes': ['/etc/kolla/prometheus-alertmanager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'prometheus:/var/lib/prometheus'], 'dimensions': {}, 'haproxy': {'prometheus_alertmanager': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True}, 'prometheus_alertmanager_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9093', 'listen_port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-openstack-exporter', 'value': {'container_name': 'prometheus_openstack_exporter', 'group': 'prometheus-openstack-exporter', 'enabled': False, 'environment': {'OS_COMPUTE_API_VERSION': 'latest'}, 'image': 'registry.osism.tech/kolla/prometheus-openstack-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-openstack-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'prometheus_openstack_exporter': {'enabled': False, 'mode': 'http', 'external': False, 'port': '9198', 'backend_http_extra': ['timeout server 45s']}, 'prometheus_openstack_exporter_external': {'enabled': False, 'mode': 'http', 'external': True, 'port': '9198', 'backend_http_extra': ['timeout server 45s']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-elasticsearch-exporter', 'value': {'container_name': 'prometheus_elasticsearch_exporter', 'group': 'prometheus-elasticsearch-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-elasticsearch-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-elasticsearch-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-blackbox-exporter', 'value': {'container_name': 'prometheus_blackbox_exporter', 'group': 'prometheus-blackbox-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-blackbox-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-blackbox-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-libvirt-exporter', 'value': {'container_name': 'prometheus_libvirt_exporter', 'group': 'prometheus-libvirt-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-libvirt-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-libvirt-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run/libvirt:/run/libvirt:ro'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [haproxy-config : Configuring firewall for prometheus] ********************",
                                "Tuesday 24 February 2026  00:54:51 +0000 (0:00:01.064)       0:05:16.728 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus_server', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9091', 'active_passive': True}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus_server_external', 'value': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9091', 'listen_port': '9091', 'active_passive': True}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus_alertmanager', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus_alertmanager_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9093', 'listen_port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus_server', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9091', 'active_passive': True}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus_server_external', 'value': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9091', 'listen_port': '9091', 'active_passive': True}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus_alertmanager', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus_alertmanager_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9093', 'listen_port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus_server', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9091', 'active_passive': True}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus_server_external', 'value': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9091', 'listen_port': '9091', 'active_passive': True}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus_alertmanager', 'value': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus_alertmanager_external', 'value': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9093', 'listen_port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [proxysql-config : Copying over prometheus ProxySQL users config] *********",
                                "Tuesday 24 February 2026  00:54:52 +0000 (0:00:01.055)       0:05:17.784 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [proxysql-config : Copying over prometheus ProxySQL rules config] *********",
                                "Tuesday 24 February 2026  00:54:53 +0000 (0:00:00.448)       0:05:18.232 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [include_role : rabbitmq] *************************************************",
                                "Tuesday 24 February 2026  00:54:54 +0000 (0:00:01.527)       0:05:19.760 ****** ",
                                "\u001b[0;36mincluded: rabbitmq for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [haproxy-config : Copying over rabbitmq haproxy config] *******************",
                                "Tuesday 24 February 2026  00:54:56 +0000 (0:00:01.910)       0:05:21.670 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'rabbitmq', 'value': {'container_name': 'rabbitmq', 'group': None, 'enabled': True, 'image': 'registry.osism.tech/kolla/rabbitmq:2024.2', 'bootstrap_environment': {'KOLLA_BOOTSTRAP': None, 'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': None, 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': None, 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'volumes': ['/etc/kolla/rabbitmq/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'rabbitmq:/var/lib/rabbitmq/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_rabbitmq'], 'timeout': '30'}, 'haproxy': {'rabbitmq_management': {'enabled': 'yes', 'mode': 'http', 'port': '15672', 'host_group': 'rabbitmq'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'rabbitmq', 'value': {'container_name': 'rabbitmq', 'group': None, 'enabled': True, 'image': 'registry.osism.tech/kolla/rabbitmq:2024.2', 'bootstrap_environment': {'KOLLA_BOOTSTRAP': None, 'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': None, 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': None, 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'volumes': ['/etc/kolla/rabbitmq/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'rabbitmq:/var/lib/rabbitmq/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_rabbitmq'], 'timeout': '30'}, 'haproxy': {'rabbitmq_management': {'enabled': 'yes', 'mode': 'http', 'port': '15672', 'host_group': 'rabbitmq'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'rabbitmq', 'value': {'container_name': 'rabbitmq', 'group': None, 'enabled': True, 'image': 'registry.osism.tech/kolla/rabbitmq:2024.2', 'bootstrap_environment': {'KOLLA_BOOTSTRAP': None, 'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': None, 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': None, 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'volumes': ['/etc/kolla/rabbitmq/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'rabbitmq:/var/lib/rabbitmq/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_rabbitmq'], 'timeout': '30'}, 'haproxy': {'rabbitmq_management': {'enabled': 'yes', 'mode': 'http', 'port': '15672', 'host_group': 'rabbitmq'}}}})\u001b[0m",
                                "",
                                "TASK [haproxy-config : Add configuration for rabbitmq when using single external frontend] ***",
                                "Tuesday 24 February 2026  00:54:59 +0000 (0:00:02.644)       0:05:24.315 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'rabbitmq', 'value': {'container_name': 'rabbitmq', 'group': None, 'enabled': True, 'image': 'registry.osism.tech/kolla/rabbitmq:2024.2', 'bootstrap_environment': {'KOLLA_BOOTSTRAP': None, 'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': None, 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': None, 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'volumes': ['/etc/kolla/rabbitmq/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'rabbitmq:/var/lib/rabbitmq/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_rabbitmq'], 'timeout': '30'}, 'haproxy': {'rabbitmq_management': {'enabled': 'yes', 'mode': 'http', 'port': '15672', 'host_group': 'rabbitmq'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'rabbitmq', 'value': {'container_name': 'rabbitmq', 'group': None, 'enabled': True, 'image': 'registry.osism.tech/kolla/rabbitmq:2024.2', 'bootstrap_environment': {'KOLLA_BOOTSTRAP': None, 'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': None, 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': None, 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'volumes': ['/etc/kolla/rabbitmq/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'rabbitmq:/var/lib/rabbitmq/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_rabbitmq'], 'timeout': '30'}, 'haproxy': {'rabbitmq_management': {'enabled': 'yes', 'mode': 'http', 'port': '15672', 'host_group': 'rabbitmq'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'rabbitmq', 'value': {'container_name': 'rabbitmq', 'group': None, 'enabled': True, 'image': 'registry.osism.tech/kolla/rabbitmq:2024.2', 'bootstrap_environment': {'KOLLA_BOOTSTRAP': None, 'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': None, 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'environment': {'KOLLA_CONFIG_STRATEGY': 'COPY_ALWAYS', 'RABBITMQ_CLUSTER_COOKIE': None, 'RABBITMQ_LOG_DIR': '/var/log/kolla/rabbitmq'}, 'volumes': ['/etc/kolla/rabbitmq/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'rabbitmq:/var/lib/rabbitmq/', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_rabbitmq'], 'timeout': '30'}, 'haproxy': {'rabbitmq_management': {'enabled': 'yes', 'mode': 'http', 'port': '15672', 'host_group': 'rabbitmq'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [haproxy-config : Configuring firewall for rabbitmq] **********************",
                                "Tuesday 24 February 2026  00:55:00 +0000 (0:00:00.969)       0:05:25.285 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'rabbitmq_management', 'value': {'enabled': 'yes', 'mode': 'http', 'port': '15672', 'host_group': 'rabbitmq'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'rabbitmq_management', 'value': {'enabled': 'yes', 'mode': 'http', 'port': '15672', 'host_group': 'rabbitmq'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'rabbitmq_management', 'value': {'enabled': 'yes', 'mode': 'http', 'port': '15672', 'host_group': 'rabbitmq'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [proxysql-config : Copying over rabbitmq ProxySQL users config] ***********",
                                "Tuesday 24 February 2026  00:55:01 +0000 (0:00:00.773)       0:05:26.058 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [proxysql-config : Copying over rabbitmq ProxySQL rules config] ***********",
                                "Tuesday 24 February 2026  00:55:01 +0000 (0:00:00.547)       0:05:26.606 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [include_role : skyline] **************************************************",
                                "Tuesday 24 February 2026  00:55:03 +0000 (0:00:01.474)       0:05:28.080 ****** ",
                                "\u001b[0;36mincluded: skyline for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [haproxy-config : Copying over skyline haproxy config] ********************",
                                "Tuesday 24 February 2026  00:55:05 +0000 (0:00:02.040)       0:05:30.121 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'skyline-apiserver', 'value': {'container_name': 'skyline_apiserver', 'group': 'skyline-apiserver', 'enabled': True, 'image': 'registry.osism.tech/kolla/skyline-apiserver:2024.2', 'volumes': ['/etc/kolla/skyline-apiserver/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9998/docs'], 'timeout': '30'}, 'haproxy': {'skyline_apiserver': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no'}, 'skyline_apiserver_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'skyline-apiserver', 'value': {'container_name': 'skyline_apiserver', 'group': 'skyline-apiserver', 'enabled': True, 'image': 'registry.osism.tech/kolla/skyline-apiserver:2024.2', 'volumes': ['/etc/kolla/skyline-apiserver/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9998/docs'], 'timeout': '30'}, 'haproxy': {'skyline_apiserver': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no'}, 'skyline_apiserver_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'skyline-apiserver', 'value': {'container_name': 'skyline_apiserver', 'group': 'skyline-apiserver', 'enabled': True, 'image': 'registry.osism.tech/kolla/skyline-apiserver:2024.2', 'volumes': ['/etc/kolla/skyline-apiserver/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9998/docs'], 'timeout': '30'}, 'haproxy': {'skyline_apiserver': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no'}, 'skyline_apiserver_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'skyline-console', 'value': {'container_name': 'skyline_console', 'group': 'skyline-console', 'enabled': True, 'image': 'registry.osism.tech/kolla/skyline-console:2024.2', 'volumes': ['/etc/kolla/skyline-console/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9999/docs'], 'timeout': '30'}, 'haproxy': {'skyline_console': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no'}, 'skyline_console_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'skyline-console', 'value': {'container_name': 'skyline_console', 'group': 'skyline-console', 'enabled': True, 'image': 'registry.osism.tech/kolla/skyline-console:2024.2', 'volumes': ['/etc/kolla/skyline-console/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9999/docs'], 'timeout': '30'}, 'haproxy': {'skyline_console': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no'}, 'skyline_console_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'skyline-console', 'value': {'container_name': 'skyline_console', 'group': 'skyline-console', 'enabled': True, 'image': 'registry.osism.tech/kolla/skyline-console:2024.2', 'volumes': ['/etc/kolla/skyline-console/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9999/docs'], 'timeout': '30'}, 'haproxy': {'skyline_console': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no'}, 'skyline_console_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no'}}}})\u001b[0m",
                                "",
                                "TASK [haproxy-config : Add configuration for skyline when using single external frontend] ***",
                                "Tuesday 24 February 2026  00:55:12 +0000 (0:00:07.155)       0:05:37.277 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'skyline-apiserver', 'value': {'container_name': 'skyline_apiserver', 'group': 'skyline-apiserver', 'enabled': True, 'image': 'registry.osism.tech/kolla/skyline-apiserver:2024.2', 'volumes': ['/etc/kolla/skyline-apiserver/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9998/docs'], 'timeout': '30'}, 'haproxy': {'skyline_apiserver': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no'}, 'skyline_apiserver_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'skyline-console', 'value': {'container_name': 'skyline_console', 'group': 'skyline-console', 'enabled': True, 'image': 'registry.osism.tech/kolla/skyline-console:2024.2', 'volumes': ['/etc/kolla/skyline-console/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9999/docs'], 'timeout': '30'}, 'haproxy': {'skyline_console': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no'}, 'skyline_console_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'skyline-apiserver', 'value': {'container_name': 'skyline_apiserver', 'group': 'skyline-apiserver', 'enabled': True, 'image': 'registry.osism.tech/kolla/skyline-apiserver:2024.2', 'volumes': ['/etc/kolla/skyline-apiserver/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9998/docs'], 'timeout': '30'}, 'haproxy': {'skyline_apiserver': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no'}, 'skyline_apiserver_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'skyline-console', 'value': {'container_name': 'skyline_console', 'group': 'skyline-console', 'enabled': True, 'image': 'registry.osism.tech/kolla/skyline-console:2024.2', 'volumes': ['/etc/kolla/skyline-console/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9999/docs'], 'timeout': '30'}, 'haproxy': {'skyline_console': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no'}, 'skyline_console_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'skyline-apiserver', 'value': {'container_name': 'skyline_apiserver', 'group': 'skyline-apiserver', 'enabled': True, 'image': 'registry.osism.tech/kolla/skyline-apiserver:2024.2', 'volumes': ['/etc/kolla/skyline-apiserver/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9998/docs'], 'timeout': '30'}, 'haproxy': {'skyline_apiserver': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no'}, 'skyline_apiserver_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'skyline-console', 'value': {'container_name': 'skyline_console', 'group': 'skyline-console', 'enabled': True, 'image': 'registry.osism.tech/kolla/skyline-console:2024.2', 'volumes': ['/etc/kolla/skyline-console/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9999/docs'], 'timeout': '30'}, 'haproxy': {'skyline_console': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no'}, 'skyline_console_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [haproxy-config : Configuring firewall for skyline] ***********************",
                                "Tuesday 24 February 2026  00:55:13 +0000 (0:00:00.650)       0:05:37.927 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'skyline_apiserver', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'skyline_apiserver_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'skyline_console', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'skyline_console_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'skyline_apiserver', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'skyline_apiserver_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'skyline_console', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'skyline_console_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'skyline_apiserver', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'skyline_apiserver_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9998', 'listen_port': '9998', 'tls_backend': 'no'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'skyline_console', 'value': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'skyline_console_external', 'value': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9999', 'listen_port': '9999', 'tls_backend': 'no'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [proxysql-config : Copying over skyline ProxySQL users config] ************",
                                "Tuesday 24 February 2026  00:55:14 +0000 (0:00:01.723)       0:05:39.650 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [proxysql-config : Copying over skyline ProxySQL rules config] ************",
                                "Tuesday 24 February 2026  00:55:16 +0000 (0:00:01.374)       0:05:41.025 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [include_role : swift] ****************************************************",
                                "Tuesday 24 February 2026  00:55:18 +0000 (0:00:02.216)       0:05:43.242 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [include_role : tacker] ***************************************************",
                                "Tuesday 24 February 2026  00:55:18 +0000 (0:00:00.374)       0:05:43.617 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [include_role : trove] ****************************************************",
                                "Tuesday 24 February 2026  00:55:19 +0000 (0:00:00.337)       0:05:43.954 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [include_role : venus] ****************************************************",
                                "Tuesday 24 February 2026  00:55:19 +0000 (0:00:00.684)       0:05:44.639 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [include_role : watcher] **************************************************",
                                "Tuesday 24 February 2026  00:55:20 +0000 (0:00:00.357)       0:05:44.997 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [include_role : zun] ******************************************************",
                                "Tuesday 24 February 2026  00:55:20 +0000 (0:00:00.374)       0:05:45.372 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [loadbalancer : Check IP addresses on the API interface] *******",
                                "Tuesday 24 February 2026  00:55:21 +0000 (0:00:01.159)       0:05:46.532 ****** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [loadbalancer : Group HA nodes by status] **********************",
                                "Tuesday 24 February 2026  00:55:22 +0000 (0:00:00.784)       0:05:47.316 ****** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [loadbalancer : Stop backup keepalived container] **************",
                                "Tuesday 24 February 2026  00:55:22 +0000 (0:00:00.401)       0:05:47.718 ****** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [loadbalancer : Stop backup haproxy container] *****************",
                                "Tuesday 24 February 2026  00:55:23 +0000 (0:00:00.974)       0:05:48.693 ****** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [loadbalancer : Stop backup proxysql container] ****************",
                                "Tuesday 24 February 2026  00:55:25 +0000 (0:00:01.285)       0:05:49.978 ****** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [loadbalancer : Start backup haproxy container] ****************",
                                "Tuesday 24 February 2026  00:55:25 +0000 (0:00:00.876)       0:05:50.855 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [loadbalancer : Wait for backup haproxy to start] **************",
                                "Tuesday 24 February 2026  00:55:35 +0000 (0:00:09.869)       0:06:00.725 ****** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [loadbalancer : Start backup proxysql container] ***************",
                                "Tuesday 24 February 2026  00:55:36 +0000 (0:00:00.861)       0:06:01.587 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [loadbalancer : Wait for backup proxysql to start] *************",
                                "Tuesday 24 February 2026  00:55:52 +0000 (0:00:15.529)       0:06:17.117 ****** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [loadbalancer : Start backup keepalived container] *************",
                                "Tuesday 24 February 2026  00:55:53 +0000 (0:00:00.898)       0:06:18.015 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [loadbalancer : Stop master haproxy container] *****************",
                                "Tuesday 24 February 2026  00:55:58 +0000 (0:00:05.413)       0:06:23.429 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [loadbalancer : Stop master proxysql container] ****************",
                                "Tuesday 24 February 2026  00:55:58 +0000 (0:00:00.452)       0:06:23.882 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [loadbalancer : Stop master keepalived container] **************",
                                "Tuesday 24 February 2026  00:55:59 +0000 (0:00:00.783)       0:06:24.665 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [loadbalancer : Start master haproxy container] ****************",
                                "Tuesday 24 February 2026  00:56:00 +0000 (0:00:00.448)       0:06:25.114 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [loadbalancer : Start master proxysql container] ***************",
                                "Tuesday 24 February 2026  00:56:00 +0000 (0:00:00.436)       0:06:25.550 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [loadbalancer : Start master keepalived container] *************",
                                "Tuesday 24 February 2026  00:56:01 +0000 (0:00:00.385)       0:06:25.935 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [loadbalancer : Wait for haproxy to listen on VIP] *************",
                                "Tuesday 24 February 2026  00:56:01 +0000 (0:00:00.391)       0:06:26.327 ****** ",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "RUNNING HANDLER [loadbalancer : Wait for proxysql to listen on VIP] ************",
                                "Tuesday 24 February 2026  00:56:06 +0000 (0:00:05.365)       0:06:31.693 ****** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=123 \u001b[0m \u001b[0;33mchanged=76  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=97  \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=122 \u001b[0m \u001b[0;33mchanged=76  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=97  \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=122 \u001b[0m \u001b[0;33mchanged=76  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=97  \u001b[0m rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Tuesday 24 February 2026  00:56:07 +0000 (0:00:00.904)       0:06:32.597 ****** ",
                                "=============================================================================== ",
                                "loadbalancer : Start backup proxysql container ------------------------- 15.53s",
                                "loadbalancer : Start backup haproxy container --------------------------- 9.87s",
                                "haproxy-config : Copying over skyline haproxy config -------------------- 7.16s",
                                "haproxy-config : Copying over opensearch haproxy config ----------------- 6.12s",
                                "loadbalancer : Start backup keepalived container ------------------------ 5.41s",
                                "loadbalancer : Wait for haproxy to listen on VIP ------------------------ 5.37s",
                                "haproxy-config : Copying over prometheus haproxy config ----------------- 5.24s",
                                "loadbalancer : Copying over proxysql config ----------------------------- 4.95s",
                                "haproxy-config : Copying over keystone haproxy config ------------------- 4.90s",
                                "haproxy-config : Copying over nova-cell:nova-novncproxy haproxy config --- 4.79s",
                                "haproxy-config : Copying over aodh haproxy config ----------------------- 4.64s",
                                "haproxy-config : Copying over designate haproxy config ------------------ 4.59s",
                                "haproxy-config : Copying over glance haproxy config --------------------- 4.54s",
                                "haproxy-config : Copying over nova haproxy config ----------------------- 4.50s",
                                "haproxy-config : Copying over manila haproxy config --------------------- 4.34s",
                                "haproxy-config : Copying over grafana haproxy config -------------------- 4.21s",
                                "haproxy-config : Copying over neutron haproxy config -------------------- 4.14s",
                                "haproxy-config : Copying over horizon haproxy config -------------------- 4.06s",
                                "haproxy-config : Copying over octavia haproxy config -------------------- 3.85s",
                                "haproxy-config : Copying over barbican haproxy config ------------------- 3.79s",
                                "\u001b[32m2026-02-24 00:56:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 00:56:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ca6226d5-0f70-4445-9f9f-dca4faf1e39d is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:56:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:56:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 21cb4ba5-3e33-4fe4-8954-1e96067dfc1b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:56:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 00:56:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ca6226d5-0f70-4445-9f9f-dca4faf1e39d is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:56:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:56:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 21cb4ba5-3e33-4fe4-8954-1e96067dfc1b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:56:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 00:56:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ca6226d5-0f70-4445-9f9f-dca4faf1e39d is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:56:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:56:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 21cb4ba5-3e33-4fe4-8954-1e96067dfc1b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:56:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 00:56:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ca6226d5-0f70-4445-9f9f-dca4faf1e39d is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:56:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:56:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 21cb4ba5-3e33-4fe4-8954-1e96067dfc1b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:56:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 00:56:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ca6226d5-0f70-4445-9f9f-dca4faf1e39d is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:56:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:56:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 21cb4ba5-3e33-4fe4-8954-1e96067dfc1b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:56:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 00:56:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ca6226d5-0f70-4445-9f9f-dca4faf1e39d is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:56:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:56:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 21cb4ba5-3e33-4fe4-8954-1e96067dfc1b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:56:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 00:56:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ca6226d5-0f70-4445-9f9f-dca4faf1e39d is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:56:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:56:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 21cb4ba5-3e33-4fe4-8954-1e96067dfc1b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:56:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 00:56:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ca6226d5-0f70-4445-9f9f-dca4faf1e39d is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:56:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:56:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 21cb4ba5-3e33-4fe4-8954-1e96067dfc1b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:56:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 00:56:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ca6226d5-0f70-4445-9f9f-dca4faf1e39d is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:56:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:56:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 21cb4ba5-3e33-4fe4-8954-1e96067dfc1b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:56:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 00:56:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ca6226d5-0f70-4445-9f9f-dca4faf1e39d is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:56:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:56:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 21cb4ba5-3e33-4fe4-8954-1e96067dfc1b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:56:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 00:56:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ca6226d5-0f70-4445-9f9f-dca4faf1e39d is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:56:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:56:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 21cb4ba5-3e33-4fe4-8954-1e96067dfc1b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:56:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 00:56:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ca6226d5-0f70-4445-9f9f-dca4faf1e39d is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:56:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:56:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 21cb4ba5-3e33-4fe4-8954-1e96067dfc1b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:56:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 00:56:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ca6226d5-0f70-4445-9f9f-dca4faf1e39d is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:56:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:56:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 21cb4ba5-3e33-4fe4-8954-1e96067dfc1b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:56:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 00:56:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ca6226d5-0f70-4445-9f9f-dca4faf1e39d is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:56:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:56:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 21cb4ba5-3e33-4fe4-8954-1e96067dfc1b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:56:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 00:56:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ca6226d5-0f70-4445-9f9f-dca4faf1e39d is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:56:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:56:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 21cb4ba5-3e33-4fe4-8954-1e96067dfc1b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:56:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 00:56:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ca6226d5-0f70-4445-9f9f-dca4faf1e39d is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:56:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:56:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 21cb4ba5-3e33-4fe4-8954-1e96067dfc1b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:56:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 00:57:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ca6226d5-0f70-4445-9f9f-dca4faf1e39d is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:57:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:57:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 21cb4ba5-3e33-4fe4-8954-1e96067dfc1b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:57:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 00:57:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ca6226d5-0f70-4445-9f9f-dca4faf1e39d is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:57:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:57:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 21cb4ba5-3e33-4fe4-8954-1e96067dfc1b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:57:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 00:57:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ca6226d5-0f70-4445-9f9f-dca4faf1e39d is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:57:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:57:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 21cb4ba5-3e33-4fe4-8954-1e96067dfc1b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:57:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 00:57:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ca6226d5-0f70-4445-9f9f-dca4faf1e39d is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:57:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:57:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 21cb4ba5-3e33-4fe4-8954-1e96067dfc1b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:57:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 00:57:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ca6226d5-0f70-4445-9f9f-dca4faf1e39d is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:57:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:57:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 21cb4ba5-3e33-4fe4-8954-1e96067dfc1b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:57:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 00:57:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ca6226d5-0f70-4445-9f9f-dca4faf1e39d is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:57:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:57:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 21cb4ba5-3e33-4fe4-8954-1e96067dfc1b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:57:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 00:57:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ca6226d5-0f70-4445-9f9f-dca4faf1e39d is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:57:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:57:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 21cb4ba5-3e33-4fe4-8954-1e96067dfc1b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:57:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 00:57:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ca6226d5-0f70-4445-9f9f-dca4faf1e39d is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:57:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:57:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 21cb4ba5-3e33-4fe4-8954-1e96067dfc1b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:57:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 00:57:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ca6226d5-0f70-4445-9f9f-dca4faf1e39d is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:57:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:57:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 21cb4ba5-3e33-4fe4-8954-1e96067dfc1b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:57:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 00:57:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ca6226d5-0f70-4445-9f9f-dca4faf1e39d is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:57:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:57:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 21cb4ba5-3e33-4fe4-8954-1e96067dfc1b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:57:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 00:57:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ca6226d5-0f70-4445-9f9f-dca4faf1e39d is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:57:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:57:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 21cb4ba5-3e33-4fe4-8954-1e96067dfc1b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:57:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 00:57:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ca6226d5-0f70-4445-9f9f-dca4faf1e39d is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:57:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:57:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 21cb4ba5-3e33-4fe4-8954-1e96067dfc1b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:57:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 00:57:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ca6226d5-0f70-4445-9f9f-dca4faf1e39d is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:57:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:57:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 21cb4ba5-3e33-4fe4-8954-1e96067dfc1b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:57:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 00:57:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ca6226d5-0f70-4445-9f9f-dca4faf1e39d is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:57:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:57:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 21cb4ba5-3e33-4fe4-8954-1e96067dfc1b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:57:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 00:57:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ca6226d5-0f70-4445-9f9f-dca4faf1e39d is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:57:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:57:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 21cb4ba5-3e33-4fe4-8954-1e96067dfc1b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:57:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 00:57:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ca6226d5-0f70-4445-9f9f-dca4faf1e39d is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:57:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:57:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 21cb4ba5-3e33-4fe4-8954-1e96067dfc1b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:57:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 00:57:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ca6226d5-0f70-4445-9f9f-dca4faf1e39d is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:57:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:57:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 21cb4ba5-3e33-4fe4-8954-1e96067dfc1b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:57:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 00:57:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ca6226d5-0f70-4445-9f9f-dca4faf1e39d is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:57:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:57:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 21cb4ba5-3e33-4fe4-8954-1e96067dfc1b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:57:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 00:57:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ca6226d5-0f70-4445-9f9f-dca4faf1e39d is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:57:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:57:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 21cb4ba5-3e33-4fe4-8954-1e96067dfc1b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:57:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 00:57:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ca6226d5-0f70-4445-9f9f-dca4faf1e39d is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:57:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:57:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 21cb4ba5-3e33-4fe4-8954-1e96067dfc1b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:57:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 00:58:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ca6226d5-0f70-4445-9f9f-dca4faf1e39d is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:58:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:58:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 21cb4ba5-3e33-4fe4-8954-1e96067dfc1b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:58:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 00:58:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ca6226d5-0f70-4445-9f9f-dca4faf1e39d is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:58:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:58:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 21cb4ba5-3e33-4fe4-8954-1e96067dfc1b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:58:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 00:58:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ca6226d5-0f70-4445-9f9f-dca4faf1e39d is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:58:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:58:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 21cb4ba5-3e33-4fe4-8954-1e96067dfc1b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:58:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 00:58:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ca6226d5-0f70-4445-9f9f-dca4faf1e39d is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:58:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:58:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 21cb4ba5-3e33-4fe4-8954-1e96067dfc1b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:58:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 00:58:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ca6226d5-0f70-4445-9f9f-dca4faf1e39d is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:58:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:58:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 21cb4ba5-3e33-4fe4-8954-1e96067dfc1b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:58:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 00:58:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ca6226d5-0f70-4445-9f9f-dca4faf1e39d is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:58:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:58:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 21cb4ba5-3e33-4fe4-8954-1e96067dfc1b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:58:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 00:58:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ca6226d5-0f70-4445-9f9f-dca4faf1e39d is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:58:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:58:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 21cb4ba5-3e33-4fe4-8954-1e96067dfc1b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:58:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 00:58:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ca6226d5-0f70-4445-9f9f-dca4faf1e39d is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:58:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:58:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 21cb4ba5-3e33-4fe4-8954-1e96067dfc1b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:58:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 00:58:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ca6226d5-0f70-4445-9f9f-dca4faf1e39d is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:58:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:58:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 21cb4ba5-3e33-4fe4-8954-1e96067dfc1b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:58:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 00:58:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ca6226d5-0f70-4445-9f9f-dca4faf1e39d is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:58:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:58:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 21cb4ba5-3e33-4fe4-8954-1e96067dfc1b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:58:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 00:58:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ca6226d5-0f70-4445-9f9f-dca4faf1e39d is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:58:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:58:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 21cb4ba5-3e33-4fe4-8954-1e96067dfc1b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:58:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 00:58:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ca6226d5-0f70-4445-9f9f-dca4faf1e39d is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:58:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:58:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 21cb4ba5-3e33-4fe4-8954-1e96067dfc1b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:58:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 00:58:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ca6226d5-0f70-4445-9f9f-dca4faf1e39d is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:58:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:58:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 21cb4ba5-3e33-4fe4-8954-1e96067dfc1b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:58:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 00:58:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ca6226d5-0f70-4445-9f9f-dca4faf1e39d is in state STARTED\u001b[0m",
                                "",
                                "\u001b[1;35m[WARNING]: Collection community.general does not support Ansible version\u001b[0m",
                                "\u001b[1;35m2.16.14\u001b[0m",
                                "",
                                "PLAY [Prepare deployment of Ceph services] *************************************",
                                "",
                                "TASK [ceph-facts : Include facts.yml] ******************************************",
                                "Tuesday 24 February 2026  00:47:15 +0000 (0:00:00.702)       0:00:00.702 ****** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-facts/tasks/facts.yml for testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [ceph-facts : Check if it is atomic host] *********************************",
                                "Tuesday 24 February 2026  00:47:16 +0000 (0:00:01.114)       0:00:01.817 ****** ",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact is_atomic] *****************************************",
                                "Tuesday 24 February 2026  00:47:18 +0000 (0:00:01.734)       0:00:03.551 ****** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Check if podman binary is present] **************************",
                                "Tuesday 24 February 2026  00:47:19 +0000 (0:00:01.156)       0:00:04.708 ****** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact container_binary] **********************************",
                                "Tuesday 24 February 2026  00:47:20 +0000 (0:00:01.104)       0:00:05.813 ****** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact ceph_cmd] ******************************************",
                                "Tuesday 24 February 2026  00:47:21 +0000 (0:00:00.859)       0:00:06.672 ****** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact discovered_interpreter_python] *********************",
                                "Tuesday 24 February 2026  00:47:21 +0000 (0:00:00.591)       0:00:07.263 ****** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact discovered_interpreter_python if not previously set] ***",
                                "Tuesday 24 February 2026  00:47:22 +0000 (0:00:00.780)       0:00:08.044 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact ceph_release ceph_stable_release] ******************",
                                "Tuesday 24 February 2026  00:47:23 +0000 (0:00:00.963)       0:00:09.008 ****** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact monitor_name ansible_facts['hostname']] ************",
                                "Tuesday 24 February 2026  00:47:25 +0000 (0:00:01.446)       0:00:10.454 ****** ",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item=testbed-node-0)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-1(192.168.16.11)] => (item=testbed-node-1)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-2(192.168.16.12)] => (item=testbed-node-2)\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact container_exec_cmd] ********************************",
                                "Tuesday 24 February 2026  00:47:25 +0000 (0:00:00.596)       0:00:11.050 ****** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Find a running mon container] *******************************",
                                "Tuesday 24 February 2026  00:47:26 +0000 (0:00:01.213)       0:00:12.264 ****** ",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item=testbed-node-0)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-1(192.168.16.11)] => (item=testbed-node-1)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-2(192.168.16.12)] => (item=testbed-node-2)\u001b[0m",
                                "",
                                "TASK [ceph-facts : Check for a ceph mon socket] ********************************",
                                "Tuesday 24 February 2026  00:47:29 +0000 (0:00:02.645)       0:00:14.913 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-0) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-1) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Check if the ceph mon socket is in-use] *********************",
                                "Tuesday 24 February 2026  00:47:30 +0000 (0:00:00.722)       0:00:15.635 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-0', 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-1', 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-2', 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact running_mon - non_container] ***********************",
                                "Tuesday 24 February 2026  00:47:31 +0000 (0:00:01.088)       0:00:16.725 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': {'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-0', 'ansible_loop_var': 'item'}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': {'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-1', 'ansible_loop_var': 'item'}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': {'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-2', 'ansible_loop_var': 'item'}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact running_mon - container] ***************************",
                                "Tuesday 24 February 2026  00:47:31 +0000 (0:00:00.464)       0:00:17.190 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'stdout': '', 'stderr': '', 'rc': 0, 'cmd': ['docker', 'ps', '-q', '--filter', 'name=ceph-mon-testbed-node-0'], 'start': '2026-02-24 00:47:27.526222', 'end': '2026-02-24 00:47:27.625319', 'delta': '0:00:00.099097', 'msg': '', 'invocation': {'module_args': {'_raw_params': 'docker ps -q --filter name=ceph-mon-testbed-node-0', '_uses_shell': False, 'expand_argument_vars': True, 'stdin_add_newline': True, 'strip_empty_ends': True, 'argv': None, 'chdir': None, 'executable': None, 'creates': None, 'removes': None, 'stdin': None}}, 'stdout_lines': [], 'stderr_lines': [], 'failed': False, 'failed_when_result': False, 'item': 'testbed-node-0', 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'stdout': '', 'stderr': '', 'rc': 0, 'cmd': ['docker', 'ps', '-q', '--filter', 'name=ceph-mon-testbed-node-1'], 'start': '2026-02-24 00:47:28.794784', 'end': '2026-02-24 00:47:28.901216', 'delta': '0:00:00.106432', 'msg': '', 'invocation': {'module_args': {'_raw_params': 'docker ps -q --filter name=ceph-mon-testbed-node-1', '_uses_shell': False, 'expand_argument_vars': True, 'stdin_add_newline': True, 'strip_empty_ends': True, 'argv': None, 'chdir': None, 'executable': None, 'creates': None, 'removes': None, 'stdin': None}}, 'stdout_lines': [], 'stderr_lines': [], 'failed': False, 'failed_when_result': False, 'item': 'testbed-node-1', 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'stdout': '', 'stderr': '', 'rc': 0, 'cmd': ['docker', 'ps', '-q', '--filter', 'name=ceph-mon-testbed-node-2'], 'start': '2026-02-24 00:47:29.340894', 'end': '2026-02-24 00:47:29.439510', 'delta': '0:00:00.098616', 'msg': '', 'invocation': {'module_args': {'_raw_params': 'docker ps -q --filter name=ceph-mon-testbed-node-2', '_uses_shell': False, 'expand_argument_vars': True, 'stdin_add_newline': True, 'strip_empty_ends': True, 'argv': None, 'chdir': None, 'executable': None, 'creates': None, 'removes': None, 'stdin': None}}, 'stdout_lines': [], 'stderr_lines': [], 'failed': False, 'failed_when_result': False, 'item': 'testbed-node-2', 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _container_exec_cmd] *******************************",
                                "Tuesday 24 February 2026  00:47:32 +0000 (0:00:00.210)       0:00:17.400 ****** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Get current fsid if cluster is already running] *************",
                                "Tuesday 24 February 2026  00:47:33 +0000 (0:00:01.197)       0:00:18.597 ****** ",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-0(192.168.16.10)]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact current_fsid rc 1] *********************************",
                                "Tuesday 24 February 2026  00:47:33 +0000 (0:00:00.659)       0:00:19.257 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Get current fsid] *******************************************",
                                "Tuesday 24 February 2026  00:47:36 +0000 (0:00:02.275)       0:00:21.532 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact fsid] **********************************************",
                                "Tuesday 24 February 2026  00:47:37 +0000 (0:00:01.272)       0:00:22.805 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact fsid from current_fsid] ****************************",
                                "Tuesday 24 February 2026  00:47:38 +0000 (0:00:01.370)       0:00:24.176 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Generate cluster fsid] **************************************",
                                "Tuesday 24 February 2026  00:47:39 +0000 (0:00:00.245)       0:00:24.421 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact fsid] **********************************************",
                                "Tuesday 24 February 2026  00:47:39 +0000 (0:00:00.657)       0:00:25.078 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Resolve device link(s)] *************************************",
                                "Tuesday 24 February 2026  00:47:40 +0000 (0:00:01.121)       0:00:26.200 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact build devices from resolved symlinks] **************",
                                "Tuesday 24 February 2026  00:47:42 +0000 (0:00:01.206)       0:00:27.407 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Resolve dedicated_device link(s)] ***************************",
                                "Tuesday 24 February 2026  00:47:42 +0000 (0:00:00.746)       0:00:28.153 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact build dedicated_devices from resolved symlinks] ****",
                                "Tuesday 24 February 2026  00:47:43 +0000 (0:00:01.100)       0:00:29.254 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Resolve bluestore_wal_device link(s)] ***********************",
                                "Tuesday 24 February 2026  00:47:44 +0000 (0:00:00.832)       0:00:30.086 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact build bluestore_wal_devices from resolved symlinks] ***",
                                "Tuesday 24 February 2026  00:47:45 +0000 (0:00:00.770)       0:00:30.857 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Collect existed devices] ************************************",
                                "Tuesday 24 February 2026  00:47:46 +0000 (0:00:00.615)       0:00:31.472 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'dm-0', 'value': {'holders': [], 'host': '', 'links': {'ids': ['dm-name-ceph--9b92f24c--dac1--5bff--a390--34d9ff9e0f4f-osd--block--9b92f24c--dac1--5bff--a390--34d9ff9e0f4f', 'dm-uuid-LVM-h1BZIKSwq0ZvfBWtXhB9IpbOvYH7db3oH2r78k3xPmfToT2WBghjfHVanEend8Zd'], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': '', 'sectors': 41934848, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'dm-1', 'value': {'holders': [], 'host': '', 'links': {'ids': ['dm-name-ceph--7f6a8fe5--6adc--52b4--b256--edcdbfd9bf4e-osd--block--7f6a8fe5--6adc--52b4--b256--edcdbfd9bf4e', 'dm-uuid-LVM-xRh1zTNtAhB8OweVwVFtzaksaCJoA7fP27eNW4gBNJ6VNIfew4kSrXXLjM4ybDHa'], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': '', 'sectors': 41934848, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'loop0', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'loop1', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'loop2', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'loop3', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'loop4', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'loop5', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'dm-0', 'value': {'holders': [], 'host': '', 'links': {'ids': ['dm-name-ceph--294dd48f--9d6d--5079--8214--59e60af9e526-osd--block--294dd48f--9d6d--5079--8214--59e60af9e526', 'dm-uuid-LVM-wjENyDIL1ovSEUDphRiOfALWwweDrSZWE8niuN9gyF3lbbybFqRYsEaNkOnZAeUN'], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': '', 'sectors': 41934848, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'dm-1', 'value': {'holders': [], 'host': '', 'links': {'ids': ['dm-name-ceph--7fbf4d63--7c6f--5c70--887c--2b73c978775b-osd--block--7fbf4d63--7c6f--5c70--887c--2b73c978775b', 'dm-uuid-LVM-aRcKlwv7ZYsrJYaz5OO903mOAEBnwLXI61utfRQTUZQKHwm4ceJ3NApz8UrwnN8y'], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': '', 'sectors': 41934848, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'loop6', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'loop0', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'loop7', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'loop1', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'loop2', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'sda', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_f90a4e80-1f36-4d14-b7da-de7b20e9fbca', 'scsi-SQEMU_QEMU_HARDDISK_f90a4e80-1f36-4d14-b7da-de7b20e9fbca'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {'sda1': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_f90a4e80-1f36-4d14-b7da-de7b20e9fbca-part1', 'scsi-SQEMU_QEMU_HARDDISK_f90a4e80-1f36-4d14-b7da-de7b20e9fbca-part1'], 'labels': ['cloudimg-rootfs'], 'masters': [], 'uuids': ['b852d8d2-8460-44aa-8998-23e4f04d73cf']}, 'sectors': 165672927, 'sectorsize': 512, 'size': '79.00 GB', 'start': '2099200', 'uuid': 'b852d8d2-8460-44aa-8998-23e4f04d73cf'}, 'sda14': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_f90a4e80-1f36-4d14-b7da-de7b20e9fbca-part14', 'scsi-SQEMU_QEMU_HARDDISK_f90a4e80-1f36-4d14-b7da-de7b20e9fbca-part14'], 'labels': [], 'masters': [], 'uuids': []}, 'sectors': 8192, 'sectorsize': 512, 'size': '4.00 MB', 'start': '2048', 'uuid': None}, 'sda15': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_f90a4e80-1f36-4d14-b7da-de7b20e9fbca-part15', 'scsi-SQEMU_QEMU_HARDDISK_f90a4e80-1f36-4d14-b7da-de7b20e9fbca-part15'], 'labels': ['UEFI'], 'masters': [], 'uuids': ['5C78-612A']}, 'sectors': 217088, 'sectorsize': 512, 'size': '106.00 MB', 'start': '10240', 'uuid': '5C78-612A'}, 'sda16': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_f90a4e80-1f36-4d14-b7da-de7b20e9fbca-part16', 'scsi-SQEMU_QEMU_HARDDISK_f90a4e80-1f36-4d14-b7da-de7b20e9fbca-part16'], 'labels': ['BOOT'], 'masters': [], 'uuids': ['09d53dc1-1e03-4286-bbb8-2b1796cf92ec']}, 'sectors': 1869825, 'sectorsize': 512, 'size': '913.00 MB', 'start': '227328', 'uuid': '09d53dc1-1e03-4286-bbb8-2b1796cf92ec'}}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 167772160, 'sectorsize': '512', 'size': '80.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'loop3', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'loop4', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'sdb', 'value': {'holders': ['ceph--9b92f24c--dac1--5bff--a390--34d9ff9e0f4f-osd--block--9b92f24c--dac1--5bff--a390--34d9ff9e0f4f'], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['lvm-pv-uuid-mec8nJ-xUip-qiWP-sj8w-uf5B-1Lt0-v8yfvH', 'scsi-0QEMU_QEMU_HARDDISK_9c6b6763-9704-41ec-9b6d-97eb80a5145c', 'scsi-SQEMU_QEMU_HARDDISK_9c6b6763-9704-41ec-9b6d-97eb80a5145c'], 'labels': [], 'masters': ['dm-0'], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'loop5', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'dm-0', 'value': {'holders': [], 'host': '', 'links': {'ids': ['dm-name-ceph--752a9ec4--4bfc--5551--82cd--92b802bdad97-osd--block--752a9ec4--4bfc--5551--82cd--92b802bdad97', 'dm-uuid-LVM-jTt9fg6XlhiXbWhW0I9T02tzYsIeoKIoaf5BxtnvIAUk6EWNKNJMqSjVsZqTXHAa'], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': '', 'sectors': 41934848, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'dm-1', 'value': {'holders': [], 'host': '', 'links': {'ids': ['dm-name-ceph--ba85a0e3--2f57--51d5--a9f1--072f0e9a8deb-osd--block--ba85a0e3--2f57--51d5--a9f1--072f0e9a8deb', 'dm-uuid-LVM-YjmBzAWpiQhdCbdayUM0w3sqNqanvbLDieybFDy5wWcdDcB5LPRFHjaRE0WCOjBW'], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': '', 'sectors': 41934848, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'loop6', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'sdc', 'value': {'holders': ['ceph--7f6a8fe5--6adc--52b4--b256--edcdbfd9bf4e-osd--block--7f6a8fe5--6adc--52b4--b256--edcdbfd9bf4e'], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['lvm-pv-uuid-HdfFdZ-81SJ-0Umt-qAzb-Dao2-7tUe-tHuhnG', 'scsi-0QEMU_QEMU_HARDDISK_6644f191-69e9-4f3a-baff-cffe4152595b', 'scsi-SQEMU_QEMU_HARDDISK_6644f191-69e9-4f3a-baff-cffe4152595b'], 'labels': [], 'masters': ['dm-1'], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'loop0', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'loop1', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'loop2', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'loop3', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'loop4', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'loop5', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'sdd', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_4534650b-23a0-46c8-8bf1-44520caee637', 'scsi-SQEMU_QEMU_HARDDISK_4534650b-23a0-46c8-8bf1-44520caee637'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'loop6', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'loop7', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'loop7', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partit\u001b[32m2026-02-24 00:58:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 5e6d114b-1478-4583-97b2-79f08cfd01e8 is in state SUCCESS\u001b[0m",
                                "ions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'sda', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_7a9487cb-ec9e-4508-9c5a-f692700fc5a4', 'scsi-SQEMU_QEMU_HARDDISK_7a9487cb-ec9e-4508-9c5a-f692700fc5a4'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {'sda1': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_7a9487cb-ec9e-4508-9c5a-f692700fc5a4-part1', 'scsi-SQEMU_QEMU_HARDDISK_7a9487cb-ec9e-4508-9c5a-f692700fc5a4-part1'], 'labels': ['cloudimg-rootfs'], 'masters': [], 'uuids': ['b852d8d2-8460-44aa-8998-23e4f04d73cf']}, 'sectors': 165672927, 'sectorsize': 512, 'size': '79.00 GB', 'start': '2099200', 'uuid': 'b852d8d2-8460-44aa-8998-23e4f04d73cf'}, 'sda14': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_7a9487cb-ec9e-4508-9c5a-f692700fc5a4-part14', 'scsi-SQEMU_QEMU_HARDDISK_7a9487cb-ec9e-4508-9c5a-f692700fc5a4-part14'], 'labels': [], 'masters': [], 'uuids': []}, 'sectors': 8192, 'sectorsize': 512, 'size': '4.00 MB', 'start': '2048', 'uuid': None}, 'sda15': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_7a9487cb-ec9e-4508-9c5a-f692700fc5a4-part15', 'scsi-SQEMU_QEMU_HARDDISK_7a9487cb-ec9e-4508-9c5a-f692700fc5a4-part15'], 'labels': ['UEFI'], 'masters': [], 'uuids': ['5C78-612A']}, 'sectors': 217088, 'sectorsize': 512, 'size': '106.00 MB', 'start': '10240', 'uuid': '5C78-612A'}, 'sda16': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_7a9487cb-ec9e-4508-9c5a-f692700fc5a4-part16', 'scsi-SQEMU_QEMU_HARDDISK_7a9487cb-ec9e-4508-9c5a-f692700fc5a4-part16'], 'labels': ['BOOT'], 'masters': [], 'uuids': ['09d53dc1-1e03-4286-bbb8-2b1796cf92ec']}, 'sectors': 1869825, 'sectorsize': 512, 'size': '913.00 MB', 'start': '227328', 'uuid': '09d53dc1-1e03-4286-bbb8-2b1796cf92ec'}}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 167772160, 'sectorsize': '512', 'size': '80.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'sr0', 'value': {'holders': [], 'host': 'IDE interface: Intel Corporation 82371SB PIIX3 IDE [Natoma/Triton II]', 'links': {'ids': ['ata-QEMU_DVD-ROM_QM00001'], 'labels': ['config-2'], 'masters': [], 'uuids': ['2026-02-24-00-01-58-00']}, 'model': 'QEMU DVD-ROM', 'partitions': {}, 'removable': '1', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'mq-deadline', 'sectors': 253, 'sectorsize': '2048', 'size': '506.00 KB', 'support_discard': '0', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'sda', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_47e600ec-05ca-44a6-93d0-5ca3104e3639', 'scsi-SQEMU_QEMU_HARDDISK_47e600ec-05ca-44a6-93d0-5ca3104e3639'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {'sda1': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_47e600ec-05ca-44a6-93d0-5ca3104e3639-part1', 'scsi-SQEMU_QEMU_HARDDISK_47e600ec-05ca-44a6-93d0-5ca3104e3639-part1'], 'labels': ['cloudimg-rootfs'], 'masters': [], 'uuids': ['b852d8d2-8460-44aa-8998-23e4f04d73cf']}, 'sectors': 165672927, 'sectorsize': 512, 'size': '79.00 GB', 'start': '2099200', 'uuid': 'b852d8d2-8460-44aa-8998-23e4f04d73cf'}, 'sda14': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_47e600ec-05ca-44a6-93d0-5ca3104e3639-part14', 'scsi-SQEMU_QEMU_HARDDISK_47e600ec-05ca-44a6-93d0-5ca3104e3639-part14'], 'labels': [], 'masters': [], 'uuids': []}, 'sectors': 8192, 'sectorsize': 512, 'size': '4.00 MB', 'start': '2048', 'uuid': None}, 'sda15': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_47e600ec-05ca-44a6-93d0-5ca3104e3639-part15', 'scsi-SQEMU_QEMU_HARDDISK_47e600ec-05ca-44a6-93d0-5ca3104e3639-part15'], 'labels': ['UEFI'], 'masters': [], 'uuids': ['5C78-612A']}, 'sectors': 217088, 'sectorsize': 512, 'size': '106.00 MB', 'start': '10240', 'uuid': '5C78-612A'}, 'sda16': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_47e600ec-05ca-44a6-93d0-5ca3104e3639-part16', 'scsi-SQEMU_QEMU_HARDDISK_47e600ec-05ca-44a6-93d0-5ca3104e3639-part16'], 'labels': ['BOOT'], 'masters': [], 'uuids': ['09d53dc1-1e03-4286-bbb8-2b1796cf92ec']}, 'sectors': 1869825, 'sectorsize': 512, 'size': '913.00 MB', 'start': '227328', 'uuid': '09d53dc1-1e03-4286-bbb8-2b1796cf92ec'}}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 167772160, 'sectorsize': '512', 'size': '80.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'sdb', 'value': {'holders': ['ceph--752a9ec4--4bfc--5551--82cd--92b802bdad97-osd--block--752a9ec4--4bfc--5551--82cd--92b802bdad97'], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['lvm-pv-uuid-vfwhPS-pgoq-FNQw-W72Z-epjC-Ziiv-qWNph9', 'scsi-0QEMU_QEMU_HARDDISK_1558935b-4c66-4614-97e5-cda14abd2641', 'scsi-SQEMU_QEMU_HARDDISK_1558935b-4c66-4614-97e5-cda14abd2641'], 'labels': [], 'masters': ['dm-0'], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'sdb', 'value': {'holders': ['ceph--294dd48f--9d6d--5079--8214--59e60af9e526-osd--block--294dd48f--9d6d--5079--8214--59e60af9e526'], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['lvm-pv-uuid-1V6jIs-3p8q-l9Rl-OJ2n-8L4P-5xCz-K0YqYS', 'scsi-0QEMU_QEMU_HARDDISK_f659b61f-c39f-426f-a640-6a753951661d', 'scsi-SQEMU_QEMU_HARDDISK_f659b61f-c39f-426f-a640-6a753951661d'], 'labels': [], 'masters': ['dm-0'], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'sdc', 'value': {'holders': ['ceph--ba85a0e3--2f57--51d5--a9f1--072f0e9a8deb-osd--block--ba85a0e3--2f57--51d5--a9f1--072f0e9a8deb'], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['lvm-pv-uuid-XuBysF-MTWF-hP8t-3CGS-0PEc-QL0q-u9CYq2', 'scsi-0QEMU_QEMU_HARDDISK_6f4643f0-329b-4ae2-97c7-d348648fd970', 'scsi-SQEMU_QEMU_HARDDISK_6f4643f0-329b-4ae2-97c7-d348648fd970'], 'labels': [], 'masters': ['dm-1'], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'sdc', 'value': {'holders': ['ceph--7fbf4d63--7c6f--5c70--887c--2b73c978775b-osd--block--7fbf4d63--7c6f--5c70--887c--2b73c978775b'], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['lvm-pv-uuid-aHCB2h-b0pT-WCWd-loOK-A09H-yFYp-FH2wIv', 'scsi-0QEMU_QEMU_HARDDISK_fb58bf16-a1c6-4abb-925b-29c6a6a6f100', 'scsi-SQEMU_QEMU_HARDDISK_fb58bf16-a1c6-4abb-925b-29c6a6a6f100'], 'labels': [], 'masters': ['dm-1'], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'sdd', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_5f47789f-2a1a-442a-8a3c-644a4f92e5a8', 'scsi-SQEMU_QEMU_HARDDISK_5f47789f-2a1a-442a-8a3c-644a4f92e5a8'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'sr0', 'value': {'holders': [], 'host': 'IDE interface: Intel Corporation 82371SB PIIX3 IDE [Natoma/Triton II]', 'links': {'ids': ['ata-QEMU_DVD-ROM_QM00001'], 'labels': ['config-2'], 'masters': [], 'uuids': ['2026-02-24-00-02-03-00']}, 'model': 'QEMU DVD-ROM', 'partitions': {}, 'removable': '1', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'mq-deadline', 'sectors': 253, 'sectorsize': '2048', 'size': '506.00 KB', 'support_discard': '0', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'sdd', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_5660bb62-e2c9-45b2-ae11-8ab0f829b194', 'scsi-SQEMU_QEMU_HARDDISK_5660bb62-e2c9-45b2-ae11-8ab0f829b194'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'sr0', 'value': {'holders': [], 'host': 'IDE interface: Intel Corporation 82371SB PIIX3 IDE [Natoma/Triton II]', 'links': {'ids': ['ata-QEMU_DVD-ROM_QM00001'], 'labels': ['config-2'], 'masters': [], 'uuids': ['2026-02-24-00-01-59-00']}, 'model': 'QEMU DVD-ROM', 'partitions': {}, 'removable': '1', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'mq-deadline', 'sectors': 253, 'sectorsize': '2048', 'size': '506.00 KB', 'support_discard': '0', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'loop0', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'loop1', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'loop2', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'loop3', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'loop4', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'loop5', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'loop6', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'loop7', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'sda', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_2d259cae-7b79-479e-b978-3fffee75aa51', 'scsi-SQEMU_QEMU_HARDDISK_2d259cae-7b79-479e-b978-3fffee75aa51'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {'sda1': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_2d259cae-7b79-479e-b978-3fffee75aa51-part1', 'scsi-SQEMU_QEMU_HARDDISK_2d259cae-7b79-479e-b978-3fffee75aa51-part1'], 'labels': ['cloudimg-rootfs'], 'masters': [], 'uuids': ['b852d8d2-8460-44aa-8998-23e4f04d73cf']}, 'sectors': 165672927, 'sectorsize': 512, 'size': '79.00 GB', 'start': '2099200', 'uuid': 'b852d8d2-8460-44aa-8998-23e4f04d73cf'}, 'sda14': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_2d259cae-7b79-479e-b978-3fffee75aa51-part14', 'scsi-SQEMU_QEMU_HARDDISK_2d259cae-7b79-479e-b978-3fffee75aa51-part14'], 'labels': [], 'masters': [], 'uuids': []}, 'sectors': 8192, 'sectorsize': 512, 'size': '4.00 MB', 'start': '2048', 'uuid': None}, 'sda15': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_2d259cae-7b79-479e-b978-3fffee75aa51-part15', 'scsi-SQEMU_QEMU_HARDDISK_2d259cae-7b79-479e-b978-3fffee75aa51-part15'], 'labels': ['UEFI'], 'masters': [], 'uuids': ['5C78-612A']}, 'sectors': 217088, 'sectorsize': 512, 'size': '106.00 MB', 'start': '10240', 'uuid': '5C78-612A'}, 'sda16': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_2d259cae-7b79-479e-b978-3fffee75aa51-part16', 'scsi-SQEMU_QEMU_HARDDISK_2d259cae-7b79-479e-b978-3fffee75aa51-part16'], 'labels': ['BOOT'], 'masters': [], 'uuids': ['09d53dc1-1e03-4286-bbb8-2b1796cf92ec']}, 'sectors': 1869825, 'sectorsize': 512, 'size': '913.00 MB', 'start': '227328', 'uuid': '09d53dc1-1e03-4286-bbb8-2b1796cf92ec'}}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 167772160, 'sectorsize': '512', 'size': '80.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'sr0', 'value': {'holders': [], 'host': 'IDE interface: Intel Corporation 82371SB PIIX3 IDE [Natoma/Triton II]', 'links': {'ids': ['ata-QEMU_DVD-ROM_QM00001'], 'labels': ['config-2'], 'masters': [], 'uuids': ['2026-02-24-00-01-56-00']}, 'model': 'QEMU DVD-ROM', 'partitions': {}, 'removable': '1', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'mq-deadline', 'sectors': 253, 'sectorsize': '2048', 'size': '506.00 KB', 'support_discard': '0', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'loop0', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'loop1', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'loop2', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'loop3', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'loop4', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'loop5', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'loop6', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'loop7', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'sda', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_6fef90f9-4ce5-456f-a170-152290f19313', 'scsi-SQEMU_QEMU_HARDDISK_6fef90f9-4ce5-456f-a170-152290f19313'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {'sda1': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_6fef90f9-4ce5-456f-a170-152290f19313-part1', 'scsi-SQEMU_QEMU_HARDDISK_6fef90f9-4ce5-456f-a170-152290f19313-part1'], 'labels': ['cloudimg-rootfs'], 'masters': [], 'uuids': ['b852d8d2-8460-44aa-8998-23e4f04d73cf']}, 'sectors': 165672927, 'sectorsize': 512, 'size': '79.00 GB', 'start': '2099200', 'uuid': 'b852d8d2-8460-44aa-8998-23e4f04d73cf'}, 'sda14': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_6fef90f9-4ce5-456f-a170-152290f19313-part14', 'scsi-SQEMU_QEMU_HARDDISK_6fef90f9-4ce5-456f-a170-152290f19313-part14'], 'labels': [], 'masters': [], 'uuids': []}, 'sectors': 8192, 'sectorsize': 512, 'size': '4.00 MB', 'start': '2048', 'uuid': None}, 'sda15': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_6fef90f9-4ce5-456f-a170-152290f19313-part15', 'scsi-SQEMU_QEMU_HARDDISK_6fef90f9-4ce5-456f-a170-152290f19313-part15'], 'labels': ['UEFI'], 'masters': [], 'uuids': ['5C78-612A']}, 'sectors': 217088, 'sectorsize': 512, 'size': '106.00 MB', 'start': '10240', 'uuid': '5C78-612A'}, 'sda16': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_6fef90f9-4ce5-456f-a170-152290f19313-part16', 'scsi-SQEMU_QEMU_HARDDISK_6fef90f9-4ce5-456f-a170-152290f19313-part16'], 'labels': ['BOOT'], 'masters': [], 'uuids': ['09d53dc1-1e03-4286-bbb8-2b1796cf92ec']}, 'sectors': 1869825, 'sectorsize': 512, 'size': '913.00 MB', 'start': '227328', 'uuid': '09d53dc1-1e03-4286-bbb8-2b1796cf92ec'}}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 167772160, 'sectorsize': '512', 'size': '80.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'sr0', 'value': {'holders': [], 'host': 'IDE interface: Intel Corporation 82371SB PIIX3 IDE [Natoma/Triton II]', 'links': {'ids': ['ata-QEMU_DVD-ROM_QM00001'], 'labels': ['config-2'], 'masters': [], 'uuids': ['2026-02-24-00-02-01-00']}, 'model': 'QEMU DVD-ROM', 'partitions': {}, 'removable': '1', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'mq-deadline', 'sectors': 253, 'sectorsize': '2048', 'size': '506.00 KB', 'support_discard': '0', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'loop0', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'loop1', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'loop2', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'loop3', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'loop4', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'loop5', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'loop6', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'loop7', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'sda', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_35b8d51e-716e-455d-9a95-995756ae5900', 'scsi-SQEMU_QEMU_HARDDISK_35b8d51e-716e-455d-9a95-995756ae5900'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {'sda1': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_35b8d51e-716e-455d-9a95-995756ae5900-part1', 'scsi-SQEMU_QEMU_HARDDISK_35b8d51e-716e-455d-9a95-995756ae5900-part1'], 'labels': ['cloudimg-rootfs'], 'masters': [], 'uuids': ['b852d8d2-8460-44aa-8998-23e4f04d73cf']}, 'sectors': 165672927, 'sectorsize': 512, 'size': '79.00 GB', 'start': '2099200', 'uuid': 'b852d8d2-8460-44aa-8998-23e4f04d73cf'}, 'sda14': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_35b8d51e-716e-455d-9a95-995756ae5900-part14', 'scsi-SQEMU_QEMU_HARDDISK_35b8d51e-716e-455d-9a95-995756ae5900-part14'], 'labels': [], 'masters': [], 'uuids': []}, 'sectors': 8192, 'sectorsize': 512, 'size': '4.00 MB', 'start': '2048', 'uuid': None}, 'sda15': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_35b8d51e-716e-455d-9a95-995756ae5900-part15', 'scsi-SQEMU_QEMU_HARDDISK_35b8d51e-716e-455d-9a95-995756ae5900-part15'], 'labels': ['UEFI'], 'masters': [], 'uuids': ['5C78-612A']}, 'sectors': 217088, 'sectorsize': 512, 'size': '106.00 MB', 'start': '10240', 'uuid': '5C78-612A'}, 'sda16': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_35b8d51e-716e-455d-9a95-995756ae5900-part16', 'scsi-SQEMU_QEMU_HARDDISK_35b8d51e-716e-455d-9a95-995756ae5900-part16'], 'labels': ['BOOT'], 'masters': [], 'uuids': ['09d53dc1-1e03-4286-bbb8-2b1796cf92ec']}, 'sectors': 1869825, 'sectorsize': 512, 'size': '913.00 MB', 'start': '227328', 'uuid': '09d53dc1-1e03-4286-bbb8-2b1796cf92ec'}}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 167772160, 'sectorsize': '512', 'size': '80.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'sr0', 'value': {'holders': [], 'host': 'IDE interface: Intel Corporation 82371SB PIIX3 IDE [Natoma/Triton II]', 'links': {'ids': ['ata-QEMU_DVD-ROM_QM00001'], 'labels': ['config-2'], 'masters': [], 'uuids': ['2026-02-24-00-02-05-00']}, 'model': 'QEMU DVD-ROM', 'partitions': {}, 'removable': '1', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'mq-deadline', 'sectors': 253, 'sectorsize': '2048', 'size': '506.00 KB', 'support_discard': '0', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact devices generate device list when osd_auto_discovery] ***",
                                "Tuesday 24 February 2026  00:47:48 +0000 (0:00:01.901)       0:00:33.374 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'dm-0', 'value': {'holders': [], 'host': '', 'links': {'ids': ['dm-name-ceph--9b92f24c--dac1--5bff--a390--34d9ff9e0f4f-osd--block--9b92f24c--dac1--5bff--a390--34d9ff9e0f4f', 'dm-uuid-LVM-h1BZIKSwq0ZvfBWtXhB9IpbOvYH7db3oH2r78k3xPmfToT2WBghjfHVanEend8Zd'], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': '', 'sectors': 41934848, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'dm-1', 'value': {'holders': [], 'host': '', 'links': {'ids': ['dm-name-ceph--7f6a8fe5--6adc--52b4--b256--edcdbfd9bf4e-osd--block--7f6a8fe5--6adc--52b4--b256--edcdbfd9bf4e', 'dm-uuid-LVM-xRh1zTNtAhB8OweVwVFtzaksaCJoA7fP27eNW4gBNJ6VNIfew4kSrXXLjM4ybDHa'], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': '', 'sectors': 41934848, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop0', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop1', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop2', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'dm-0', 'value': {'holders': [], 'host': '', 'links': {'ids': ['dm-name-ceph--294dd48f--9d6d--5079--8214--59e60af9e526-osd--block--294dd48f--9d6d--5079--8214--59e60af9e526', 'dm-uuid-LVM-wjENyDIL1ovSEUDphRiOfALWwweDrSZWE8niuN9gyF3lbbybFqRYsEaNkOnZAeUN'], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': '', 'sectors': 41934848, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop3', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop4', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'dm-1', 'value': {'holders': [], 'host': '', 'links': {'ids': ['dm-name-ceph--7fbf4d63--7c6f--5c70--887c--2b73c978775b-osd--block--7fbf4d63--7c6f--5c70--887c--2b73c978775b', 'dm-uuid-LVM-aRcKlwv7ZYsrJYaz5OO903mOAEBnwLXI61utfRQTUZQKHwm4ceJ3NApz8UrwnN8y'], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': '', 'sectors': 41934848, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop5', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop6', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop0', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop0', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop1', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'dm-0', 'value': {'holders': [], 'host': '', 'links': {'ids': ['dm-name-ceph--752a9ec4--4bfc--5551--82cd--92b802bdad97-osd--block--752a9ec4--4bfc--5551--82cd--92b802bdad97', 'dm-uuid-LVM-jTt9fg6XlhiXbWhW0I9T02tzYsIeoKIoaf5BxtnvIAUk6EWNKNJMqSjVsZqTXHAa'], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': '', 'sectors': 41934848, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop7', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop2', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop1', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop3', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop4', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop5', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'dm-1', 'value': {'holders': [], 'host': '', 'links': {'ids': ['dm-name-ceph--ba85a0e3--2f57--51d5--a9f1--072f0e9a8deb-osd--block--ba85a0e3--2f57--51d5--a9f1--072f0e9a8deb', 'dm-uuid-LVM-YjmBzAWpiQhdCbdayUM0w3sqNqanvbLDieybFDy5wWcdDcB5LPRFHjaRE0WCOjBW'], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': '', 'sectors': 41934848, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop6', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop2', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sda', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_f90a4e80-1f36-4d14-b7da-de7b20e9fbca', 'scsi-SQEMU_QEMU_HARDDISK_f90a4e80-1f36-4d14-b7da-de7b20e9fbca'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {'sda1': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_f90a4e80-1f36-4d14-b7da-de7b20e9fbca-part1', 'scsi-SQEMU_QEMU_HARDDISK_f90a4e80-1f36-4d14-b7da-de7b20e9fbca-part1'], 'labels': ['cloudimg-rootfs'], 'masters': [], 'uuids': ['b852d8d2-8460-44aa-8998-23e4f04d73cf']}, 'sectors': 165672927, 'sectorsize': 512, 'size': '79.00 GB', 'start': '2099200', 'uuid': 'b852d8d2-8460-44aa-8998-23e4f04d73cf'}, 'sda14': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_f90a4e80-1f36-4d14-b7da-de7b20e9fbca-part14', 'scsi-SQEMU_QEMU_HARDDISK_f90a4e80-1f36-4d14-b7da-de7b20e9fbca-part14'], 'labels': [], 'masters': [], 'uuids': []}, 'sectors': 8192, 'sectorsize': 512, 'size': '4.00 MB', 'start': '2048', 'uuid': None}, 'sda15': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_f90a4e80-1f36-4d14-b7da-de7b20e9fbca-part15', 'scsi-SQEMU_QEMU_HARDDISK_f90a4e80-1f36-4d14-b7da-de7b20e9fbca-part15'], 'labels': ['UEFI'], 'masters': [], 'uuids': ['5C78-612A']}, 'sectors': 217088, 'sectorsize': 512, 'size': '106.00 MB', 'start': '10240', 'uuid': '5C78-612A'}, 'sda16': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_f90a4e80-1f36-4d14-b7da-de7b20e9fbca-part16', 'scsi-SQEMU_QEMU_HARDDISK_f90a4e80-1f36-4d14-b7da-de7b20e9fbca-part16'], 'labels': ['BOOT'], 'masters': [], 'uuids': ['09d53dc1-1e03-4286-bbb8-2b1796cf92ec']}, 'sectors': 1869825, 'sectorsize': 512, 'size': '913.00 MB', 'start': '227328', 'uuid': '09d53dc1-1e03-4286-bbb8-2b1796cf92ec'}}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 167772160, 'sectorsize': '512', 'size': '80.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop7', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'sda', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_2d259cae-7b79-479e-b978-3fffee75aa51', 'scsi-SQEMU_QEMU_HARDDISK_2d259cae-7b79-479e-b978-3fffee75aa51'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {'sda1': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_2d259cae-7b79-479e-b978-3fffee75aa51-part1', 'scsi-SQEMU_QEMU_HARDDISK_2d259cae-7b79-479e-b978-3fffee75aa51-part1'], 'labels': ['cloudimg-rootfs'], 'masters': [], 'uuids': ['b852d8d2-8460-44aa-8998-23e4f04d73cf']}, 'sectors': 165672927, 'sectorsize': 512, 'size': '79.00 GB', 'start': '2099200', 'uuid': 'b852d8d2-8460-44aa-8998-23e4f04d73cf'}, 'sda14': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_2d259cae-7b79-479e-b978-3fffee75aa51-part14', 'scsi-SQEMU_QEMU_HARDDISK_2d259cae-7b79-479e-b978-3fffee75aa51-part14'], 'labels': [], 'masters': [], 'uuids': []}, 'sectors': 8192, 'sectorsize': 512, 'size': '4.00 MB', 'start': '2048', 'uuid': None}, 'sda15': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_2d259cae-7b79-479e-b978-3fffee75aa51-part15', 'scsi-SQEMU_QEMU_HARDDISK_2d259cae-7b79-479e-b978-3fffee75aa51-part15'], 'labels': ['UEFI'], 'masters': [], 'uuids': ['5C78-612A']}, 'sectors': 217088, 'sectorsize': 512, 'size': '106.00 MB', 'start': '10240', 'uuid': '5C78-612A'}, 'sda16': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_2d259cae-7b79-479e-b978-3fffee75aa51-part16', 'scsi-SQEMU_QEMU_HARDDISK_2d259cae-7b79-479e-b978-3fffee75aa51-part16'], 'labels': ['BOOT'], 'masters': [], 'uuids': ['09d53dc1-1e03-4286-bbb8-2b1796cf92ec']}, 'sectors': 1869825, 'sectorsize': 512, 'size': '913.00 MB', 'start': '227328', 'uuid': '09d53dc1-1e03-4286-bbb8-2b1796cf92ec'}}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 167772160, 'sectorsize': '512', 'size': '80.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop0', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'sr0', 'value': {'holders': [], 'host': 'IDE interface: Intel Corporation 82371SB PIIX3 IDE [Natoma/Triton II]', 'links': {'ids': ['ata-QEMU_DVD-ROM_QM00001'], 'labels': ['config-2'], 'masters': [], 'uuids': ['2026-02-24-00-01-56-00']}, 'model': 'QEMU DVD-ROM', 'partitions': {}, 'removable': '1', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'mq-deadline', 'sectors': 253, 'sectorsize': '2048', 'size': '506.00 KB', 'support_discard': '0', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop3', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sdb', 'value': {'holders': ['ceph--9b92f24c--dac1--5bff--a390--34d9ff9e0f4f-osd--block--9b92f24c--dac1--5bff--a390--34d9ff9e0f4f'], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['lvm-pv-uuid-mec8nJ-xUip-qiWP-sj8w-uf5B-1Lt0-v8yfvH', 'scsi-0QEMU_QEMU_HARDDISK_9c6b6763-9704-41ec-9b6d-97eb80a5145c', 'scsi-SQEMU_QEMU_HARDDISK_9c6b6763-9704-41ec-9b6d-97eb80a5145c'], 'labels': [], 'masters': ['dm-0'], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop1', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop4', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop2', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop5', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sdc', 'value': {'holders': ['ceph--7f6a8fe5--6adc--52b4--b256--edcdbfd9bf4e-osd--block--7f6a8fe5--6adc--52b4--b256--edcdbfd9bf4e'], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['lvm-pv-uuid-HdfFdZ-81SJ-0Umt-qAzb-Dao2-7tUe-tHuhnG', 'scsi-0QEMU_QEMU_HARDDISK_6644f191-69e9-4f3a-baff-cffe4152595b', 'scsi-SQEMU_QEMU_HARDDISK_6644f191-69e9-4f3a-baff-cffe4152595b'], 'labels': [], 'masters': ['dm-1'], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop0', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop6', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop3', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop7', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop1', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sda', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_47e600ec-05ca-44a6-93d0-5ca3104e3639', 'scsi-SQEMU_QEMU_HARDDISK_47e600ec-05ca-44a6-93d0-5ca3104e3639'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {'sda1': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_47e600ec-05ca-44a6-93d0-5ca3104e3639-part1', 'scsi-SQEMU_QEMU_HARDDISK_47e600ec-05ca-44a6-93d0-5ca3104e3639-part1'], 'labels': ['cloudimg-rootfs'], 'masters': [], 'uuids': ['b852d8d2-8460-44aa-8998-23e4f04d73cf']}, 'sectors': 165672927, 'sectorsize': 512, 'size': '79.00 GB', 'start': '2099200', 'uuid': 'b852d8d2-8460-44aa-8998-23e4f04d73cf'}, 'sda14': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_47e600ec-05ca-44a6-93d0-5ca3104e3639-part14', 'scsi-SQEMU_QEMU_HARDDISK_47e600ec-05ca-44a6-93d0-5ca3104e3639-part14'], 'labels': [], 'masters': [], 'uuids': []}, 'sectors': 8192, 'sectorsize': 512, 'size': '4.00 MB', 'start': '2048', 'uuid': None}, 'sda15': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_47e600ec-05ca-44a6-93d0-5ca3104e3639-part15', 'scsi-SQEMU_QEMU_HARDDISK_47e600ec-05ca-44a6-93d0-5ca3104e3639-part15'], 'labels': ['UEFI'], 'masters': [], 'uuids': ['5C78-612A']}, 'sectors': 217088, 'sectorsize': 512, 'size': '106.00 MB', 'start': '10240', 'uuid': '5C78-612A'}, 'sda16': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_47e600ec-05ca-44a6-93d0-5ca3104e3639-part16', 'scsi-SQEMU_QEMU_HARDDISK_47e600ec-05ca-44a6-93d0-5ca3104e3639-part16'], 'labels': ['BOOT'], 'masters': [], 'uuids': ['09d53dc1-1e03-4286-bbb8-2b1796cf92ec']}, 'sectors': 1869825, 'sectorsize': 512, 'size': '913.00 MB', 'start': '227328', 'uuid': '09d53dc1-1e03-4286-bbb8-2b1796cf92ec'}}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 167772160, 'sectorsize': '512', 'size': '80.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop2', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sdb', 'value': {'holders': ['ceph--294dd48f--9d6d--5079--8214--59e60af9e526-osd--block--294dd48f--9d6d--5079--8214--59e60af9e526'], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['lvm-pv-uuid-1V6jIs-3p8q-l9Rl-OJ2n-8L4P-5xCz-K0YqYS', 'scsi-0QEMU_QEMU_HARDDISK_f659b61f-c39f-426f-a640-6a753951661d', 'scsi-SQEMU_QEMU_HARDDISK_f659b61f-c39f-426f-a640-6a753951661d'], 'labels': [], 'masters': ['dm-0'], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop4', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sdc', 'value': {'holders': ['ceph--7fbf4d63--7c6f--5c70--887c--2b73c978775b-osd--block--7fbf4d63--7c6f--5c70--887c--2b73c978775b'], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['lvm-pv-uuid-aHCB2h-b0pT-WCWd-loOK-A09H-yFYp-FH2wIv', 'scsi-0QEMU_QEMU_HARDDISK_fb58bf16-a1c6-4abb-925b-29c6a6a6f100', 'scsi-SQEMU_QEMU_HARDDISK_fb58bf16-a1c6-4abb-925b-29c6a6a6f100'], 'labels': [], 'masters': ['dm-1'], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop3', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sdd', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_5660bb62-e2c9-45b2-ae11-8ab0f829b194', 'scsi-SQEMU_QEMU_HARDDISK_5660bb62-e2c9-45b2-ae11-8ab0f829b194'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop5', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sr0', 'value': {'holders': [], 'host': 'IDE interface: Intel Corporation 82371SB PIIX3 IDE [Natoma/Triton II]', 'links': {'ids': ['ata-QEMU_DVD-ROM_QM00001'], 'labels': ['config-2'], 'masters': [], 'uuids': ['2026-02-24-00-01-59-00']}, 'model': 'QEMU DVD-ROM', 'partitions': {}, 'removable': '1', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'mq-deadline', 'sectors': 253, 'sectorsize': '2048', 'size': '506.00 KB', 'support_discard': '0', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop6', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sdd', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_4534650b-23a0-46c8-8bf1-44520caee637', 'scsi-SQEMU_QEMU_HARDDISK_4534650b-23a0-46c8-8bf1-44520caee637'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop7', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sda', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_7a9487cb-ec9e-4508-9c5a-f692700fc5a4', 'scsi-SQEMU_QEMU_HARDDISK_7a9487cb-ec9e-4508-9c5a-f692700fc5a4'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {'sda1': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_7a9487cb-ec9e-4508-9c5a-f692700fc5a4-part1', 'scsi-SQEMU_QEMU_HARDDISK_7a9487cb-ec9e-4508-9c5a-f692700fc5a4-part1'], 'labels': ['cloudimg-rootfs'], 'masters': [], 'uuids': ['b852d8d2-8460-44aa-8998-23e4f04d73cf']}, 'sectors': 165672927, 'sectorsize': 512, 'size': '79.00 GB', 'start': '2099200', 'uuid': 'b852d8d2-8460-44aa-8998-23e4f04d73cf'}, 'sda14': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_7a9487cb-ec9e-4508-9c5a-f692700fc5a4-part14', 'scsi-SQEMU_QEMU_HARDDISK_7a9487cb-ec9e-4508-9c5a-f692700fc5a4-part14'], 'labels': [], 'masters': [], 'uuids': []}, 'sectors': 8192, 'sectorsize': 512, 'size': '4.00 MB', 'start': '2048', 'uuid': None}, 'sda15': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_7a9487cb-ec9e-4508-9c5a-f692700fc5a4-part15', 'scsi-SQEMU_QEMU_HARDDISK_7a9487cb-ec9e-4508-9c5a-f692700fc5a4-part15'], 'labels': ['UEFI'], 'masters': [], 'uuids': ['5C78-612A']}, 'sectors': 217088, 'sectorsize': 512, 'size': '106.00 MB', 'start': '10240', 'uuid': '5C78-612A'}, 'sda16': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_7a9487cb-ec9e-4508-9c5a-f692700fc5a4-part16', 'scsi-SQEMU_QEMU_HARDDISK_7a9487cb-ec9e-4508-9c5a-f692700fc5a4-part16'], 'labels': ['BOOT'], 'masters': [], 'uuids': ['09d53dc1-1e03-4286-bbb8-2b1796cf92ec']}, 'sectors': 1869825, 'sectorsize': 512, 'size': '913.00 MB', 'start': '227328', 'uuid': '09d53dc1-1e03-4286-bbb8-2b1796cf92ec'}}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 167772160, 'sectorsize': '512', 'size': '80.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop0', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sdb', 'value': {'holders': ['ceph--752a9ec4--4bfc--5551--82cd--92b802bdad97-osd--block--752a9ec4--4bfc--5551--82cd--92b802bdad97'], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['lvm-pv-uuid-vfwhPS-pgoq-FNQw-W72Z-epjC-Ziiv-qWNph9', 'scsi-0QEMU_QEMU_HARDDISK_1558935b-4c66-4614-97e5-cda14abd2641', 'scsi-SQEMU_QEMU_HARDDISK_1558935b-4c66-4614-97e5-cda14abd2641'], 'labels': [], 'masters': ['dm-0'], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop4', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop1', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sdc', 'value': {'holders': ['ceph--ba85a0e3--2f57--51d5--a9f1--072f0e9a8deb-osd--block--ba85a0e3--2f57--51d5--a9f1--072f0e9a8deb'], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['lvm-pv-uuid-XuBysF-MTWF-hP8t-3CGS-0PEc-QL0q-u9CYq2', 'scsi-0QEMU_QEMU_HARDDISK_6f4643f0-329b-4ae2-97c7-d348648fd970', 'scsi-SQEMU_QEMU_HARDDISK_6f4643f0-329b-4ae2-97c7-d348648fd970'], 'labels': [], 'masters': ['dm-1'], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop2', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sdd', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_5f47789f-2a1a-442a-8a3c-644a4f92e5a8', 'scsi-SQEMU_QEMU_HARDDISK_5f47789f-2a1a-442a-8a3c-644a4f92e5a8'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sr0', 'value': {'holders': [], 'host': 'IDE interface: Intel Corporation 82371SB PIIX3 IDE [Natoma/Triton II]', 'links': {'ids': ['ata-QEMU_DVD-ROM_QM00001'], 'labels': ['config-2'], 'masters': [], 'uuids': ['2026-02-24-00-02-03-00']}, 'model': 'QEMU DVD-ROM', 'partitions': {}, 'removable': '1', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'mq-deadline', 'sectors': 253, 'sectorsize': '2048', 'size': '506.00 KB', 'support_discard': '0', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop3', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop5', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop6', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sr0', 'value': {'holders': [], 'host': 'IDE interface: Intel Corporation 82371SB PIIX3 IDE [Natoma/Triton II]', 'links': {'ids': ['ata-QEMU_DVD-ROM_QM00001'], 'labels': ['config-2'], 'masters': [], 'uuids': ['2026-02-24-00-01-58-00']}, 'model': 'QEMU DVD-ROM', 'partitions': {}, 'removable': '1', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'mq-deadline', 'sectors': 253, 'sectorsize': '2048', 'size': '506.00 KB', 'support_discard': '0', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop7', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'sda', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_6fef90f9-4ce5-456f-a170-152290f19313', 'scsi-SQEMU_QEMU_HARDDISK_6fef90f9-4ce5-456f-a170-152290f19313'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {'sda1': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_6fef90f9-4ce5-456f-a170-152290f19313-part1', 'scsi-SQEMU_QEMU_HARDDISK_6fef90f9-4ce5-456f-a170-152290f19313-part1'], 'labels': ['cloudimg-rootfs'], 'masters': [], 'uuids': ['b852d8d2-8460-44aa-8998-23e4f04d73cf']}, 'sectors': 165672927, 'sectorsize': 512, 'size': '79.00 GB', 'start': '2099200', 'uuid': 'b852d8d2-8460-44aa-8998-23e4f04d73cf'}, 'sda14': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_6fef90f9-4ce5-456f-a170-152290f19313-part14', 'scsi-SQEMU_QEMU_HARDDISK_6fef90f9-4ce5-456f-a170-152290f19313-part14'], 'labels': [], 'masters': [], 'uuids': []}, 'sectors': 8192, 'sectorsize': 512, 'size': '4.00 MB', 'start': '2048', 'uuid': None}, 'sda15': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_6fef90f9-4ce5-456f-a170-152290f19313-part15', 'scsi-SQEMU_QEMU_HARDDISK_6fef90f9-4ce5-456f-a170-152290f19313-part15'], 'labels': ['UEFI'], 'masters': [], 'uuids': ['5C78-612A']}, 'sectors': 217088, 'sectorsize': 512, 'size': '106.00 MB', 'start': '10240', 'uuid': '5C78-612A'}, 'sda16': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_6fef90f9-4ce5-456f-a170-152290f19313-part16', 'scsi-SQEMU_QEMU_HARDDISK_6fef90f9-4ce5-456f-a170-152290f19313-part16'], 'labels': ['BOOT'], 'masters': [], 'uuids': ['09d53dc1-1e03-4286-bbb8-2b1796cf92ec']}, 'sectors': 1869825, 'sectorsize': 512, 'size': '913.00 MB', 'start': '227328', 'uuid': '09d53dc1-1e03-4286-bbb8-2b1796cf92ec'}}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 167772160, 'sectorsize': '512', 'size': '80.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'sr0', 'value': {'holders': [], 'host': 'IDE interface: Intel Corporation 82371SB PIIX3 IDE [Natoma/Triton II]', 'links': {'ids': ['ata-QEMU_DVD-ROM_QM00001'], 'labels': ['config-2'], 'masters': [], 'uuids': ['2026-02-24-00-02-01-00']}, 'model': 'QEMU DVD-ROM', 'partitions': {}, 'removable': '1', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'mq-deadline', 'sectors': 253, 'sectorsize': '2048', 'size': '506.00 KB', 'support_discard': '0', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop4', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop5', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop6', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'loop7', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'sda', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_35b8d51e-716e-455d-9a95-995756ae5900', 'scsi-SQEMU_QEMU_HARDDISK_35b8d51e-716e-455d-9a95-995756ae5900'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {'sda1': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_35b8d51e-716e-455d-9a95-995756ae5900-part1', 'scsi-SQEMU_QEMU_HARDDISK_35b8d51e-716e-455d-9a95-995756ae5900-part1'], 'labels': ['cloudimg-rootfs'], 'masters': [], 'uuids': ['b852d8d2-8460-44aa-8998-23e4f04d73cf']}, 'sectors': 165672927, 'sectorsize': 512, 'size': '79.00 GB', 'start': '2099200', 'uuid': 'b852d8d2-8460-44aa-8998-23e4f04d73cf'}, 'sda14': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_35b8d51e-716e-455d-9a95-995756ae5900-part14', 'scsi-SQEMU_QEMU_HARDDISK_35b8d51e-716e-455d-9a95-995756ae5900-part14'], 'labels': [], 'masters': [], 'uuids': []}, 'sectors': 8192, 'sectorsize': 512, 'size': '4.00 MB', 'start': '2048', 'uuid': None}, 'sda15': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_35b8d51e-716e-455d-9a95-995756ae5900-part15', 'scsi-SQEMU_QEMU_HARDDISK_35b8d51e-716e-455d-9a95-995756ae5900-part15'], 'labels': ['UEFI'], 'masters': [], 'uuids': ['5C78-612A']}, 'sectors': 217088, 'sectorsize': 512, 'size': '106.00 MB', 'start': '10240', 'uuid': '5C78-612A'}, 'sda16': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_35b8d51e-716e-455d-9a95-995756ae5900-part16', 'scsi-SQEMU_QEMU_HARDDISK_35b8d51e-716e-455d-9a95-995756ae5900-part16'], 'labels': ['BOOT'], 'masters': [], 'uuids': ['09d53dc1-1e03-4286-bbb8-2b1796cf92ec']}, 'sectors': 1869825, 'sectorsize': 512, 'size': '913.00 MB', 'start': '227328', 'uuid': '09d53dc1-1e03-4286-bbb8-2b1796cf92ec'}}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 167772160, 'sectorsize': '512', 'size': '80.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'inventory_hostname in groups.get(osd_group_name, [])', 'item': {'key': 'sr0', 'value': {'holders': [], 'host': 'IDE interface: Intel Corporation 82371SB PIIX3 IDE [Natoma/Triton II]', 'links': {'ids': ['ata-QEMU_DVD-ROM_QM00001'], 'labels': ['config-2'], 'masters': [], 'uuids': ['2026-02-24-00-02-05-00']}, 'model': 'QEMU DVD-ROM', 'partitions': {}, 'removable': '1', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'mq-deadline', 'sectors': 253, 'sectorsize': '2048', 'size': '506.00 KB', 'support_discard': '0', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Check if the ceph conf exists] ******************************",
                                "Tuesday 24 February 2026  00:47:49 +0000 (0:00:01.257)       0:00:34.631 ****** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set default osd_pool_default_crush_rule fact] ***************",
                                "Tuesday 24 February 2026  00:47:50 +0000 (0:00:01.654)       0:00:36.285 ****** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Read osd pool default crush rule] ***************************",
                                "Tuesday 24 February 2026  00:47:52 +0000 (0:00:01.251)       0:00:37.537 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set osd_pool_default_crush_rule fact] ***********************",
                                "Tuesday 24 February 2026  00:47:53 +0000 (0:00:01.080)       0:00:38.617 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Read osd pool default crush rule] ***************************",
                                "Tuesday 24 February 2026  00:47:54 +0000 (0:00:00.863)       0:00:39.481 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set osd_pool_default_crush_rule fact] ***********************",
                                "Tuesday 24 February 2026  00:47:55 +0000 (0:00:01.067)       0:00:40.548 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _monitor_addresses - ipv4] *************************",
                                "Tuesday 24 February 2026  00:47:56 +0000 (0:00:01.307)       0:00:41.856 ****** ",
                                "\u001b[0;32mok: [testbed-node-4] => (item=testbed-node-0)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item=testbed-node-0)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item=testbed-node-1)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item=testbed-node-0)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item=testbed-node-2)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0] => (item=testbed-node-0)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item=testbed-node-1)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item=testbed-node-0)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item=testbed-node-2)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item=testbed-node-1)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0] => (item=testbed-node-1)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item=testbed-node-0)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item=testbed-node-1)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item=testbed-node-2)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0] => (item=testbed-node-2)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item=testbed-node-1)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item=testbed-node-2)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item=testbed-node-2)\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _monitor_addresses - ipv6] *************************",
                                "Tuesday 24 February 2026  00:48:00 +0000 (0:00:03.976)       0:00:45.832 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-0) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-1) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=testbed-node-0) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=testbed-node-1) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=testbed-node-0) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=testbed-node-2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=testbed-node-1) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=testbed-node-2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-0) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-1) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=testbed-node-0) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=testbed-node-1) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=testbed-node-2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=testbed-node-0) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=testbed-node-1) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=testbed-node-2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Import_tasks set_radosgw_address.yml] ***********************",
                                "Tuesday 24 February 2026  00:48:01 +0000 (0:00:01.151)       0:00:46.984 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-facts/tasks/set_radosgw_address.yml for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set current radosgw_address_block, radosgw_address, radosgw_interface  from node \"{{ ceph_dashboard_call_item }}\"] ***",
                                "Tuesday 24 February 2026  00:48:02 +0000 (0:00:01.304)       0:00:48.288 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _radosgw_address to radosgw_address_block ipv4] ****",
                                "Tuesday 24 February 2026  00:48:03 +0000 (0:00:00.621)       0:00:48.909 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _radosgw_address to radosgw_address_block ipv6] ****",
                                "Tuesday 24 February 2026  00:48:03 +0000 (0:00:00.367)       0:00:49.277 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _radosgw_address to radosgw_address] ***************",
                                "Tuesday 24 February 2026  00:48:04 +0000 (0:00:00.863)       0:00:50.141 ****** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _interface] ****************************************",
                                "Tuesday 24 February 2026  00:48:05 +0000 (0:00:00.791)       0:00:50.932 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-3) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-4) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-5) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _radosgw_address to radosgw_interface - ipv4] ******",
                                "Tuesday 24 February 2026  00:48:06 +0000 (0:00:00.474)       0:00:51.406 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-3) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-4) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-5) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _radosgw_address to radosgw_interface - ipv6] ******",
                                "Tuesday 24 February 2026  00:48:06 +0000 (0:00:00.718)       0:00:52.124 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-3) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-4) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-5) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Reset rgw_instances (workaround)] ***************************",
                                "Tuesday 24 February 2026  00:48:07 +0000 (0:00:00.851)       0:00:52.976 ****** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact rgw_instances] *************************************",
                                "Tuesday 24 February 2026  00:48:08 +0000 (0:00:00.656)       0:00:53.633 ****** ",
                                "\u001b[0;32mok: [testbed-node-4] => (item=0)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item=0)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item=0)\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact ceph_run_cmd] **************************************",
                                "Tuesday 24 February 2026  00:48:09 +0000 (0:00:01.533)       0:00:55.166 ****** ",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item=testbed-node-0)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-1(192.168.16.11)] => (item=testbed-node-1)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-2(192.168.16.12)] => (item=testbed-node-2)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item=testbed-node-3)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-4(192.168.16.14)] => (item=testbed-node-4)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-5(192.168.16.15)] => (item=testbed-node-5)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-manager(192.168.16.5)] => (item=testbed-manager)\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact ceph_admin_command] ********************************",
                                "Tuesday 24 February 2026  00:48:11 +0000 (0:00:01.558)       0:00:56.725 ****** ",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item=testbed-node-0)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-1(192.168.16.11)] => (item=testbed-node-1)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-2(192.168.16.12)] => (item=testbed-node-2)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item=testbed-node-3)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-4(192.168.16.14)] => (item=testbed-node-4)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-5(192.168.16.15)] => (item=testbed-node-5)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-manager(192.168.16.5)] => (item=testbed-manager)\u001b[0m",
                                "",
                                "TASK [ceph-handler : Include check_running_cluster.yml] ************************",
                                "Tuesday 24 February 2026  00:48:13 +0000 (0:00:02.119)       0:00:58.845 ****** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-handler/tasks/check_running_cluster.yml for testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [ceph-handler : Include check_running_containers.yml] *********************",
                                "Tuesday 24 February 2026  00:48:14 +0000 (0:00:01.000)       0:00:59.845 ****** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-handler/tasks/check_running_containers.yml for testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a mon container] ********************************",
                                "Tuesday 24 February 2026  00:48:15 +0000 (0:00:01.165)       0:01:01.011 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for an osd container] *******************************",
                                "Tuesday 24 February 2026  00:48:16 +0000 (0:00:01.167)       0:01:02.178 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a mds container] ********************************",
                                "Tuesday 24 February 2026  00:48:17 +0000 (0:00:00.863)       0:01:03.042 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a rgw container] ********************************",
                                "Tuesday 24 February 2026  00:48:18 +0000 (0:00:00.699)       0:01:03.742 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a mgr container] ********************************",
                                "Tuesday 24 February 2026  00:48:19 +0000 (0:00:00.783)       0:01:04.525 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a rbd mirror container] *************************",
                                "Tuesday 24 February 2026  00:48:20 +0000 (0:00:01.199)       0:01:05.725 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a nfs container] ********************************",
                                "Tuesday 24 February 2026  00:48:21 +0000 (0:00:00.689)       0:01:06.414 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a ceph-crash container] *************************",
                                "Tuesday 24 February 2026  00:48:21 +0000 (0:00:00.724)       0:01:07.138 ****** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a ceph-exporter container] **********************",
                                "Tuesday 24 February 2026  00:48:22 +0000 (0:00:01.029)       0:01:08.168 ****** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Include check_socket_non_container.yml] *******************",
                                "Tuesday 24 February 2026  00:48:23 +0000 (0:00:01.139)       0:01:09.308 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_mon_status] ******************************",
                                "Tuesday 24 February 2026  00:48:24 +0000 (0:00:00.519)       0:01:09.828 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_osd_status] ******************************",
                                "Tuesday 24 February 2026  00:48:25 +0000 (0:00:00.774)       0:01:10.602 ****** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_mds_status] ******************************",
                                "Tuesday 24 February 2026  00:48:25 +0000 (0:00:00.542)       0:01:11.144 ****** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_rgw_status] ******************************",
                                "Tuesday 24 February 2026  00:48:26 +0000 (0:00:00.734)       0:01:11.879 ****** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_nfs_status] ******************************",
                                "Tuesday 24 February 2026  00:48:27 +0000 (0:00:00.864)       0:01:12.743 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_rbd_status] ******************************",
                                "Tuesday 24 February 2026  00:48:28 +0000 (0:00:00.962)       0:01:13.706 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_mgr_status] ******************************",
                                "Tuesday 24 February 2026  00:48:28 +0000 (0:00:00.629)       0:01:14.335 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_crash_status] ****************************",
                                "Tuesday 24 February 2026  00:48:29 +0000 (0:00:00.830)       0:01:15.165 ****** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_exporter_status] *************************",
                                "Tuesday 24 February 2026  00:48:30 +0000 (0:00:00.755)       0:01:15.921 ****** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Generate systemd ceph target file] ***************",
                                "Tuesday 24 February 2026  00:48:32 +0000 (0:00:01.558)       0:01:17.479 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Enable ceph.target] ******************************",
                                "Tuesday 24 February 2026  00:48:34 +0000 (0:00:02.107)       0:01:19.587 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Include prerequisites.yml] ***********************",
                                "Tuesday 24 February 2026  00:48:37 +0000 (0:00:03.160)       0:01:22.748 ****** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-container-common/tasks/prerequisites.yml for testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Stop lvmetad] ************************************",
                                "Tuesday 24 February 2026  00:48:38 +0000 (0:00:01.606)       0:01:24.354 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Disable and mask lvmetad service] ****************",
                                "Tuesday 24 February 2026  00:48:39 +0000 (0:00:00.682)       0:01:25.037 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Remove ceph udev rules] **************************",
                                "Tuesday 24 February 2026  00:48:40 +0000 (0:00:01.144)       0:01:26.182 ****** ",
                                "\u001b[0;32mok: [testbed-node-3] => (item=/usr/lib/udev/rules.d/95-ceph-osd.rules)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item=/usr/lib/udev/rules.d/95-ceph-osd.rules)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item=/usr/lib/udev/rules.d/95-ceph-osd.rules)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0] => (item=/usr/lib/udev/rules.d/95-ceph-osd.rules)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item=/usr/lib/udev/rules.d/60-ceph-by-parttypeuuid.rules)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item=/usr/lib/udev/rules.d/60-ceph-by-parttypeuuid.rules)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item=/usr/lib/udev/rules.d/95-ceph-osd.rules)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item=/usr/lib/udev/rules.d/60-ceph-by-parttypeuuid.rules)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0] => (item=/usr/lib/udev/rules.d/60-ceph-by-parttypeuuid.rules)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item=/usr/lib/udev/rules.d/95-ceph-osd.rules)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item=/usr/lib/udev/rules.d/60-ceph-by-parttypeuuid.rules)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item=/usr/lib/udev/rules.d/60-ceph-by-parttypeuuid.rules)\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Ensure tmpfiles.d is present] ********************",
                                "Tuesday 24 February 2026  00:48:42 +0000 (0:00:01.601)       0:01:27.783 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Restore certificates selinux context] ************",
                                "Tuesday 24 February 2026  00:48:43 +0000 (0:00:01.176)       0:01:28.960 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Install python3 on osd nodes] ********************",
                                "Tuesday 24 February 2026  00:48:44 +0000 (0:00:00.521)       0:01:29.481 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Include registry.yml] ****************************",
                                "Tuesday 24 February 2026  00:48:44 +0000 (0:00:00.825)       0:01:30.306 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Include fetch_image.yml] *************************",
                                "Tuesday 24 February 2026  00:48:45 +0000 (0:00:00.512)       0:01:30.819 ****** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-container-common/tasks/fetch_image.yml for testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Pulling Ceph container image] ********************",
                                "Tuesday 24 February 2026  00:48:46 +0000 (0:00:01.045)       0:01:31.865 ****** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Pulling alertmanager/prometheus/grafana container images] ***",
                                "Tuesday 24 February 2026  00:49:26 +0000 (0:00:40.427)       0:02:12.292 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=docker.io/prom/alertmanager:v0.16.2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=docker.io/prom/prometheus:v2.7.2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=docker.io/grafana/grafana:6.7.4) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=docker.io/prom/alertmanager:v0.16.2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=docker.io/prom/prometheus:v2.7.2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=docker.io/grafana/grafana:6.7.4) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=docker.io/prom/alertmanager:v0.16.2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=docker.io/prom/prometheus:v2.7.2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=docker.io/grafana/grafana:6.7.4) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=docker.io/prom/alertmanager:v0.16.2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=docker.io/prom/prometheus:v2.7.2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=docker.io/grafana/grafana:6.7.4) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=docker.io/prom/alertmanager:v0.16.2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=docker.io/prom/prometheus:v2.7.2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=docker.io/grafana/grafana:6.7.4) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=docker.io/prom/alertmanager:v0.16.2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=docker.io/prom/prometheus:v2.7.2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=docker.io/grafana/grafana:6.7.4) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Pulling node-exporter container image] ***********",
                                "Tuesday 24 February 2026  00:49:27 +0000 (0:00:00.585)       0:02:12.877 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Export local ceph dev image] *********************",
                                "Tuesday 24 February 2026  00:49:28 +0000 (0:00:00.635)       0:02:13.512 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Copy ceph dev image file] ************************",
                                "Tuesday 24 February 2026  00:49:28 +0000 (0:00:00.136)       0:02:13.649 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Load ceph dev image] *****************************",
                                "Tuesday 24 February 2026  00:49:28 +0000 (0:00:00.520)       0:02:14.169 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Remove tmp ceph dev image file] ******************",
                                "Tuesday 24 February 2026  00:49:29 +0000 (0:00:00.795)       0:02:14.965 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Get ceph version] ********************************",
                                "Tuesday 24 February 2026  00:49:30 +0000 (0:00:00.597)       0:02:15.563 ****** ",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Set_fact ceph_version ceph_version.stdout.split] ***",
                                "Tuesday 24 February 2026  00:49:32 +0000 (0:00:02.522)       0:02:18.086 ****** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Include release.yml] *****************************",
                                "Tuesday 24 February 2026  00:49:33 +0000 (0:00:00.633)       0:02:18.719 ****** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-container-common/tasks/release.yml for testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Set_fact ceph_release jewel] *********************",
                                "Tuesday 24 February 2026  00:49:34 +0000 (0:00:01.159)       0:02:19.878 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Set_fact ceph_release kraken] ********************",
                                "Tuesday 24 February 2026  00:49:35 +0000 (0:00:00.656)       0:02:20.535 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Set_fact ceph_release luminous] ******************",
                                "Tuesday 24 February 2026  00:49:35 +0000 (0:00:00.538)       0:02:21.074 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Set_fact ceph_release mimic] *********************",
                                "Tuesday 24 February 2026  00:49:36 +0000 (0:00:00.681)       0:02:21.755 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Set_fact ceph_release nautilus] ******************",
                                "Tuesday 24 February 2026  00:49:37 +0000 (0:00:00.658)       0:02:22.413 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Set_fact ceph_release octopus] *******************",
                                "Tuesday 24 February 2026  00:49:37 +0000 (0:00:00.641)       0:02:23.055 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Set_fact ceph_release pacific] *******************",
                                "Tuesday 24 February 2026  00:49:38 +0000 (0:00:00.554)       0:02:23.609 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Set_fact ceph_release quincy] ********************",
                                "Tuesday 24 February 2026  00:49:38 +0000 (0:00:00.676)       0:02:24.286 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-container-common : Set_fact ceph_release reef] **********************",
                                "Tuesday 24 February 2026  00:49:39 +0000 (0:00:00.883)       0:02:25.170 ****** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-config : Include create_ceph_initial_dirs.yml] **********************",
                                "Tuesday 24 February 2026  00:49:40 +0000 (0:00:01.073)       0:02:26.243 ****** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-config/tasks/create_ceph_initial_dirs.yml for testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [ceph-config : Create ceph initial directories] ***************************",
                                "Tuesday 24 February 2026  00:49:41 +0000 (0:00:00.943)       0:02:27.187 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=/etc/ceph)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=/etc/ceph)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=/etc/ceph)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=/etc/ceph)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=/etc/ceph)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=/etc/ceph)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=/var/lib/ceph/)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=/var/lib/ceph/)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=/var/lib/ceph/)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=/var/lib/ceph/)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=/var/lib/ceph/)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=/var/lib/ceph/)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=/var/lib/ceph/mon)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=/var/lib/ceph/mon)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=/var/lib/ceph/mon)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=/var/lib/ceph/mon)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=/var/lib/ceph/mon)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=/var/lib/ceph/mon)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=/var/lib/ceph/osd)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=/var/lib/ceph/osd)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=/var/lib/ceph/osd)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=/var/lib/ceph/osd)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=/var/lib/ceph/osd)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=/var/lib/ceph/osd)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=/var/lib/ceph/mds)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=/var/lib/ceph/mds)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=/var/lib/ceph/mds)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=/var/lib/ceph/mds)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=/var/lib/ceph/mds)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=/var/lib/ceph/mds)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=/var/lib/ceph/tmp)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=/var/lib/ceph/tmp)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=/var/lib/ceph/tmp)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=/var/lib/ceph/tmp)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=/var/lib/ceph/tmp)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=/var/lib/ceph/tmp)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=/var/lib/ceph/crash)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=/var/lib/ceph/crash)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=/var/lib/ceph/crash)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=/var/lib/ceph/crash)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=/var/lib/ceph/crash)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=/var/lib/ceph/radosgw)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=/var/lib/ceph/radosgw)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=/var/lib/ceph/radosgw)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=/var/lib/ceph/radosgw)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=/var/lib/ceph/crash)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=/var/lib/ceph/bootstrap-rgw)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=/var/lib/ceph/radosgw)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=/var/lib/ceph/bootstrap-rgw)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=/var/lib/ceph/bootstrap-rgw)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=/var/lib/ceph/bootstrap-rgw)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=/var/lib/ceph/radosgw)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=/var/lib/ceph/bootstrap-rgw)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=/var/lib/ceph/bootstrap-mgr)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=/var/lib/ceph/bootstrap-mgr)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=/var/lib/ceph/bootstrap-mgr)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=/var/lib/ceph/bootstrap-mgr)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=/var/lib/ceph/bootstrap-rgw)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=/var/lib/ceph/bootstrap-mgr)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=/var/lib/ceph/bootstrap-mds)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=/var/lib/ceph/bootstrap-mds)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=/var/lib/ceph/bootstrap-mds)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=/var/lib/ceph/bootstrap-mds)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=/var/lib/ceph/bootstrap-mgr)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=/var/lib/ceph/bootstrap-mds)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=/var/lib/ceph/bootstrap-osd)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=/var/lib/ceph/bootstrap-osd)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=/var/lib/ceph/bootstrap-osd)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=/var/lib/ceph/bootstrap-osd)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=/var/lib/ceph/bootstrap-osd)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=/var/lib/ceph/bootstrap-mds)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=/var/lib/ceph/bootstrap-rbd)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=/var/lib/ceph/bootstrap-rbd)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=/var/lib/ceph/bootstrap-rbd)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=/var/lib/ceph/bootstrap-rbd)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=/var/lib/ceph/bootstrap-osd)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=/var/lib/ceph/bootstrap-rbd)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=/var/lib/ceph/bootstrap-rbd-mirror)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=/var/lib/ceph/bootstrap-rbd-mirror)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=/var/lib/ceph/bootstrap-rbd-mirror)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=/var/lib/ceph/bootstrap-rbd-mirror)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=/var/lib/ceph/bootstrap-rbd-mirror)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=/var/lib/ceph/bootstrap-rbd)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=/var/run/ceph)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=/var/run/ceph)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=/var/run/ceph)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=/var/run/ceph)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=/var/run/ceph)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=/var/log/ceph)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=/var/lib/ceph/bootstrap-rbd-mirror)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=/var/log/ceph)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=/var/log/ceph)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=/var/log/ceph)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=/var/log/ceph)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=/var/run/ceph)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=/var/log/ceph)\u001b[0m",
                                "",
                                "TASK [ceph-config : Include_tasks rgw_systemd_environment_file.yml] ************",
                                "Tuesday 24 February 2026  00:49:48 +0000 (0:00:06.659)       0:02:33.847 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-config/tasks/rgw_systemd_environment_file.yml for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m",
                                "",
                                "TASK [ceph-config : Create rados gateway instance directories] *****************",
                                "Tuesday 24 February 2026  00:49:49 +0000 (0:00:00.865)       0:02:34.713 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'instance_name': 'rgw0', 'radosgw_address': '192.168.16.13', 'radosgw_frontend_port': 8081})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'instance_name': 'rgw0', 'radosgw_address': '192.168.16.14', 'radosgw_frontend_port': 8081})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'instance_name': 'rgw0', 'radosgw_address': '192.168.16.15', 'radosgw_frontend_port': 8081})\u001b[0m",
                                "",
                                "TASK [ceph-config : Generate environment file] *********************************",
                                "Tuesday 24 February 2026  00:49:50 +0000 (0:00:00.933)       0:02:35.647 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'instance_name': 'rgw0', 'radosgw_address': '192.168.16.13', 'radosgw_frontend_port': 8081})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'instance_name': 'rgw0', 'radosgw_address': '192.168.16.14', 'radosgw_frontend_port': 8081})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'instance_name': 'rgw0', 'radosgw_address': '192.168.16.15', 'radosgw_frontend_port': 8081})\u001b[0m",
                                "",
                                "TASK [ceph-config : Reset num_osds] ********************************************",
                                "Tuesday 24 February 2026  00:49:51 +0000 (0:00:01.067)       0:02:36.715 ****** ",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-config : Count number of osds for lvm scenario] *********************",
                                "Tuesday 24 February 2026  00:49:52 +0000 (0:00:00.715)       0:02:37.430 ****** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-config : Look up for ceph-volume rejected devices] ******************",
                                "Tuesday 24 February 2026  00:49:52 +0000 (0:00:00.870)       0:02:38.301 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-config : Set_fact rejected_devices] *********************************",
                                "Tuesday 24 February 2026  00:49:53 +0000 (0:00:00.652)       0:02:38.953 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-config : Set_fact _devices] *****************************************",
                                "Tuesday 24 February 2026  00:49:54 +0000 (0:00:00.605)       0:02:39.559 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-config : Run 'ceph-volume lvm batch --report' to see how many osds are to be created] ***",
                                "Tuesday 24 February 2026  00:49:54 +0000 (0:00:00.550)       0:02:40.109 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-config : Set_fact num_osds from the output of 'ceph-volume lvm batch --report' (legacy report)] ***",
                                "Tuesday 24 February 2026  00:49:55 +0000 (0:00:00.576)       0:02:40.686 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-config : Set_fact num_osds from the output of 'ceph-volume lvm batch --report' (new report)] ***",
                                "Tuesday 24 February 2026  00:49:55 +0000 (0:00:00.486)       0:02:41.172 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-config : Run 'ceph-volume lvm list' to see how many osds have already been created] ***",
                                "Tuesday 24 February 2026  00:49:56 +0000 (0:00:00.578)       0:02:41.751 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-config : Set_fact num_osds (add existing osds)] *********************",
                                "Tuesday 24 February 2026  00:49:59 +0000 (0:00:02.995)       0:02:44.746 ****** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-config : Set_fact _osd_memory_target] *******************************",
                                "Tuesday 24 February 2026  00:50:00 +0000 (0:00:00.780)       0:02:45.527 ****** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-config : Set osd_memory_target to cluster host config] **************",
                                "Tuesday 24 February 2026  00:50:00 +0000 (0:00:00.538)       0:02:46.065 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-config : Render rgw configs] ****************************************",
                                "Tuesday 24 February 2026  00:50:01 +0000 (0:00:00.763)       0:02:46.829 ****** ",
                                "\u001b[0;32mok: [testbed-node-3] => (item={'instance_name': 'rgw0', 'radosgw_address': '192.168.16.13', 'radosgw_frontend_port': 8081})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item={'instance_name': 'rgw0', 'radosgw_address': '192.168.16.14', 'radosgw_frontend_port': 8081})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item={'instance_name': 'rgw0', 'radosgw_address': '192.168.16.15', 'radosgw_frontend_port': 8081})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-config : Set config to cluster] *************************************",
                                "Tuesday 24 February 2026  00:50:02 +0000 (0:00:00.628)       0:02:47.457 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=[{'key': 'client.rgw.default.testbed-node-3.rgw0', 'value': {'log_file': '/var/log/ceph/ceph-rgw-default-testbed-node-3.rgw0.log', 'rgw_frontends': 'beast endpoint=192.168.16.13:8081'}}, {'key': 'log_file', 'value': '/var/log/ceph/ceph-rgw-default-testbed-node-3.rgw0.log'}]) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=[{'key': 'client.rgw.default.testbed-node-3.rgw0', 'value': {'log_file': '/var/log/ceph/ceph-rgw-default-testbed-node-3.rgw0.log', 'rgw_frontends': 'beast endpoint=192.168.16.13:8081'}}, {'key': 'rgw_frontends', 'value': 'beast endpoint=192.168.16.13:8081'}]) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=[{'key': 'client.rgw.default.testbed-node-4.rgw0', 'value': {'log_file': '/var/log/ceph/ceph-rgw-default-testbed-node-4.rgw0.log', 'rgw_frontends': 'beast endpoint=192.168.16.14:8081'}}, {'key': 'log_file', 'value': '/var/log/ceph/ceph-rgw-default-testbed-node-4.rgw0.log'}]) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=[{'key': 'client.rgw.default.testbed-node-4.rgw0', 'value': {'log_file': '/var/log/ceph/ceph-rgw-default-testbed-node-4.rgw0.log', 'rgw_frontends': 'beast endpoint=192.168.16.14:8081'}}, {'key': 'rgw_frontends', 'value': 'beast endpoint=192.168.16.14:8081'}]) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=[{'key': 'client.rgw.default.testbed-node-5.rgw0', 'value': {'log_file': '/var/log/ceph/ceph-rgw-default-testbed-node-5.rgw0.log', 'rgw_frontends': 'beast endpoint=192.168.16.15:8081'}}, {'key': 'log_file', 'value': '/var/log/ceph/ceph-rgw-default-testbed-node-5.rgw0.log'}]) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=[{'key': 'client.rgw.default.testbed-node-5.rgw0', 'value': {'log_file': '/var/log/ceph/ceph-rgw-default-testbed-node-5.rgw0.log', 'rgw_frontends': 'beast endpoint=192.168.16.15:8081'}}, {'key': 'rgw_frontends', 'value': 'beast endpoint=192.168.16.15:8081'}]) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-config : Set rgw configs to file] ***********************************",
                                "Tuesday 24 February 2026  00:50:02 +0000 (0:00:00.687)       0:02:48.145 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-config : Create ceph conf directory] ********************************",
                                "Tuesday 24 February 2026  00:50:03 +0000 (0:00:00.744)       0:02:48.890 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set current radosgw_address_block, radosgw_address, radosgw_interface  from node \"{{ ceph_dashboard_call_item }}\"] ***",
                                "Tuesday 24 February 2026  00:50:04 +0000 (0:00:00.953)       0:02:49.843 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _radosgw_address to radosgw_address_block ipv4] ****",
                                "Tuesday 24 February 2026  00:50:05 +0000 (0:00:00.716)       0:02:50.559 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _radosgw_address to radosgw_address_block ipv6] ****",
                                "Tuesday 24 February 2026  00:50:06 +0000 (0:00:00.998)       0:02:51.558 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _radosgw_address to radosgw_address] ***************",
                                "Tuesday 24 February 2026  00:50:07 +0000 (0:00:00.838)       0:02:52.396 ****** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _interface] ****************************************",
                                "Tuesday 24 February 2026  00:50:08 +0000 (0:00:01.233)       0:02:53.630 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-3) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-4) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-5) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _radosgw_address to radosgw_interface - ipv4] ******",
                                "Tuesday 24 February 2026  00:50:08 +0000 (0:00:00.358)       0:02:53.988 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-3) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-4) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-5) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _radosgw_address to radosgw_interface - ipv6] ******",
                                "Tuesday 24 February 2026  00:50:08 +0000 (0:00:00.351)       0:02:54.339 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-3) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-4) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-5) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Reset rgw_instances (workaround)] ***************************",
                                "Tuesday 24 February 2026  00:50:09 +0000 (0:00:00.355)       0:02:54.694 ****** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact rgw_instances] *************************************",
                                "Tuesday 24 February 2026  00:50:09 +0000 (0:00:00.439)       0:02:55.134 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=0) \u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item=0)\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item=0)\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=0) \u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item=0)\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=0) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-config : Generate Ceph file] ****************************************",
                                "Tuesday 24 February 2026  00:50:11 +0000 (0:00:01.794)       0:02:56.929 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Make tempdir for scripts] **********************",
                                "Tuesday 24 February 2026  00:50:14 +0000 (0:00:03.310)       0:03:00.240 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Mons handler] **********************************",
                                "Tuesday 24 February 2026  00:50:16 +0000 (0:00:01.379)       0:03:01.619 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-handler/tasks/handler_mons.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Set _mon_handler_called before restart] ********",
                                "Tuesday 24 February 2026  00:50:17 +0000 (0:00:00.940)       0:03:02.560 ****** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Copy mon restart script] ***********************",
                                "Tuesday 24 February 2026  00:50:17 +0000 (0:00:00.281)       0:03:02.841 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Restart ceph mon daemon(s)] ********************",
                                "Tuesday 24 February 2026  00:50:18 +0000 (0:00:01.312)       0:03:04.154 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-0) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-1) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Set _mon_handler_called after restart] *********",
                                "Tuesday 24 February 2026  00:50:19 +0000 (0:00:00.950)       0:03:05.104 ****** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Osds handler] **********************************",
                                "Tuesday 24 February 2026  00:50:20 +0000 (0:00:00.312)       0:03:05.416 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-handler/tasks/handler_osds.yml for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Set_fact trigger_restart] **********************",
                                "Tuesday 24 February 2026  00:50:21 +0000 (0:00:00.992)       0:03:06.409 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-3) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-4) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-5) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Set _osd_handler_called before restart] ********",
                                "Tuesday 24 February 2026  00:50:21 +0000 (0:00:00.419)       0:03:06.828 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Unset noup flag] *******************************",
                                "Tuesday 24 February 2026  00:50:21 +0000 (0:00:00.378)       0:03:07.207 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Copy osd restart script] ***********************",
                                "Tuesday 24 February 2026  00:50:22 +0000 (0:00:00.218)       0:03:07.425 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Get pool list] *********************************",
                                "Tuesday 24 February 2026  00:50:22 +0000 (0:00:00.322)       0:03:07.748 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Get balancer module status] ********************",
                                "Tuesday 24 February 2026  00:50:22 +0000 (0:00:00.220)       0:03:07.968 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Set_fact pools_pgautoscaler_mode] **************",
                                "Tuesday 24 February 2026  00:50:22 +0000 (0:00:00.267)       0:03:08.236 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Disable balancer] ******************************",
                                "Tuesday 24 February 2026  00:50:23 +0000 (0:00:00.149)       0:03:08.385 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Disable pg autoscale on pools] *****************",
                                "Tuesday 24 February 2026  00:50:23 +0000 (0:00:00.750)       0:03:09.135 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Restart ceph osds daemon(s)] *******************",
                                "Tuesday 24 February 2026  00:50:24 +0000 (0:00:00.238)       0:03:09.374 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-3) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-4) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-5) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Set _osd_handler_called after restart] *********",
                                "Tuesday 24 February 2026  00:50:24 +0000 (0:00:00.440)       0:03:09.814 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Re-enable pg autoscale on pools] ***************",
                                "Tuesday 24 February 2026  00:50:24 +0000 (0:00:00.336)       0:03:10.151 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Re-enable balancer] ****************************",
                                "Tuesday 24 February 2026  00:50:25 +0000 (0:00:00.245)       0:03:10.396 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Mdss handler] **********************************",
                                "Tuesday 24 February 2026  00:50:25 +0000 (0:00:00.205)       0:03:10.602 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-handler/tasks/handler_mdss.yml for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Set _mds_handler_called before restart] ********",
                                "Tuesday 24 February 2026  00:50:26 +0000 (0:00:01.091)       0:03:11.694 ****** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Copy mds restart script] ***********************",
                                "Tuesday 24 February 2026  00:50:26 +0000 (0:00:00.344)       0:03:12.038 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Restart ceph mds daemon(s)] ********************",
                                "Tuesday 24 February 2026  00:50:27 +0000 (0:00:01.250)       0:03:13.289 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-3) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-4) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-5) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Set _mds_handler_called after restart] *********",
                                "Tuesday 24 February 2026  00:50:28 +0000 (0:00:00.756)       0:03:14.046 ****** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Rgws handler] **********************************",
                                "Tuesday 24 February 2026  00:50:29 +0000 (0:00:00.412)       0:03:14.458 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-handler/tasks/handler_rgws.yml for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Set _rgw_handler_called before restart] ********",
                                "Tuesday 24 February 2026  00:50:30 +0000 (0:00:01.223)       0:03:15.681 ****** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Copy rgw restart script] ***********************",
                                "Tuesday 24 February 2026  00:50:30 +0000 (0:00:00.475)       0:03:16.157 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Restart ceph rgw daemon(s)] ********************",
                                "Tuesday 24 February 2026  00:50:32 +0000 (0:00:02.006)       0:03:18.163 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-3) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-4) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-5) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Set _rgw_handler_called after restart] *********",
                                "Tuesday 24 February 2026  00:50:33 +0000 (0:00:00.527)       0:03:18.691 ****** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Rbdmirrors handler] ****************************",
                                "Tuesday 24 February 2026  00:50:33 +0000 (0:00:00.302)       0:03:18.993 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Mgrs handler] **********************************",
                                "Tuesday 24 February 2026  00:50:34 +0000 (0:00:00.668)       0:03:19.662 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-handler/tasks/handler_mgrs.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Set _mgr_handler_called before restart] ********",
                                "Tuesday 24 February 2026  00:50:35 +0000 (0:00:00.754)       0:03:20.417 ****** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Copy mgr restart script] ***********************",
                                "Tuesday 24 February 2026  00:50:35 +0000 (0:00:00.543)       0:03:20.961 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Restart ceph mgr daemon(s)] ********************",
                                "Tuesday 24 February 2026  00:50:36 +0000 (0:00:01.261)       0:03:22.222 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-0) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-1) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Set _mgr_handler_called after restart] *********",
                                "Tuesday 24 February 2026  00:50:37 +0000 (0:00:00.600)       0:03:22.823 ****** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "PLAY [Apply role ceph-mon] *****************************************************",
                                "",
                                "TASK [ceph-handler : Include check_running_cluster.yml] ************************",
                                "Tuesday 24 February 2026  00:50:38 +0000 (0:00:00.672)       0:03:23.496 ****** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-handler/tasks/check_running_cluster.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [ceph-handler : Include check_running_containers.yml] *********************",
                                "Tuesday 24 February 2026  00:50:38 +0000 (0:00:00.742)       0:03:24.239 ****** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-handler/tasks/check_running_containers.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a mon container] ********************************",
                                "Tuesday 24 February 2026  00:50:39 +0000 (0:00:00.536)       0:03:24.776 ****** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for an osd container] *******************************",
                                "Tuesday 24 February 2026  00:50:40 +0000 (0:00:00.951)       0:03:25.728 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a mds container] ********************************",
                                "Tuesday 24 February 2026  00:50:40 +0000 (0:00:00.320)       0:03:26.048 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a rgw container] ********************************",
                                "Tuesday 24 February 2026  00:50:40 +0000 (0:00:00.269)       0:03:26.318 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a mgr container] ********************************",
                                "Tuesday 24 February 2026  00:50:41 +0000 (0:00:00.298)       0:03:26.617 ****** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a rbd mirror container] *************************",
                                "Tuesday 24 February 2026  00:50:42 +0000 (0:00:00.929)       0:03:27.546 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a nfs container] ********************************",
                                "Tuesday 24 February 2026  00:50:42 +0000 (0:00:00.272)       0:03:27.819 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a ceph-crash container] *************************",
                                "Tuesday 24 February 2026  00:50:42 +0000 (0:00:00.250)       0:03:28.069 ****** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a ceph-exporter container] **********************",
                                "Tuesday 24 February 2026  00:50:43 +0000 (0:00:00.630)       0:03:28.700 ****** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Include check_socket_non_container.yml] *******************",
                                "Tuesday 24 February 2026  00:50:44 +0000 (0:00:00.898)       0:03:29.599 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_mon_status] ******************************",
                                "Tuesday 24 February 2026  00:50:44 +0000 (0:00:00.263)       0:03:29.863 ****** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_osd_status] ******************************",
                                "Tuesday 24 February 2026  00:50:44 +0000 (0:00:00.379)       0:03:30.242 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_mds_status] ******************************",
                                "Tuesday 24 February 2026  00:50:45 +0000 (0:00:00.227)       0:03:30.470 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_rgw_status] ******************************",
                                "Tuesday 24 February 2026  00:50:45 +0000 (0:00:00.253)       0:03:30.723 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_nfs_status] ******************************",
                                "Tuesday 24 February 2026  00:50:45 +0000 (0:00:00.431)       0:03:31.155 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_rbd_status] ******************************",
                                "Tuesday 24 February 2026  00:50:46 +0000 (0:00:00.272)       0:03:31.428 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_mgr_status] ******************************",
                                "Tuesday 24 February 2026  00:50:46 +0000 (0:00:00.502)       0:03:31.930 ****** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_crash_status] ****************************",
                                "Tuesday 24 February 2026  00:50:46 +0000 (0:00:00.330)       0:03:32.261 ****** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_exporter_status] *************************",
                                "Tuesday 24 February 2026  00:50:47 +0000 (0:00:00.514)       0:03:32.775 ****** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-mon : Set_fact container_exec_cmd] **********************************",
                                "Tuesday 24 February 2026  00:50:47 +0000 (0:00:00.516)       0:03:33.292 ****** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-mon : Include deploy_monitors.yml] **********************************",
                                "Tuesday 24 February 2026  00:50:48 +0000 (0:00:00.292)       0:03:33.584 ****** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-mon/tasks/deploy_monitors.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [ceph-mon : Check if monitor initial keyring already exists] **************",
                                "Tuesday 24 February 2026  00:50:48 +0000 (0:00:00.699)       0:03:34.283 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-mon : Generate monitor initial keyring] *****************************",
                                "Tuesday 24 February 2026  00:50:49 +0000 (0:00:00.112)       0:03:34.395 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0 -> localhost]\u001b[0m",
                                "",
                                "TASK [ceph-mon : Set_fact _initial_mon_key_success] ****************************",
                                "Tuesday 24 February 2026  00:50:49 +0000 (0:00:00.790)       0:03:35.186 ****** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-mon : Get initial keyring when it already exists] *******************",
                                "Tuesday 24 February 2026  00:50:50 +0000 (0:00:00.299)       0:03:35.486 ****** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-mon : Create monitor initial keyring] *******************************",
                                "Tuesday 24 February 2026  00:50:50 +0000 (0:00:00.276)       0:03:35.763 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-mon : Copy the initial key in /etc/ceph (for containers)] ***********",
                                "Tuesday 24 February 2026  00:50:51 +0000 (0:00:01.342)       0:03:37.106 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-mon : Create monitor directory] *************************************",
                                "Tuesday 24 February 2026  00:50:52 +0000 (0:00:00.750)       0:03:37.856 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-mon : Recursively fix ownership of monitor directory] ***************",
                                "Tuesday 24 February 2026  00:50:53 +0000 (0:00:00.692)       0:03:38.548 ****** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-mon : Create admin keyring] *****************************************",
                                "Tuesday 24 February 2026  00:50:54 +0000 (0:00:01.253)       0:03:39.803 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-mon : Slurp admin keyring] ******************************************",
                                "Tuesday 24 February 2026  00:50:56 +0000 (0:00:01.678)       0:03:41.481 ****** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-mon : Copy admin keyring over to mons] ******************************",
                                "Tuesday 24 February 2026  00:50:56 +0000 (0:00:00.637)       0:03:42.119 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=None)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=None)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0 -> testbed-node-1(192.168.16.11)] => (item=None)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2 -> testbed-node-1(192.168.16.11)] => (item=None)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0 -> testbed-node-2(192.168.16.12)] => (item=None)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1 -> testbed-node-2(192.168.16.12)] => (item=None)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0 -> {{ item }}]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1 -> {{ item }}]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item=None)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2 -> {{ item }}]\u001b[0m",
                                "",
                                "TASK [ceph-mon : Import admin keyring into mon keyring] ************************",
                                "Tuesday 24 February 2026  00:51:00 +0000 (0:00:03.529)       0:03:45.649 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-mon : Set_fact ceph-mon container command] **************************",
                                "Tuesday 24 February 2026  00:51:01 +0000 (0:00:01.219)       0:03:46.869 ****** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-mon : Set_fact monmaptool container command] ************************",
                                "Tuesday 24 February 2026  00:51:01 +0000 (0:00:00.316)       0:03:47.185 ****** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-mon : Generate initial monmap] **************************************",
                                "Tuesday 24 February 2026  00:51:02 +0000 (0:00:00.453)       0:03:47.639 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-mon : Ceph monitor mkfs with keyring] *******************************",
                                "Tuesday 24 February 2026  00:51:03 +0000 (0:00:01.484)       0:03:49.123 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-mon : Ceph monitor mkfs without keyring] ****************************",
                                "Tuesday 24 February 2026  00:51:05 +0000 (0:00:01.367)       0:03:50.491 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-mon : Include start_monitor.yml] ************************************",
                                "Tuesday 24 February 2026  00:51:05 +0000 (0:00:00.365)       0:03:50.857 ****** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-mon/tasks/start_monitor.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [ceph-mon : Ensure systemd service override directory exists] *************",
                                "Tuesday 24 February 2026  00:51:06 +0000 (0:00:00.613)       0:03:51.471 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-mon : Add ceph-mon systemd service overrides] ***********************",
                                "Tuesday 24 February 2026  00:51:06 +0000 (0:00:00.264)       0:03:51.735 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-mon : Include_tasks systemd.yml] ************************************",
                                "Tuesday 24 February 2026  00:51:06 +0000 (0:00:00.276)       0:03:52.012 ****** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-mon/tasks/systemd.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [ceph-mon : Generate systemd unit file for mon container] *****************",
                                "Tuesday 24 February 2026  00:51:07 +0000 (0:00:00.822)       0:03:52.834 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-mon : Generate systemd ceph-mon target file] ************************",
                                "Tuesday 24 February 2026  00:51:09 +0000 (0:00:02.202)       0:03:55.036 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-mon : Enable ceph-mon.target] ***************************************",
                                "Tuesday 24 February 2026  00:51:11 +0000 (0:00:01.387)       0:03:56.424 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-mon : Start the monitor service] ************************************",
                                "Tuesday 24 February 2026  00:51:13 +0000 (0:00:02.089)       0:03:58.514 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-mon : Include_tasks ceph_keys.yml] **********************************",
                                "Tuesday 24 February 2026  00:51:15 +0000 (0:00:02.509)       0:04:01.023 ****** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-mon/tasks/ceph_keys.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [ceph-mon : Waiting for the monitor(s) to form the quorum...] *************",
                                "Tuesday 24 February 2026  00:51:16 +0000 (0:00:00.483)       0:04:01.507 ****** ",
                                "\u001b[1;30mFAILED - RETRYING: [testbed-node-0]: Waiting for the monitor(s) to form the quorum... (10 retries left).\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [ceph-mon : Fetch ceph initial keys] **************************************",
                                "Tuesday 24 February 2026  00:51:38 +0000 (0:00:21.909)       0:04:23.416 ****** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-mon : Include secure_cluster.yml] ***********************************",
                                "Tuesday 24 February 2026  00:51:47 +0000 (0:00:09.272)       0:04:32.688 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-mon : Set cluster configs] ******************************************",
                                "Tuesday 24 February 2026  00:51:47 +0000 (0:00:00.463)       0:04:33.151 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=[{'key': 'global', 'value': {'public_network': '192.168.16.0/20', 'cluster_network': '192.168.16.0/20', 'osd_pool_default_crush_rule': -1, 'ms_bind_ipv6': 'False', 'ms_bind_ipv4': 'True', 'osd_crush_chooseleaf_type': '__omit_place_holder__a1d5f0c2cf158f4a69bae0c92e95cd3badd9c9e3'}}, {'key': 'public_network', 'value': '192.168.16.0/20'}])\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=[{'key': 'global', 'value': {'public_network': '192.168.16.0/20', 'cluster_network': '192.168.16.0/20', 'osd_pool_default_crush_rule': -1, 'ms_bind_ipv6': 'False', 'ms_bind_ipv4': 'True', 'osd_crush_chooseleaf_type': '__omit_place_holder__a1d5f0c2cf158f4a69bae0c92e95cd3badd9c9e3'}}, {'key': 'cluster_network', 'value': '192.168.16.0/20'}])\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=[{'key': 'global', 'value': {'public_network': '192.168.16.0/20', 'cluster_network': '192.168.16.0/20', 'osd_pool_default_crush_rule': -1, 'ms_bind_ipv6': 'False', 'ms_bind_ipv4': 'True', 'osd_crush_chooseleaf_type': '__omit_place_holder__a1d5f0c2cf158f4a69bae0c92e95cd3badd9c9e3'}}, {'key': 'osd_pool_default_crush_rule', 'value': -1}])\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=[{'key': 'global', 'value': {'public_network': '192.168.16.0/20', 'cluster_network': '192.168.16.0/20', 'osd_pool_default_crush_rule': -1, 'ms_bind_ipv6': 'False', 'ms_bind_ipv4': 'True', 'osd_crush_chooseleaf_type': '__omit_place_holder__a1d5f0c2cf158f4a69bae0c92e95cd3badd9c9e3'}}, {'key': 'ms_bind_ipv6', 'value': 'False'}])\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=[{'key': 'global', 'value': {'public_network': '192.168.16.0/20', 'cluster_network': '192.168.16.0/20', 'osd_pool_default_crush_rule': -1, 'ms_bind_ipv6': 'False', 'ms_bind_ipv4': 'True', 'osd_crush_chooseleaf_type': '__omit_place_holder__a1d5f0c2cf158f4a69bae0c92e95cd3badd9c9e3'}}, {'key': 'ms_bind_ipv4', 'value': 'True'}])\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=[{'key': 'global', 'value': {'public_network': '192.168.16.0/20', 'cluster_network': '192.168.16.0/20', 'osd_pool_default_crush_rule': -1, 'ms_bind_ipv6': 'False', 'ms_bind_ipv4': 'True', 'osd_crush_chooseleaf_type': '__omit_place_holder__a1d5f0c2cf158f4a69bae0c92e95cd3badd9c9e3'}}, {'key': 'osd_crush_chooseleaf_type', 'value': '__omit_place_holder__a1d5f0c2cf158f4a69bae0c92e95cd3badd9c9e3'}]) \u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Make tempdir for scripts] **********************",
                                "Tuesday 24 February 2026  00:52:02 +0000 (0:00:14.800)       0:04:47.952 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Mons handler] **********************************",
                                "Tuesday 24 February 2026  00:52:02 +0000 (0:00:00.395)       0:04:48.348 ****** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-handler/tasks/handler_mons.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Set _mon_handler_called before restart] ********",
                                "Tuesday 24 February 2026  00:52:03 +0000 (0:00:00.868)       0:04:49.217 ****** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Copy mon restart script] ***********************",
                                "Tuesday 24 February 2026  00:52:04 +0000 (0:00:00.472)       0:04:49.689 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Restart ceph mon daemon(s)] ********************",
                                "Tuesday 24 February 2026  00:52:04 +0000 (0:00:00.489)       0:04:50.179 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-0) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-1) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Set _mon_handler_called after restart] *********",
                                "Tuesday 24 February 2026  00:52:06 +0000 (0:00:01.477)       0:04:51.656 ****** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "PLAY [Apply role ceph-mgr] *****************************************************",
                                "",
                                "TASK [ceph-handler : Include check_running_cluster.yml] ************************",
                                "Tuesday 24 February 2026  00:52:07 +0000 (0:00:00.780)       0:04:52.436 ****** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-handler/tasks/check_running_cluster.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [ceph-handler : Include check_running_containers.yml] *********************",
                                "Tuesday 24 February 2026  00:52:07 +0000 (0:00:00.577)       0:04:53.013 ****** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-handler/tasks/check_running_containers.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a mon container] ********************************",
                                "Tuesday 24 February 2026  00:52:08 +0000 (0:00:00.579)       0:04:53.593 ****** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for an osd container] *******************************",
                                "Tuesday 24 February 2026  00:52:09 +0000 (0:00:00.773)       0:04:54.366 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a mds container] ********************************",
                                "Tuesday 24 February 2026  00:52:09 +0000 (0:00:00.387)       0:04:54.753 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a rgw container] ********************************",
                                "Tuesday 24 February 2026  00:52:09 +0000 (0:00:00.376)       0:04:55.130 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a mgr container] ********************************",
                                "Tuesday 24 February 2026  00:52:10 +0000 (0:00:00.303)       0:04:55.433 ****** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a rbd mirror container] *************************",
                                "Tuesday 24 February 2026  00:52:10 +0000 (0:00:00.843)       0:04:56.277 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a nfs container] ********************************",
                                "Tuesday 24 February 2026  00:52:11 +0000 (0:00:00.313)       0:04:56.590 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a ceph-crash container] *************************",
                                "Tuesday 24 February 2026  00:52:11 +0000 (0:00:00.679)       0:04:57.269 ****** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a ceph-exporter container] **********************",
                                "Tuesday 24 February 2026  00:52:12 +0000 (0:00:01.017)       0:04:58.287 ****** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Include check_socket_non_container.yml] *******************",
                                "Tuesday 24 February 2026  00:52:13 +0000 (0:00:00.785)       0:04:59.073 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_mon_status] ******************************",
                                "Tuesday 24 February 2026  00:52:13 +0000 (0:00:00.265)       0:04:59.338 ****** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_osd_status] ******************************",
                                "Tuesday 24 February 2026  00:52:14 +0000 (0:00:00.277)       0:04:59.616 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_mds_status] ******************************",
                                "Tuesday 24 February 2026  00:52:14 +0000 (0:00:00.426)       0:05:00.042 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_rgw_status] ******************************",
                                "Tuesday 24 February 2026  00:52:14 +0000 (0:00:00.256)       0:05:00.299 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_nfs_status] ******************************",
                                "Tuesday 24 February 2026  00:52:15 +0000 (0:00:00.267)       0:05:00.567 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_rbd_status] ******************************",
                                "Tuesday 24 February 2026  00:52:15 +0000 (0:00:00.300)       0:05:00.868 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_mgr_status] ******************************",
                                "Tuesday 24 February 2026  00:52:15 +0000 (0:00:00.455)       0:05:01.323 ****** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_crash_status] ****************************",
                                "Tuesday 24 February 2026  00:52:16 +0000 (0:00:00.318)       0:05:01.641 ****** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_exporter_status] *************************",
                                "Tuesday 24 February 2026  00:52:16 +0000 (0:00:00.275)       0:05:01.916 ****** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-mgr : Set_fact container_exec_cmd] **********************************",
                                "Tuesday 24 February 2026  00:52:17 +0000 (0:00:00.720)       0:05:02.637 ****** ",
                                "\u001b[0;32mok: [testbed-node-0] => (item=testbed-node-0)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0 -> testbed-node-1(192.168.16.11)] => (item=testbed-node-1)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0 -> testbed-node-2(192.168.16.12)] => (item=testbed-node-2)\u001b[0m",
                                "",
                                "TASK [ceph-mgr : Include common.yml] *******************************************",
                                "Tuesday 24 February 2026  00:52:17 +0000 (0:00:00.474)       0:05:03.112 ****** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-mgr/tasks/common.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [ceph-mgr : Create mgr directory] *****************************************",
                                "Tuesday 24 February 2026  00:52:18 +0000 (0:00:00.515)       0:05:03.627 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-mgr : Fetch ceph mgr keyring] ***************************************",
                                "Tuesday 24 February 2026  00:52:18 +0000 (0:00:00.605)       0:05:04.233 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-mgr : Create ceph mgr keyring(s) on a mon node] *********************",
                                "Tuesday 24 February 2026  00:52:19 +0000 (0:00:00.430)       0:05:04.663 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=None)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=None)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=None)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0 -> {{ groups[mon_group_name][0] }}]\u001b[0m",
                                "",
                                "TASK [ceph-mgr : Set_fact _mgr_keys] *******************************************",
                                "Tuesday 24 February 2026  00:52:29 +0000 (0:00:10.213)       0:05:14.877 ****** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-mgr : Get keys from monitors] ***************************************",
                                "Tuesday 24 February 2026  00:52:29 +0000 (0:00:00.371)       0:05:15.248 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=None) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=None) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=None) \u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0] => (item=None)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m",
                                "",
                                "TASK [ceph-mgr : Copy ceph key(s) if needed] ***********************************",
                                "Tuesday 24 February 2026  00:52:32 +0000 (0:00:02.351)       0:05:17.600 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=None) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=None) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=None) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=None)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=None)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=None)\u001b[0m",
                                "",
                                "TASK [ceph-mgr : Set mgr key permissions] **************************************",
                                "Tuesday 24 February 2026  00:52:33 +0000 (0:00:01.488)       0:05:19.089 ****** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-mgr : Append dashboard modules to ceph_mgr_modules] *****************",
                                "Tuesday 24 February 2026  00:52:35 +0000 (0:00:01.523)       0:05:20.613 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-mgr : Include pre_requisite.yml] ************************************",
                                "Tuesday 24 February 2026  00:52:35 +0000 (0:00:00.332)       0:05:20.945 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-mgr : Include start_mgr.yml] ****************************************",
                                "Tuesday 24 February 2026  00:52:35 +0000 (0:00:00.298)       0:05:21.244 ****** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-mgr/tasks/start_mgr.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [ceph-mgr : Ensure systemd service override directory exists] *************",
                                "Tuesday 24 February 2026  00:52:36 +0000 (0:00:00.723)       0:05:21.968 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-mgr : Add ceph-mgr systemd service overrides] ***********************",
                                "Tuesday 24 February 2026  00:52:36 +0000 (0:00:00.324)       0:05:22.292 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-mgr : Include_tasks systemd.yml] ************************************",
                                "Tuesday 24 February 2026  00:52:37 +0000 (0:00:00.329)       0:05:22.621 ****** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-mgr/tasks/systemd.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [ceph-mgr : Generate systemd unit file] ***********************************",
                                "Tuesday 24 February 2026  00:52:38 +0000 (0:00:00.805)       0:05:23.427 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [ceph-mgr : Generate systemd ceph-mgr target file] ************************",
                                "Tuesday 24 February 2026  00:52:39 +0000 (0:00:01.273)       0:05:24.701 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-mgr : Enable ceph-mgr.target] ***************************************",
                                "Tuesday 24 February 2026  00:52:40 +0000 (0:00:01.225)       0:05:25.927 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-mgr : Systemd start mgr] ********************************************",
                                "Tuesday 24 February 2026  00:52:42 +0000 (0:00:02.019)       0:05:27.946 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-mgr : Include mgr_modules.yml] **************************************",
                                "Tuesday 24 February 2026  00:52:44 +0000 (0:00:02.114)       0:05:30.060 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-mgr/tasks/mgr_modules.yml for testbed-node-2\u001b[0m",
                                "",
                                "TASK [ceph-mgr : Wait for all mgr to be up] ************************************",
                                "Tuesday 24 February 2026  00:52:45 +0000 (0:00:00.542)       0:05:30.603 ****** ",
                                "\u001b[1;30mFAILED - RETRYING: [testbed-node-2 -> testbed-node-0]: Wait for all mgr to be up (30 retries left).\u001b[0m",
                                "\u001b[1;30mFAILED - RETRYING: [testbed-node-2 -> testbed-node-0]: Wait for all mgr to be up (29 retries left).\u001b[0m",
                                "\u001b[1;30mFAILED - RETRYING: [testbed-node-2 -> testbed-node-0]: Wait for all mgr to be up (28 retries left).\u001b[0m",
                                "\u001b[1;30mFAILED - RETRYING: [testbed-node-2 -> testbed-node-0]: Wait for all mgr to be up (27 retries left).\u001b[0m",
                                "\u001b[1;30mFAILED - RETRYING: [testbed-node-2 -> testbed-node-0]: Wait for all mgr to be up (26 retries left).\u001b[0m",
                                "\u001b[1;30mFAILED - RETRYING: [testbed-node-2 -> testbed-node-0]: Wait for all mgr to be up (25 retries left).\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2 -> testbed-node-0(192.168.16.10)]\u001b[0m",
                                "",
                                "TASK [ceph-mgr : Get enabled modules from ceph-mgr] ****************************",
                                "Tuesday 24 February 2026  00:53:21 +0000 (0:00:36.565)       0:06:07.169 ****** ",
                                "\u001b[0;32mok: [testbed-node-2 -> testbed-node-0(192.168.16.10)]\u001b[0m",
                                "",
                                "TASK [ceph-mgr : Set _ceph_mgr_modules fact (convert _ceph_mgr_modules.stdout to a dict)] ***",
                                "Tuesday 24 February 2026  00:53:23 +0000 (0:00:01.367)       0:06:08.536 ****** ",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-mgr : Set _disabled_ceph_mgr_modules fact] **************************",
                                "Tuesday 24 February 2026  00:53:23 +0000 (0:00:00.277)       0:06:08.813 ****** ",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-mgr : Disable ceph mgr enabled modules] *****************************",
                                "Tuesday 24 February 2026  00:53:23 +0000 (0:00:00.140)       0:06:08.954 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-2 -> testbed-node-0(192.168.16.10)] => (item=iostat)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2 -> testbed-node-0(192.168.16.10)] => (item=nfs)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2 -> testbed-node-0(192.168.16.10)] => (item=restful)\u001b[0m",
                                "",
                                "TASK [ceph-mgr : Add modules to ceph-mgr] **************************************",
                                "Tuesday 24 February 2026  00:53:30 +0000 (0:00:06.715)       0:06:15.670 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=balancer) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2 -> testbed-node-0(192.168.16.10)] => (item=dashboard)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2 -> testbed-node-0(192.168.16.10)] => (item=prometheus)\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=status) \u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Make tempdir for scripts] **********************",
                                "Tuesday 24 February 2026  00:53:36 +0000 (0:00:05.722)       0:06:21.392 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Mgrs handler] **********************************",
                                "Tuesday 24 February 2026  00:53:36 +0000 (0:00:00.724)       0:06:22.116 ****** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-handler/tasks/handler_mgrs.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Set _mgr_handler_called before restart] ********",
                                "Tuesday 24 February 2026  00:53:37 +0000 (0:00:00.907)       0:06:23.024 ****** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Copy mgr restart script] ***********************",
                                "Tuesday 24 February 2026  00:53:38 +0000 (0:00:00.373)       0:06:23.398 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Restart ceph mgr daemon(s)] ********************",
                                "Tuesday 24 February 2026  00:53:39 +0000 (0:00:01.324)       0:06:24.723 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-0) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-1) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=testbed-node-2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Set _mgr_handler_called after restart] *********",
                                "Tuesday 24 February 2026  00:53:39 +0000 (0:00:00.618)       0:06:25.341 ****** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "PLAY [Apply role ceph-osd] *****************************************************",
                                "",
                                "TASK [ceph-handler : Include check_running_cluster.yml] ************************",
                                "Tuesday 24 February 2026  00:53:40 +0000 (0:00:00.952)       0:06:26.293 ****** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-handler/tasks/check_running_cluster.yml for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m",
                                "",
                                "TASK [ceph-handler : Include check_running_containers.yml] *********************",
                                "Tuesday 24 February 2026  00:53:41 +0000 (0:00:00.506)       0:06:26.800 ****** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-handler/tasks/check_running_containers.yml for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a mon container] ********************************",
                                "Tuesday 24 February 2026  00:53:42 +0000 (0:00:00.764)       0:06:27.565 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for an osd container] *******************************",
                                "Tuesday 24 February 2026  00:53:42 +0000 (0:00:00.314)       0:06:27.879 ****** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a mds container] ********************************",
                                "Tuesday 24 February 2026  00:53:43 +0000 (0:00:00.810)       0:06:28.690 ****** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a rgw container] ********************************",
                                "Tuesday 24 February 2026  00:53:44 +0000 (0:00:00.843)       0:06:29.533 ****** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a mgr container] ********************************",
                                "Tuesday 24 February 2026  00:53:45 +0000 (0:00:01.264)       0:06:30.797 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a rbd mirror container] *************************",
                                "Tuesday 24 February 2026  00:53:45 +0000 (0:00:00.368)       0:06:31.166 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a nfs container] ********************************",
                                "Tuesday 24 February 2026  00:53:46 +0000 (0:00:00.307)       0:06:31.473 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a ceph-crash container] *************************",
                                "Tuesday 24 February 2026  00:53:46 +0000 (0:00:00.323)       0:06:31.797 ****** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a ceph-exporter container] **********************",
                                "Tuesday 24 February 2026  00:53:47 +0000 (0:00:01.020)       0:06:32.818 ****** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Include check_socket_non_container.yml] *******************",
                                "Tuesday 24 February 2026  00:53:48 +0000 (0:00:00.694)       0:06:33.512 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_mon_status] ******************************",
                                "Tuesday 24 February 2026  00:53:48 +0000 (0:00:00.339)       0:06:33.852 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_osd_status] ******************************",
                                "Tuesday 24 February 2026  00:53:48 +0000 (0:00:00.338)       0:06:34.190 ****** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_mds_status] ******************************",
                                "Tuesday 24 February 2026  00:53:49 +0000 (0:00:00.629)       0:06:34.819 ****** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_rgw_status] ******************************",
                                "Tuesday 24 February 2026  00:53:49 +0000 (0:00:00.366)       0:06:35.186 ****** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_nfs_status] ******************************",
                                "Tuesday 24 February 2026  00:53:50 +0000 (0:00:00.346)       0:06:35.533 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_rbd_status] ******************************",
                                "Tuesday 24 February 2026  00:53:50 +0000 (0:00:00.309)       0:06:35.842 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_mgr_status] ******************************",
                                "Tuesday 24 February 2026  00:53:51 +0000 (0:00:00.702)       0:06:36.544 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_crash_status] ****************************",
                                "Tuesday 24 February 2026  00:53:51 +0000 (0:00:00.384)       0:06:36.928 ****** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_exporter_status] *************************",
                                "Tuesday 24 February 2026  00:53:51 +0000 (0:00:00.370)       0:06:37.299 ****** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-osd : Set_fact add_osd] *********************************************",
                                "Tuesday 24 February 2026  00:53:52 +0000 (0:00:00.554)       0:06:37.853 ****** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-osd : Set_fact container_exec_cmd] **********************************",
                                "Tuesday 24 February 2026  00:53:53 +0000 (0:00:00.625)       0:06:38.478 ****** ",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item=testbed-node-0)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-1(192.168.16.11)] => (item=testbed-node-1)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-2(192.168.16.12)] => (item=testbed-node-2)\u001b[0m",
                                "",
                                "TASK [ceph-osd : Include_tasks system_tuning.yml] ******************************",
                                "Tuesday 24 February 2026  00:53:53 +0000 (0:00:00.643)       0:06:39.122 ****** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-osd/tasks/system_tuning.yml for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m",
                                "",
                                "TASK [ceph-osd : Create tmpfiles.d directory] **********************************",
                                "Tuesday 24 February 2026  00:53:54 +0000 (0:00:00.654)       0:06:39.777 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-osd : Disable transparent hugepage] *********************************",
                                "Tuesday 24 February 2026  00:53:54 +0000 (0:00:00.584)       0:06:40.361 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-osd : Get default vm.min_free_kbytes] *******************************",
                                "Tuesday 24 February 2026  00:53:55 +0000 (0:00:00.355)       0:06:40.716 ****** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-osd : Set_fact vm_min_free_kbytes] **********************************",
                                "Tuesday 24 February 2026  00:53:55 +0000 (0:00:00.617)       0:06:41.334 ****** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-osd : Apply operating system tuning] ********************************",
                                "Tuesday 24 February 2026  00:53:56 +0000 (0:00:00.394)       0:06:41.728 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'name': 'fs.aio-max-nr', 'value': '1048576', 'enable': True})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'name': 'fs.aio-max-nr', 'value': '1048576', 'enable': True})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'name': 'fs.file-max', 'value': 26234859})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'name': 'fs.file-max', 'value': 26234859})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'name': 'vm.zone_reclaim_mode', 'value': 0})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'name': 'fs.aio-max-nr', 'value': '1048576', 'enable': True})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'name': 'vm.zone_reclaim_mode', 'value': 0})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'name': 'vm.swappiness', 'value': 10})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'name': 'vm.swappiness', 'value': 10})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'name': 'vm.min_free_kbytes', 'value': '67584'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'name': 'fs.file-max', 'value': 26234859})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'name': 'vm.zone_reclaim_mode', 'value': 0})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'name': 'vm.swappiness', 'value': 10})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'name': 'vm.min_free_kbytes', 'value': '67584'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'name': 'vm.min_free_kbytes', 'value': '67584'})\u001b[0m",
                                "",
                                "TASK [ceph-osd : Install dependencies] *****************************************",
                                "Tuesday 24 February 2026  00:54:00 +0000 (0:00:04.284)       0:06:46.013 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-osd : Include_tasks common.yml] *************************************",
                                "Tuesday 24 February 2026  00:54:00 +0000 (0:00:00.308)       0:06:46.321 ****** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-osd/tasks/common.yml for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m",
                                "",
                                "TASK [ceph-osd : Create bootstrap-osd and osd directories] *********************",
                                "Tuesday 24 February 2026  00:54:01 +0000 (0:00:00.544)       0:06:46.866 ****** ",
                                "\u001b[0;32mok: [testbed-node-3] => (item=/var/lib/ceph/bootstrap-osd/)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item=/var/lib/ceph/bootstrap-osd/)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item=/var/lib/ceph/bootstrap-osd/)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item=/var/lib/ceph/osd/)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item=/var/lib/ceph/osd/)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item=/var/lib/ceph/osd/)\u001b[0m",
                                "",
                                "TASK [ceph-osd : Get keys from monitors] ***************************************",
                                "Tuesday 24 February 2026  00:54:02 +0000 (0:00:01.375)       0:06:48.241 ****** ",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=None) \u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3 -> {{ groups.get(mon_group_name)[0] }}]\u001b[0m",
                                "",
                                "TASK [ceph-osd : Copy ceph key(s) if needed] ***********************************",
                                "Tuesday 24 February 2026  00:54:05 +0000 (0:00:02.637)       0:06:50.878 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=None)\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=None) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=None)\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=None) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=None)\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=None) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-osd : Set noup flag] ************************************************",
                                "Tuesday 24 February 2026  00:54:06 +0000 (0:00:01.342)       0:06:52.221 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-3 -> testbed-node-0(192.168.16.10)]\u001b[0m",
                                "",
                                "TASK [ceph-osd : Include_tasks scenarios/lvm.yml] ******************************",
                                "Tuesday 24 February 2026  00:54:09 +0000 (0:00:02.618)       0:06:54.839 ****** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-osd/tasks/scenarios/lvm.yml for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m",
                                "",
                                "TASK [ceph-osd : Use ceph-volume to create osds] *******************************",
                                "Tuesday 24 February 2026  00:54:10 +0000 (0:00:00.555)       0:06:55.395 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'data': 'osd-block-752a9ec4-4bfc-5551-82cd-92b802bdad97', 'data_vg': 'ceph-752a9ec4-4bfc-5551-82cd-92b802bdad97'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'data': 'osd-block-9b92f24c-dac1-5bff-a390-34d9ff9e0f4f', 'data_vg': 'ceph-9b92f24c-dac1-5bff-a390-34d9ff9e0f4f'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'data': 'osd-block-294dd48f-9d6d-5079-8214-59e60af9e526', 'data_vg': 'ceph-294dd48f-9d6d-5079-8214-59e60af9e526'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'data': 'osd-block-ba85a0e3-2f57-51d5-a9f1-072f0e9a8deb', 'data_vg': 'ceph-ba85a0e3-2f57-51d5-a9f1-072f0e9a8deb'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'data': 'osd-block-7f6a8fe5-6adc-52b4-b256-edcdbfd9bf4e', 'data_vg': 'ceph-7f6a8fe5-6adc-52b4-b256-edcdbfd9bf4e'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'data': 'osd-block-7fbf4d63-7c6f-5c70-887c-2b73c978775b', 'data_vg': 'ceph-7fbf4d63-7c6f-5c70-887c-2b73c978775b'})\u001b[0m",
                                "",
                                "TASK [ceph-osd : Include_tasks scenarios/lvm-batch.yml] ************************",
                                "Tuesday 24 February 2026  00:54:58 +0000 (0:00:48.430)       0:07:43.825 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-osd : Include_tasks start_osds.yml] *********************************",
                                "Tuesday 24 February 2026  00:54:58 +0000 (0:00:00.408)       0:07:44.234 ****** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-osd/tasks/start_osds.yml for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m",
                                "",
                                "TASK [ceph-osd : Get osd ids] **************************************************",
                                "Tuesday 24 February 2026  00:54:59 +0000 (0:00:00.681)       0:07:44.915 ****** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-osd : Collect osd ids] **********************************************",
                                "Tuesday 24 February 2026  00:55:00 +0000 (0:00:01.064)       0:07:45.979 ****** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [ceph-osd : Include_tasks systemd.yml] ************************************",
                                "Tuesday 24 February 2026  00:55:03 +0000 (0:00:02.510)       0:07:48.490 ****** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-osd/tasks/systemd.yml for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m",
                                "",
                                "TASK [ceph-osd : Generate systemd unit file] ***********************************",
                                "Tuesday 24 February 2026  00:55:03 +0000 (0:00:00.502)       0:07:48.992 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-osd : Generate systemd ceph-osd target file] ************************",
                                "Tuesday 24 February 2026  00:55:05 +0000 (0:00:01.478)       0:07:50.471 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-osd : Enable ceph-osd.target] ***************************************",
                                "Tuesday 24 February 2026  00:55:06 +0000 (0:00:01.145)       0:07:51.617 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [ceph-osd : Ensure systemd service override directory exists] *************",
                                "Tuesday 24 February 2026  00:55:08 +0000 (0:00:01.909)       0:07:53.526 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-osd : Add ceph-osd systemd service overrides] ***********************",
                                "Tuesday 24 February 2026  00:55:08 +0000 (0:00:00.359)       0:07:53.886 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-osd : Ensure /var/lib/ceph/osd/<cluster>-<item> is present] *********",
                                "Tuesday 24 February 2026  00:55:09 +0000 (0:00:00.618)       0:07:54.505 ****** ",
                                "\u001b[0;32mok: [testbed-node-3] => (item=0)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item=5)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item=3)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item=4)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item=1)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item=2)\u001b[0m",
                                "",
                                "TASK [ceph-osd : Write run file in /var/lib/ceph/osd/xxxx/run] *****************",
                                "Tuesday 24 February 2026  00:55:10 +0000 (0:00:01.020)       0:07:55.525 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=0)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=3)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=5)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=4)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=1)\u001b[0m",
                                "",
                                "TASK [ceph-osd : Systemd start osd] ********************************************",
                                "Tuesday 24 February 2026  00:55:12 +0000 (0:00:02.525)       0:07:58.051 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=3)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=0)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=5)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=4)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=1)\u001b[0m",
                                "",
                                "TASK [ceph-osd : Unset noup flag] **********************************************",
                                "Tuesday 24 February 2026  00:55:16 +0000 (0:00:03.882)       0:08:01.933 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-0(192.168.16.10)]\u001b[0m",
                                "",
                                "TASK [ceph-osd : Wait for all osd to be up] ************************************",
                                "Tuesday 24 February 2026  00:55:19 +0000 (0:00:03.041)       0:08:04.975 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[1;30mFAILED - RETRYING: [testbed-node-5 -> testbed-node-0]: Wait for all osd to be up (60 retries left).\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5 -> testbed-node-0(192.168.16.10)]\u001b[0m",
                                "",
                                "TASK [ceph-osd : Include crush_rules.yml] **************************************",
                                "Tuesday 24 February 2026  00:55:32 +0000 (0:00:12.824)       0:08:17.799 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Make tempdir for scripts] **********************",
                                "Tuesday 24 February 2026  00:55:33 +0000 (0:00:01.281)       0:08:19.081 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Osds handler] **********************************",
                                "Tuesday 24 February 2026  00:55:34 +0000 (0:00:00.400)       0:08:19.482 ****** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-handler/tasks/handler_osds.yml for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Set_fact trigger_restart] **********************",
                                "Tuesday 24 February 2026  00:55:34 +0000 (0:00:00.543)       0:08:20.026 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-3) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-4) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-5) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Set _osd_handler_called before restart] ********",
                                "Tuesday 24 February 2026  00:55:35 +0000 (0:00:01.062)       0:08:21.088 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Unset noup flag] *******************************",
                                "Tuesday 24 February 2026  00:55:36 +0000 (0:00:00.332)       0:08:21.421 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Copy osd restart script] ***********************",
                                "Tuesday 24 February 2026  00:55:36 +0000 (0:00:00.232)       0:08:21.654 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Get pool list] *********************************",
                                "Tuesday 24 February 2026  00:55:36 +0000 (0:00:00.388)       0:08:22.042 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Get balancer module status] ********************",
                                "Tuesday 24 February 2026  00:55:36 +0000 (0:00:00.231)       0:08:22.273 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Set_fact pools_pgautoscaler_mode] **************",
                                "Tuesday 24 February 2026  00:55:37 +0000 (0:00:00.215)       0:08:22.489 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Disable balancer] ******************************",
                                "Tuesday 24 February 2026  00:55:37 +0000 (0:00:00.122)       0:08:22.611 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Disable pg autoscale on pools] *****************",
                                "Tuesday 24 February 2026  00:55:37 +0000 (0:00:00.371)       0:08:22.983 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Restart ceph osds daemon(s)] *******************",
                                "Tuesday 24 February 2026  00:55:38 +0000 (0:00:01.051)       0:08:24.035 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-3) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-4) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-5) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Set _osd_handler_called after restart] *********",
                                "Tuesday 24 February 2026  00:55:39 +0000 (0:00:00.466)       0:08:24.501 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Re-enable pg autoscale on pools] ***************",
                                "Tuesday 24 February 2026  00:55:39 +0000 (0:00:00.355)       0:08:24.857 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Re-enable balancer] ****************************",
                                "Tuesday 24 February 2026  00:55:39 +0000 (0:00:00.261)       0:08:25.118 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "PLAY [Apply role ceph-crash] ***************************************************",
                                "",
                                "TASK [ceph-handler : Include check_running_cluster.yml] ************************",
                                "Tuesday 24 February 2026  00:55:40 +0000 (0:00:00.733)       0:08:25.852 ****** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-handler/tasks/check_running_cluster.yml for testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [ceph-handler : Include check_running_containers.yml] *********************",
                                "Tuesday 24 February 2026  00:55:41 +0000 (0:00:01.317)       0:08:27.169 ****** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-handler/tasks/check_running_containers.yml for testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a mon container] ********************************",
                                "Tuesday 24 February 2026  00:55:43 +0000 (0:00:01.401)       0:08:28.571 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for an osd container] *******************************",
                                "Tuesday 24 February 2026  00:55:44 +0000 (0:00:01.389)       0:08:29.961 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a mds container] ********************************",
                                "Tuesday 24 February 2026  00:55:45 +0000 (0:00:00.719)       0:08:30.681 ****** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a rgw container] ********************************",
                                "Tuesday 24 February 2026  00:55:46 +0000 (0:00:01.022)       0:08:31.703 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a mgr container] ********************************",
                                "Tuesday 24 February 2026  00:55:47 +0000 (0:00:00.710)       0:08:32.413 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a rbd mirror container] *************************",
                                "Tuesday 24 February 2026  00:55:48 +0000 (0:00:01.334)       0:08:33.748 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a nfs container] ********************************",
                                "Tuesday 24 February 2026  00:55:49 +0000 (0:00:00.717)       0:08:34.466 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a ceph-crash container] *************************",
                                "Tuesday 24 February 2026  00:55:50 +0000 (0:00:00.931)       0:08:35.398 ****** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a ceph-exporter container] **********************",
                                "Tuesday 24 February 2026  00:55:51 +0000 (0:00:01.101)       0:08:36.499 ****** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Include check_socket_non_container.yml] *******************",
                                "Tuesday 24 February 2026  00:55:52 +0000 (0:00:01.492)       0:08:37.991 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_mon_status] ******************************",
                                "Tuesday 24 February 2026  00:55:53 +0000 (0:00:00.663)       0:08:38.655 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_osd_status] ******************************",
                                "Tuesday 24 February 2026  00:55:54 +0000 (0:00:01.115)       0:08:39.771 ****** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_mds_status] ******************************",
                                "Tuesday 24 February 2026  00:55:55 +0000 (0:00:00.796)       0:08:40.567 ****** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_rgw_status] ******************************",
                                "Tuesday 24 February 2026  00:55:56 +0000 (0:00:01.059)       0:08:41.627 ****** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_nfs_status] ******************************",
                                "Tuesday 24 February 2026  00:55:57 +0000 (0:00:00.835)       0:08:42.462 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_rbd_status] ******************************",
                                "Tuesday 24 February 2026  00:55:58 +0000 (0:00:00.942)       0:08:43.405 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_mgr_status] ******************************",
                                "Tuesday 24 February 2026  00:55:58 +0000 (0:00:00.771)       0:08:44.177 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_crash_status] ****************************",
                                "Tuesday 24 February 2026  00:55:59 +0000 (0:00:01.109)       0:08:45.286 ****** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_exporter_status] *************************",
                                "Tuesday 24 February 2026  00:56:00 +0000 (0:00:00.723)       0:08:46.010 ****** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-crash : Create client.crash keyring] ********************************",
                                "Tuesday 24 February 2026  00:56:02 +0000 (0:00:01.441)       0:08:47.452 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-3 -> testbed-node-0(192.168.16.10)]\u001b[0m",
                                "",
                                "TASK [ceph-crash : Get keys from monitors] *************************************",
                                "Tuesday 24 February 2026  00:56:06 +0000 (0:00:04.655)       0:08:52.108 ****** ",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-0(192.168.16.10)]\u001b[0m",
                                "",
                                "TASK [ceph-crash : Copy ceph key(s) if needed] *********************************",
                                "Tuesday 24 February 2026  00:56:09 +0000 (0:00:02.380)       0:08:54.489 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-crash : Create /var/lib/ceph/crash/posted] **************************",
                                "Tuesday 24 February 2026  00:56:11 +0000 (0:00:02.195)       0:08:56.684 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-crash : Include_tasks systemd.yml] **********************************",
                                "Tuesday 24 February 2026  00:56:12 +0000 (0:00:01.034)       0:08:57.719 ****** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-crash/tasks/systemd.yml for testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [ceph-crash : Generate systemd unit file for ceph-crash container] ********",
                                "Tuesday 24 February 2026  00:56:13 +0000 (0:00:01.392)       0:08:59.112 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [ceph-crash : Start the ceph-crash service] *******************************",
                                "Tuesday 24 February 2026  00:56:15 +0000 (0:00:02.048)       0:09:01.160 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Ceph crash handler] ****************************",
                                "Tuesday 24 February 2026  00:56:19 +0000 (0:00:03.697)       0:09:04.857 ****** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-handler/tasks/handler_crash.yml for testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Set _crash_handler_called before restart] ******",
                                "Tuesday 24 February 2026  00:56:22 +0000 (0:00:02.520)       0:09:07.378 ****** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Restart the ceph-crash service] ****************",
                                "Tuesday 24 February 2026  00:56:23 +0000 (0:00:01.480)       0:09:08.859 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Set _crash_handler_called after restart] *******",
                                "Tuesday 24 February 2026  00:56:27 +0000 (0:00:03.599)       0:09:12.458 ****** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "PLAY [Apply role ceph-mds] *****************************************************",
                                "",
                                "TASK [ceph-handler : Include check_running_cluster.yml] ************************",
                                "Tuesday 24 February 2026  00:56:28 +0000 (0:00:01.310)       0:09:13.769 ****** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-handler/tasks/check_running_cluster.yml for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m",
                                "",
                                "TASK [ceph-handler : Include check_running_containers.yml] *********************",
                                "Tuesday 24 February 2026  00:56:29 +0000 (0:00:00.776)       0:09:14.545 ****** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-handler/tasks/check_running_containers.yml for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a mon container] ********************************",
                                "Tuesday 24 February 2026  00:56:29 +0000 (0:00:00.814)       0:09:15.360 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for an osd container] *******************************",
                                "Tuesday 24 February 2026  00:56:30 +0000 (0:00:00.436)       0:09:15.796 ****** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a mds container] ********************************",
                                "Tuesday 24 February 2026  00:56:31 +0000 (0:00:00.758)       0:09:16.554 ****** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a rgw container] ********************************",
                                "Tuesday 24 February 2026  00:56:32 +0000 (0:00:01.039)       0:09:17.594 ****** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a mgr container] ********************************",
                                "Tuesday 24 February 2026  00:56:32 +0000 (0:00:00.734)       0:09:18.329 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a rbd mirror container] *************************",
                                "Tuesday 24 February 2026  00:56:33 +0000 (0:00:00.345)       0:09:18.675 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a nfs container] ********************************",
                                "Tuesday 24 February 2026  00:56:33 +0000 (0:00:00.307)       0:09:18.982 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a ceph-crash container] *************************",
                                "Tuesday 24 February 2026  00:56:34 +0000 (0:00:00.656)       0:09:19.638 ****** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a ceph-exporter container] **********************",
                                "Tuesday 24 February 2026  00:56:35 +0000 (0:00:00.763)       0:09:20.402 ****** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Include check_socket_non_container.yml] *******************",
                                "Tuesday 24 February 2026  00:56:35 +0000 (0:00:00.936)       0:09:21.338 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_mon_status] ******************************",
                                "Tuesday 24 February 2026  00:56:36 +0000 (0:00:00.347)       0:09:21.685 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_osd_status] ******************************",
                                "Tuesday 24 February 2026  00:56:37 +0000 (0:00:00.850)       0:09:22.536 ****** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_mds_status] ******************************",
                                "Tuesday 24 February 2026  00:56:37 +0000 (0:00:00.412)       0:09:22.949 ****** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_rgw_status] ******************************",
                                "Tuesday 24 February 2026  00:56:38 +0000 (0:00:00.480)       0:09:23.429 ****** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_nfs_status] ******************************",
                                "Tuesday 24 February 2026  00:56:38 +0000 (0:00:00.413)       0:09:23.843 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_rbd_status] ******************************",
                                "Tuesday 24 February 2026  00:56:39 +0000 (0:00:00.698)       0:09:24.541 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_mgr_status] ******************************",
                                "Tuesday 24 February 2026  00:56:39 +0000 (0:00:00.336)       0:09:24.877 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_crash_status] ****************************",
                                "Tuesday 24 February 2026  00:56:39 +0000 (0:00:00.370)       0:09:25.247 ****** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_exporter_status] *************************",
                                "Tuesday 24 February 2026  00:56:40 +0000 (0:00:00.482)       0:09:25.731 ****** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-mds : Include create_mds_filesystems.yml] ***************************",
                                "Tuesday 24 February 2026  00:56:41 +0000 (0:00:00.951)       0:09:26.682 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-mds/tasks/create_mds_filesystems.yml for testbed-node-3\u001b[0m",
                                "",
                                "TASK [ceph-facts : Get current default crush rule details] *********************",
                                "Tuesday 24 February 2026  00:56:41 +0000 (0:00:00.412)       0:09:27.095 ****** ",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-0(192.168.16.10)]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Get current default crush rule name] ************************",
                                "Tuesday 24 February 2026  00:56:44 +0000 (0:00:02.379)       0:09:29.475 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'rule_id': 0, 'rule_name': 'replicated_rule', 'type': 1, 'steps': [{'op': 'take', 'item': -1, 'item_name': 'default'}, {'op': 'chooseleaf_firstn', 'num': 0, 'type': 'host'}, {'op': 'emit'}]}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-mds : Create filesystem pools] **************************************",
                                "Tuesday 24 February 2026  00:56:44 +0000 (0:00:00.246)       0:09:29.721 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item={'application': 'cephfs', 'erasure_profile': '', 'expected_num_objects': '', 'min_size': 0, 'name': 'cephfs_data', 'pg_num': 16, 'pgp_num': 16, 'rule_name': 'replicated_rule', 'size': 3, 'type': 1})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item={'application': 'cephfs', 'erasure_profile': '', 'expected_num_objects': '', 'min_size': 0, 'name': 'cephfs_metadata', 'pg_num': 16, 'pgp_num': 16, 'rule_name': 'replicated_rule', 'size': 3, 'type': 1})\u001b[0m",
                                "",
                                "TASK [ceph-mds : Create ceph filesystem] ***************************************",
                                "Tuesday 24 February 2026  00:56:54 +0000 (0:00:10.517)       0:09:40.239 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-3 -> testbed-node-0(192.168.16.10)]\u001b[0m",
                                "",
                                "TASK [ceph-mds : Include common.yml] *******************************************",
                                "Tuesday 24 February 2026  00:56:58 +0000 (0:00:03.807)       0:09:44.046 ****** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-mds/tasks/common.yml for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m",
                                "",
                                "TASK [ceph-mds : Create bootstrap-mds and mds directories] *********************",
                                "Tuesday 24 February 2026  00:56:59 +0000 (0:00:00.698)       0:09:44.745 ****** ",
                                "\u001b[0;32mok: [testbed-node-3] => (item=/var/lib/ceph/bootstrap-mds/)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item=/var/lib/ceph/bootstrap-mds/)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item=/var/lib/ceph/bootstrap-mds/)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=/var/lib/ceph/mds/ceph-testbed-node-4)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=/var/lib/ceph/mds/ceph-testbed-node-3)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=/var/lib/ceph/mds/ceph-testbed-node-5)\u001b[0m",
                                "",
                                "TASK [ceph-mds : Get keys from monitors] ***************************************",
                                "Tuesday 24 February 2026  00:57:00 +0000 (0:00:01.112)       0:09:45.858 ****** ",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=None) \u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3 -> {{ groups.get(mon_group_name)[0] }}]\u001b[0m",
                                "",
                                "TASK [ceph-mds : Copy ceph key(s) if needed] ***********************************",
                                "Tuesday 24 February 2026  00:57:03 +0000 (0:00:02.730)       0:09:48.589 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=None)\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=None) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=None)\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=None) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=None)\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=None) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-mds : Create mds keyring] *******************************************",
                                "Tuesday 24 February 2026  00:57:04 +0000 (0:00:01.619)       0:09:50.209 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [ceph-mds : Non_containerized.yml] ****************************************",
                                "Tuesday 24 February 2026  00:57:07 +0000 (0:00:02.951)       0:09:53.160 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-mds : Containerized.yml] ********************************************",
                                "Tuesday 24 February 2026  00:57:08 +0000 (0:00:00.381)       0:09:53.542 ****** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-mds/tasks/containerized.yml for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m",
                                "",
                                "TASK [ceph-mds : Include_tasks systemd.yml] ************************************",
                                "Tuesday 24 February 2026  00:57:09 +0000 (0:00:01.029)       0:09:54.571 ****** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-mds/tasks/systemd.yml for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m",
                                "",
                                "TASK [ceph-mds : Generate systemd unit file] ***********************************",
                                "Tuesday 24 February 2026  00:57:09 +0000 (0:00:00.670)       0:09:55.242 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-mds : Generate systemd ceph-mds target file] ************************",
                                "Tuesday 24 February 2026  00:57:11 +0000 (0:00:01.400)       0:09:56.643 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-mds : Enable ceph-mds.target] ***************************************",
                                "Tuesday 24 February 2026  00:57:13 +0000 (0:00:01.768)       0:09:58.411 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-mds : Systemd start mds container] **********************************",
                                "Tuesday 24 February 2026  00:57:14 +0000 (0:00:01.862)       0:10:00.274 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-mds : Wait for mds socket to exist] *********************************",
                                "Tuesday 24 February 2026  00:57:17 +0000 (0:00:02.157)       0:10:02.431 ****** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Make tempdir for scripts] **********************",
                                "Tuesday 24 February 2026  00:57:18 +0000 (0:00:01.577)       0:10:04.009 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Mdss handler] **********************************",
                                "Tuesday 24 February 2026  00:57:19 +0000 (0:00:00.753)       0:10:04.763 ****** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-handler/tasks/handler_mdss.yml for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Set _mds_handler_called before restart] ********",
                                "Tuesday 24 February 2026  00:57:20 +0000 (0:00:00.985)       0:10:05.749 ****** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Copy mds restart script] ***********************",
                                "Tuesday 24 February 2026  00:57:20 +0000 (0:00:00.391)       0:10:06.140 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Restart ceph mds daemon(s)] ********************",
                                "Tuesday 24 February 2026  00:57:22 +0000 (0:00:01.286)       0:10:07.427 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-3) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-4) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-5) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Set _mds_handler_called after restart] *********",
                                "Tuesday 24 February 2026  00:57:23 +0000 (0:00:01.240)       0:10:08.668 ****** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "PLAY [Apply role ceph-rgw] *****************************************************",
                                "",
                                "TASK [ceph-handler : Include check_running_cluster.yml] ************************",
                                "Tuesday 24 February 2026  00:57:24 +0000 (0:00:01.020)       0:10:09.688 ****** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-handler/tasks/check_running_cluster.yml for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m",
                                "",
                                "TASK [ceph-handler : Include check_running_containers.yml] *********************",
                                "Tuesday 24 February 2026  00:57:24 +0000 (0:00:00.589)       0:10:10.277 ****** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-handler/tasks/check_running_containers.yml for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a mon container] ********************************",
                                "Tuesday 24 February 2026  00:57:25 +0000 (0:00:00.975)       0:10:11.253 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for an osd container] *******************************",
                                "Tuesday 24 February 2026  00:57:26 +0000 (0:00:00.340)       0:10:11.594 ****** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a mds container] ********************************",
                                "Tuesday 24 February 2026  00:57:26 +0000 (0:00:00.759)       0:10:12.354 ****** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a rgw container] ********************************",
                                "Tuesday 24 February 2026  00:57:28 +0000 (0:00:01.124)       0:10:13.479 ****** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a mgr container] ********************************",
                                "Tuesday 24 February 2026  00:57:28 +0000 (0:00:00.779)       0:10:14.258 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a rbd mirror container] *************************",
                                "Tuesday 24 February 2026  00:57:29 +0000 (0:00:00.342)       0:10:14.601 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a nfs container] ********************************",
                                "Tuesday 24 February 2026  00:57:29 +0000 (0:00:00.343)       0:10:14.944 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a ceph-crash container] *************************",
                                "Tuesday 24 February 2026  00:57:29 +0000 (0:00:00.348)       0:10:15.293 ****** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Check for a ceph-exporter container] **********************",
                                "Tuesday 24 February 2026  00:57:31 +0000 (0:00:01.224)       0:10:16.518 ****** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Include check_socket_non_container.yml] *******************",
                                "Tuesday 24 February 2026  00:57:31 +0000 (0:00:00.785)       0:10:17.303 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_mon_status] ******************************",
                                "Tuesday 24 February 2026  00:57:32 +0000 (0:00:00.326)       0:10:17.629 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_osd_status] ******************************",
                                "Tuesday 24 February 2026  00:57:32 +0000 (0:00:00.345)       0:10:17.975 ****** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_mds_status] ******************************",
                                "Tuesday 24 February 2026  00:57:33 +0000 (0:00:00.805)       0:10:18.780 ****** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_rgw_status] ******************************",
                                "Tuesday 24 February 2026  00:57:33 +0000 (0:00:00.361)       0:10:19.141 ****** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_nfs_status] ******************************",
                                "Tuesday 24 February 2026  00:57:34 +0000 (0:00:00.414)       0:10:19.556 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_rbd_status] ******************************",
                                "Tuesday 24 February 2026  00:57:34 +0000 (0:00:00.317)       0:10:19.874 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_mgr_status] ******************************",
                                "Tuesday 24 February 2026  00:57:35 +0000 (0:00:00.657)       0:10:20.531 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_crash_status] ****************************",
                                "Tuesday 24 February 2026  00:57:35 +0000 (0:00:00.396)       0:10:20.928 ****** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-handler : Set_fact handler_exporter_status] *************************",
                                "Tuesday 24 February 2026  00:57:35 +0000 (0:00:00.415)       0:10:21.343 ****** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-rgw : Include common.yml] *******************************************",
                                "Tuesday 24 February 2026  00:57:36 +0000 (0:00:00.855)       0:10:22.198 ****** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-rgw/tasks/common.yml for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m",
                                "",
                                "TASK [ceph-rgw : Get keys from monitors] ***************************************",
                                "Tuesday 24 February 2026  00:57:37 +0000 (0:00:00.591)       0:10:22.790 ****** ",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=None) \u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3 -> {{ groups.get(mon_group_name)[0] }}]\u001b[0m",
                                "",
                                "TASK [ceph-rgw : Copy ceph key(s) if needed] ***********************************",
                                "Tuesday 24 February 2026  00:57:39 +0000 (0:00:02.439)       0:10:25.230 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=None)\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=None) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=None)\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=None) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=None)\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=None) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-rgw : Copy SSL certificate & key data to certificate path] **********",
                                "Tuesday 24 February 2026  00:57:41 +0000 (0:00:01.656)       0:10:26.886 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-rgw : Include_tasks pre_requisite.yml] ******************************",
                                "Tuesday 24 February 2026  00:57:41 +0000 (0:00:00.376)       0:10:27.262 ****** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-rgw/tasks/pre_requisite.yml for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m",
                                "",
                                "TASK [ceph-rgw : Create rados gateway directories] *****************************",
                                "Tuesday 24 February 2026  00:57:42 +0000 (0:00:00.552)       0:10:27.814 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item={'instance_name': 'rgw0', 'radosgw_address': '192.168.16.13', 'radosgw_frontend_port': 8081})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4 -> testbed-node-0(192.168.16.10)] => (item={'instance_name': 'rgw0', 'radosgw_address': '192.168.16.14', 'radosgw_frontend_port': 8081})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item={'instance_name': 'rgw0', 'radosgw_address': '192.168.16.15', 'radosgw_frontend_port': 8081})\u001b[0m",
                                "",
                                "TASK [ceph-rgw : Create rgw keyrings] ******************************************",
                                "Tuesday 24 February 2026  00:57:44 +0000 (0:00:01.634)       0:10:29.449 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3 -> {{ groups[mon_group_name][0] if groups.get(mon_group_name, []) | length > 0 else 'localhost' }}]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4 -> {{ groups[mon_group_name][0] if groups.get(mon_group_name, []) | length > 0 else 'localhost' }}]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5 -> {{ groups[mon_group_name][0] if groups.get(mon_group_name, []) | length > 0 else 'localhost' }}]\u001b[0m",
                                "",
                                "TASK [ceph-rgw : Get keys from monitors] ***************************************",
                                "Tuesday 24 February 2026  00:57:49 +0000 (0:00:05.535)       0:10:34.985 ****** ",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3 -> {{ groups.get(mon_group_name)[0] }}]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5 -> {{ groups.get(mon_group_name)[0] }}]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4 -> {{ groups.get(mon_group_name)[0] }}]\u001b[0m",
                                "",
                                "TASK [ceph-rgw : Copy ceph key(s) if needed] ***********************************",
                                "Tuesday 24 February 2026  00:57:52 +0000 (0:00:02.978)       0:10:37.963 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=None)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=None)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=None)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-rgw : Rgw pool creation tasks] **************************************",
                                "Tuesday 24 February 2026  00:57:53 +0000 (0:00:01.388)       0:10:39.351 ****** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-rgw/tasks/rgw_create_pools.yml for testbed-node-3\u001b[0m",
                                "",
                                "TASK [ceph-rgw : Create ec profile] ********************************************",
                                "Tuesday 24 February 2026  00:57:54 +0000 (0:00:00.251)       0:10:39.603 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'default.rgw.buckets.data', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'default.rgw.buckets.index', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'default.rgw.control', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'default.rgw.log', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'default.rgw.meta', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-rgw : Set crush rule] ***********************************************",
                                "Tuesday 24 February 2026  00:57:55 +0000 (0:00:01.301)       0:10:40.905 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'default.rgw.buckets.data', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'default.rgw.buckets.index', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'default.rgw.control', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'default.rgw.log', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'default.rgw.meta', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-rgw : Create rgw pools] *********************************************",
                                "Tuesday 24 February 2026  00:57:56 +0000 (0:00:00.671)       0:10:41.577 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item={'key': 'default.rgw.buckets.data', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item={'key': 'default.rgw.buckets.index', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item={'key': 'default.rgw.control', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item={'key': 'default.rgw.log', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item={'key': 'default.rgw.meta', 'value': {'pg_num': 8, 'size': 3, 'type': 'replicated'}})\u001b[0m",
                                "",
                                "TASK [ceph-rgw : Include_tasks openstack-keystone.yml] *************************",
                                "Tuesday 24 February 2026  00:58:27 +0000 (0:00:31.785)       0:11:13.362 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-rgw : Include_tasks start_radosgw.yml] ******************************",
                                "Tuesday 24 February 2026  00:58:28 +0000 (0:00:00.358)       0:11:13.721 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-rgw : Include start_docker_rgw.yml] *********************************",
                                "Tuesday 24 February 2026  00:58:28 +0000 (0:00:00.350)       0:11:14.072 ****** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-rgw/tasks/start_docker_rgw.yml for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m",
                                "",
                                "TASK [ceph-rgw : Include_task systemd.yml] *************************************",
                                "Tuesday 24 February 2026  00:58:29 +0000 (0:00:00.875)       0:11:14.947 ****** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-rgw/tasks/systemd.yml for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m",
                                "",
                                "TASK [ceph-rgw : Generate systemd unit file] ***********************************",
                                "Tuesday 24 February 2026  00:58:30 +0000 (0:00:00.553)       0:11:15.500 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-rgw : Generate systemd ceph-radosgw target file] ********************",
                                "Tuesday 24 February 2026  00:58:31 +0000 (0:00:01.332)       0:11:16.833 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-rgw : Enable ceph-radosgw.target] ***********************************",
                                "Tuesday 24 February 2026  00:58:33 +0000 (0:00:01.637)       0:11:18.470 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-rgw : Systemd start rgw container] **********************************",
                                "Tuesday 24 February 2026  00:58:35 +0000 (0:00:02.045)       0:11:20.515 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'instance_name': 'rgw0', 'radosgw_address': '192.168.16.14', 'radosgw_frontend_port': 8081})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'instance_name': 'rgw0', 'radosgw_address': '192.168.16.13', 'radosgw_frontend_port': 8081})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'instance_name': 'rgw0', 'radosgw_address': '192.168.16.15', 'radosgw_frontend_port': 8081})\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Make tempdir for scripts] **********************",
                                "Tuesday 24 February 2026  00:58:37 +0000 (0:00:02.766)       0:11:23.282 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Rgws handler] **********************************",
                                "Tuesday 24 February 2026  00:58:38 +0000 (0:00:00.390)       0:11:23.672 ****** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-handler/tasks/handler_rgws.yml for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Set _rgw_handler_called before restart] ********",
                                "Tuesday 24 February 2026  00:58:38 +0000 (0:00:00.572)       0:11:24.245 ****** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Copy rgw restart script] ***********************",
                                "Tuesday 24 February 2026  00:58:39 +0000 (0:00:00.663)       0:11:24.908 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Restart ceph rgw daemon(s)] ********************",
                                "Tuesday 24 February 2026  00:58:39 +0000 (0:00:00.389)       0:11:25.298 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-3) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-4) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-5) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "RUNNING HANDLER [ceph-handler : Set _rgw_handler_called after restart] *********",
                                "Tuesday 24 February 2026  00:58:40 +0000 (0:00:00.704)       0:11:26.003 ****** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=134 \u001b[0m \u001b[0;33mchanged=35  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=125 \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=127 \u001b[0m \u001b[0;33mchanged=32  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=120 \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=134 \u001b[0m \u001b[0;33mchanged=33  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=119 \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-3\u001b[0m             : \u001b[0;32mok=193 \u001b[0m \u001b[0;33mchanged=45  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=162 \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-4\u001b[0m             : \u001b[0;32mok=175 \u001b[0m \u001b[0;33mchanged=40  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=123 \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-5\u001b[0m             : \u001b[0;32mok=177 \u001b[0m \u001b[0;33mchanged=41  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=121 \u001b[0m rescued=0    ignored=0   ",
                                "",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Tuesday 24 February 2026  00:58:40 +0000 (0:00:00.266)       0:11:26.270 ****** ",
                                "=============================================================================== ",
                                "ceph-osd : Use ceph-volume to create osds ------------------------------ 48.43s",
                                "ceph-container-common : Pulling Ceph container image ------------------- 40.43s",
                                "ceph-mgr : Wait for all mgr to be up ----------------------------------- 36.57s",
                                "ceph-rgw : Create rgw pools -------------------------------------------- 31.79s",
                                "ceph-mon : Waiting for the monitor(s) to form the quorum... ------------ 21.91s",
                                "ceph-mon : Set cluster configs ----------------------------------------- 14.80s",
                                "ceph-osd : Wait for all osd to be up ----------------------------------- 12.82s",
                                "ceph-mds : Create filesystem pools ------------------------------------- 10.52s",
                                "ceph-mgr : Create ceph mgr keyring(s) on a mon node -------------------- 10.21s",
                                "ceph-mon : Fetch ceph initial keys -------------------------------------- 9.27s",
                                "ceph-mgr : Disable ceph mgr enabled modules ----------------------------- 6.72s",
                                "ceph-config : Create ceph initial directories --------------------------- 6.66s",
                                "ceph-mgr : Add modules to ceph-mgr -------------------------------------- 5.72s",
                                "ceph-rgw : Create rgw keyrings ------------------------------------------ 5.54s",
                                "ceph-crash : Create client.crash keyring -------------------------------- 4.66s",
                                "ceph-osd : Apply operating system tuning -------------------------------- 4.28s",
                                "ceph-facts : Set_fact _monitor_addresses - ipv4 ------------------------- 3.98s",
                                "ceph-osd : Systemd start osd -------------------------------------------- 3.88s",
                                "ceph-mds : Create ceph filesystem --------------------------------------- 3.81s",
                                "ceph-crash : Start the ceph-crash service ------------------------------- 3.70s",
                                "\u001b[32m2026-02-24 00:58:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 21cb4ba5-3e33-4fe4-8954-1e96067dfc1b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:58:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 00:58:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ca6226d5-0f70-4445-9f9f-dca4faf1e39d is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:58:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 460230b5-3593-4a25-becb-4c4727892a2e is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:58:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 21cb4ba5-3e33-4fe4-8954-1e96067dfc1b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:58:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 00:58:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ca6226d5-0f70-4445-9f9f-dca4faf1e39d is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:58:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 460230b5-3593-4a25-becb-4c4727892a2e is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:58:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 21cb4ba5-3e33-4fe4-8954-1e96067dfc1b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:58:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 00:58:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ca6226d5-0f70-4445-9f9f-dca4faf1e39d is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:58:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 460230b5-3593-4a25-becb-4c4727892a2e is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:58:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 21cb4ba5-3e33-4fe4-8954-1e96067dfc1b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:58:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 00:58:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ca6226d5-0f70-4445-9f9f-dca4faf1e39d is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:58:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 460230b5-3593-4a25-becb-4c4727892a2e is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:58:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 21cb4ba5-3e33-4fe4-8954-1e96067dfc1b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:58:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 00:58:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ca6226d5-0f70-4445-9f9f-dca4faf1e39d is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:58:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 460230b5-3593-4a25-becb-4c4727892a2e is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:58:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 21cb4ba5-3e33-4fe4-8954-1e96067dfc1b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:58:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 00:59:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ca6226d5-0f70-4445-9f9f-dca4faf1e39d is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:59:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 460230b5-3593-4a25-becb-4c4727892a2e is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:59:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 21cb4ba5-3e33-4fe4-8954-1e96067dfc1b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:59:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 00:59:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ca6226d5-0f70-4445-9f9f-dca4faf1e39d is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:59:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 460230b5-3593-4a25-becb-4c4727892a2e is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:59:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 21cb4ba5-3e33-4fe4-8954-1e96067dfc1b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:59:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 00:59:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ca6226d5-0f70-4445-9f9f-dca4faf1e39d is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:59:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 460230b5-3593-4a25-becb-4c4727892a2e is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:59:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 21cb4ba5-3e33-4fe4-8954-1e96067dfc1b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:59:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 00:59:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ca6226d5-0f70-4445-9f9f-dca4faf1e39d is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:59:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 460230b5-3593-4a25-becb-4c4727892a2e is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:59:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 21cb4ba5-3e33-4fe4-8954-1e96067dfc1b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:59:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 00:59:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ca6226d5-0f70-4445-9f9f-dca4faf1e39d is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:59:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 460230b5-3593-4a25-becb-4c4727892a2e is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:59:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 21cb4ba5-3e33-4fe4-8954-1e96067dfc1b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:59:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 00:59:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ca6226d5-0f70-4445-9f9f-dca4faf1e39d is in state SUCCESS\u001b[0m",
                                "",
                                "",
                                "PLAY [Group hosts based on configuration] **************************************",
                                "",
                                "TASK [Group hosts based on Kolla action] ***************************************",
                                "Tuesday 24 February 2026  00:56:14 +0000 (0:00:00.367)       0:00:00.367 ****** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [Group hosts based on enabled services] ***********************************",
                                "Tuesday 24 February 2026  00:56:14 +0000 (0:00:00.377)       0:00:00.744 ****** ",
                                "\u001b[0;32mok: [testbed-node-0] => (item=enable_opensearch_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item=enable_opensearch_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item=enable_opensearch_True)\u001b[0m",
                                "",
                                "PLAY [Apply role opensearch] ***************************************************",
                                "",
                                "TASK [opensearch : include_tasks] **********************************************",
                                "Tuesday 24 February 2026  00:56:15 +0000 (0:00:00.495)       0:00:01.240 ****** ",
                                "\u001b[0;36mincluded: /ansible/roles/opensearch/tasks/deploy.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [opensearch : Setting sysctl values] **************************************",
                                "Tuesday 24 February 2026  00:56:15 +0000 (0:00:00.627)       0:00:01.868 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'name': 'vm.max_map_count', 'value': 262144})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'name': 'vm.max_map_count', 'value': 262144})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'name': 'vm.max_map_count', 'value': 262144})\u001b[0m",
                                "",
                                "TASK [opensearch : Ensuring config directories exist] **************************",
                                "Tuesday 24 February 2026  00:56:16 +0000 (0:00:00.877)       0:00:02.746 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/opensearch:2024.2', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/opensearch:2024.2', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/opensearch:2024.2', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/opensearch-dashboards:2024.2', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/opensearch-dashboards:2024.2', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/opensearch-dashboards:2024.2', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}}}})\u001b[0m",
                                "",
                                "TASK [opensearch : include_tasks] **********************************************",
                                "Tuesday 24 February 2026  00:56:18 +0000 (0:00:02.098)       0:00:04.844 ****** ",
                                "\u001b[0;36mincluded: /ansible/roles/opensearch/tasks/copy-certs.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [service-cert-copy : opensearch | Copying over extra CA certificates] *****",
                                "Tuesday 24 February 2026  00:56:19 +0000 (0:00:00.746)       0:00:05.591 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/opensearch:2024.2', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/opensearch:2024.2', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/opensearch:2024.2', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/opensearch-dashboards:2024.2', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/opensearch-dashboards:2024.2', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/opensearch-dashboards:2024.2', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}}}})\u001b[0m",
                                "",
                                "TASK [service-cert-copy : opensearch | Copying over backend internal TLS certificate] ***",
                                "Tuesday 24 February 2026  00:56:23 +0000 (0:00:03.963)       0:00:09.554 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/opensearch:2024.2', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/opensearch-dashboards:2024.2', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/opensearch:2024.2', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/opensearch-dashboards:2024.2', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/opensearch:2024.2', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/opensearch-dashboards:2024.2', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [service-cert-copy : opensearch | Copying over backend internal TLS key] ***",
                                "Tuesday 24 February 2026  00:56:25 +0000 (0:00:01.992)       0:00:11.547 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/opensearch:2024.2', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/opensearch-dashboards:2024.2', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/opensearch:2024.2', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/opensearch-dashboards:2024.2', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/opensearch:2024.2', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/opensearch-dashboards:2024.2', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [opensearch : Copying over config.json files for services] ****************",
                                "Tuesday 24 February 2026  00:56:26 +0000 (0:00:01.278)       0:00:12.826 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/opensearch:2024.2', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/opensearch:2024.2', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/opensearch:2024.2', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/opensearch-dashboards:2024.2', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/opensearch-dashboards:2024.2', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/opensearch-dashboards:2024.2', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}}}})\u001b[0m",
                                "",
                                "TASK [opensearch : Copying over opensearch service config file] ****************",
                                "Tuesday 24 February 2026  00:56:29 +0000 (0:00:02.769)       0:00:15.595 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [opensearch : Copying over opensearch-dashboards config file] *************",
                                "Tuesday 24 February 2026  00:56:32 +0000 (0:00:03.450)       0:00:19.046 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [opensearch : Check opensearch containers] ********************************",
                                "Tuesday 24 February 2026  00:56:35 +0000 (0:00:02.551)       0:00:21.597 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/opensearch:2024.2', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/opensearch:2024.2', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'opensearch', 'value': {'container_name': 'opensearch', 'group': 'opensearch', 'enabled': True, 'image': 'registry.osism.tech/kolla/opensearch:2024.2', 'environment': {'OPENSEARCH_JAVA_OPTS': '-Xms1g -Xmx1g -Dlog4j2.formatMsgNoLookups=true'}, 'volumes': ['/etc/kolla/opensearch/:/var/lib/kolla/config_files/', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'opensearch:/var/lib/opensearch/data', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9200'], 'timeout': '30'}, 'haproxy': {'opensearch': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9200', 'frontend_http_extra': ['option dontlog-normal']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/opensearch-dashboards:2024.2', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/opensearch-dashboards:2024.2', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'opensearch-dashboards', 'value': {'container_name': 'opensearch_dashboards', 'group': 'opensearch-dashboards', 'enabled': True, 'environment': {'OPENSEARCH_DASHBOARDS_SECURITY_PLUGIN': 'False'}, 'image': 'registry.osism.tech/kolla/opensearch-dashboards:2024.2', 'volumes': ['/etc/kolla/opensearch-dashboards/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:5601'], 'timeout': '30'}, 'haproxy': {'opensearch-dashboards': {'enabled': True, 'mode': 'http', 'external': False, 'port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}, 'opensearch_dashboards_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '5601', 'listen_port': '5601', 'auth_user': 'opensearch', 'auth_pass': 'password'}}}})\u001b[0m",
                                "",
                                "TASK [opensearch : include_tasks] **********************************************",
                                "Tuesday 24 February 2026  00:56:38 +0000 (0:00:02.653)       0:00:24.251 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [opensearch : Flush handlers] *********************************************",
                                "Tuesday 24 February 2026  00:56:38 +0000 (0:00:00.367)       0:00:24.619 ****** ",
                                "",
                                "TASK [opensearch : Flush handlers] *********************************************",
                                "Tuesday 24 February 2026  00:56:38 +0000 (0:00:00.095)       0:00:24.715 ****** ",
                                "",
                                "TASK [opensearch : Flush handlers] *********************************************",
                                "Tuesday 24 February 2026  00:56:38 +0000 (0:00:00.076)       0:00:24.791 ****** ",
                                "",
                                "RUNNING HANDLER [opensearch : Disable shard allocation] ************************",
                                "Tuesday 24 February 2026  00:56:38 +0000 (0:00:00.076)       0:00:24.867 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "RUNNING HANDLER [opensearch : Perform a flush] *********************************",
                                "Tuesday 24 February 2026  00:56:39 +0000 (0:00:00.857)       0:00:25.725 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "RUNNING HANDLER [opensearch : Restart opensearch container] ********************",
                                "Tuesday 24 February 2026  00:56:39 +0000 (0:00:00.269)       0:00:25.994 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "",
                                "RUNNING HANDLER [opensearch : Restart opensearch-dashboards container] *********",
                                "Tuesday 24 February 2026  00:57:43 +0000 (0:01:04.121)       0:01:30.115 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [opensearch : include_tasks] **********************************************",
                                "Tuesday 24 February 2026  00:58:58 +0000 (0:01:14.250)       0:02:44.366 ****** ",
                                "\u001b[0;36mincluded: /ansible/roles/opensearch/tasks/post-config.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [opensearch : Wait for OpenSearch to become ready] ************************",
                                "Tuesday 24 February 2026  00:58:59 +0000 (0:00:00.848)       0:02:45.214 ****** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [opensearch : Wait for OpenSearch cluster to become healthy] **************",
                                "Tuesday 24 February 2026  00:59:02 +0000 (0:00:03.085)       0:02:48.300 ****** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [opensearch : Check if a log retention policy exists] *********************",
                                "Tuesday 24 February 2026  00:59:05 +0000 (0:00:02.915)       0:02:51.215 ****** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [opensearch : Create new log retention policy] ****************************",
                                "Tuesday 24 February 2026  00:59:07 +0000 (0:00:02.839)       0:02:54.055 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [opensearch : Apply retention policy to existing indices] *****************",
                                "Tuesday 24 February 2026  00:59:11 +0000 (0:00:03.362)       0:02:57.418 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=19  \u001b[0m \u001b[0;33mchanged=11  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=5   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=14  \u001b[0m \u001b[0;33mchanged=9   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=3   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=14  \u001b[0m \u001b[0;33mchanged=9   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=3   \u001b[0m rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Tuesday 24 February 2026  00:59:14 +0000 (0:00:03.258)       0:03:00.676 ****** ",
                                "=============================================================================== ",
                                "opensearch : Restart opensearch-dashboards container ------------------- 74.25s",
                                "opensearch : Restart opensearch container ------------------------------ 64.12s",
                                "service-cert-copy : opensearch | Copying over extra CA certificates ----- 3.96s",
                                "opensearch : Copying over opensearch service config file ---------------- 3.45s",
                                "opensearch : Create new log retention policy ---------------------------- 3.36s",
                                "opensearch : Apply retention policy to existing indices ----------------- 3.26s",
                                "opensearch : Wait for OpenSearch to become ready ------------------------ 3.09s",
                                "opensearch : Wait for OpenSearch cluster to become healthy -------------- 2.92s",
                                "opensearch : Check if a log retention policy exists --------------------- 2.84s",
                                "opensearch : Copying over config.json files for services ---------------- 2.77s",
                                "opensearch : Check opensearch containers -------------------------------- 2.65s",
                                "opensearch : Copying over opensearch-dashboards config file ------------- 2.55s",
                                "opensearch : Ensuring config directories exist -------------------------- 2.10s",
                                "service-cert-copy : opensearch | Copying over backend internal TLS certificate --- 1.99s",
                                "service-cert-copy : opensearch | Copying over backend internal TLS key --- 1.28s",
                                "opensearch : Setting sysctl values -------------------------------------- 0.88s",
                                "opensearch : Disable shard allocation ----------------------------------- 0.86s",
                                "opensearch : include_tasks ---------------------------------------------- 0.85s",
                                "opensearch : include_tasks ---------------------------------------------- 0.75s",
                                "opensearch : include_tasks ---------------------------------------------- 0.63s",
                                "\u001b[32m2026-02-24 00:59:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 460230b5-3593-4a25-becb-4c4727892a2e is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:59:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 21cb4ba5-3e33-4fe4-8954-1e96067dfc1b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:59:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 00:59:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 460230b5-3593-4a25-becb-4c4727892a2e is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:59:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 21cb4ba5-3e33-4fe4-8954-1e96067dfc1b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:59:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 00:59:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 460230b5-3593-4a25-becb-4c4727892a2e is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:59:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 21cb4ba5-3e33-4fe4-8954-1e96067dfc1b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:59:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 00:59:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 460230b5-3593-4a25-becb-4c4727892a2e is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:59:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 21cb4ba5-3e33-4fe4-8954-1e96067dfc1b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:59:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 00:59:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 460230b5-3593-4a25-becb-4c4727892a2e is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:59:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 21cb4ba5-3e33-4fe4-8954-1e96067dfc1b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:59:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 00:59:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 460230b5-3593-4a25-becb-4c4727892a2e is in state STARTED\u001b[0m",
                                "",
                                "",
                                "PLAY [Set kolla_action_mariadb] ************************************************",
                                "",
                                "TASK [Inform the user about the following task] ********************************",
                                "Tuesday 24 February 2026  00:56:13 +0000 (0:00:00.107)       0:00:00.107 ****** ",
                                "\u001b[0;32mok: [localhost] => {\u001b[0m",
                                "\u001b[0;32m    \"msg\": \"The task 'Check MariaDB service' fails if the MariaDB service has not yet been deployed. This is fine.\"\u001b[0m",
                                "\u001b[0;32m}\u001b[0m",
                                "",
                                "TASK [Check MariaDB service] ***************************************************",
                                "Tuesday 24 February 2026  00:56:13 +0000 (0:00:00.068)       0:00:00.175 ****** ",
                                "\u001b[0;31mfatal: [localhost]: FAILED! => {\"changed\": false, \"elapsed\": 2, \"msg\": \"Timeout when waiting for search string MariaDB in 192.168.16.9:3306\"}\u001b[0m",
                                "\u001b[0;36m...ignoring\u001b[0m",
                                "",
                                "TASK [Set kolla_action_mariadb = upgrade if MariaDB is already running] ********",
                                "Tuesday 24 February 2026  00:56:16 +0000 (0:00:02.965)       0:00:03.141 ****** ",
                                "\u001b[0;36mskipping: [localhost]\u001b[0m",
                                "",
                                "TASK [Set kolla_action_mariadb = kolla_action_ng] ******************************",
                                "Tuesday 24 February 2026  00:56:16 +0000 (0:00:00.062)       0:00:03.203 ****** ",
                                "\u001b[0;32mok: [localhost]\u001b[0m",
                                "",
                                "PLAY [Group hosts based on configuration] **************************************",
                                "",
                                "TASK [Group hosts based on Kolla action] ***************************************",
                                "Tuesday 24 February 2026  00:56:17 +0000 (0:00:00.166)       0:00:03.370 ****** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [Group hosts based on enabled services] ***********************************",
                                "Tuesday 24 February 2026  00:56:17 +0000 (0:00:00.350)       0:00:03.721 ****** ",
                                "\u001b[0;32mok: [testbed-node-0] => (item=enable_mariadb_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item=enable_mariadb_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item=enable_mariadb_True)\u001b[0m",
                                "",
                                "PLAY [Apply role mariadb] ******************************************************",
                                "",
                                "TASK [mariadb : Group MariaDB hosts based on shards] ***************************",
                                "Tuesday 24 February 2026  00:56:18 +0000 (0:00:00.639)       0:00:04.360 ****** ",
                                "\u001b[0;32mok: [testbed-node-0] => (item=testbed-node-0)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0] => (item=testbed-node-1)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0] => (item=testbed-node-2)\u001b[0m",
                                "",
                                "TASK [mariadb : include_tasks] *************************************************",
                                "Tuesday 24 February 2026  00:56:18 +0000 (0:00:00.480)       0:00:04.841 ****** ",
                                "\u001b[0;36mincluded: /ansible/roles/mariadb/tasks/deploy.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [mariadb : Ensuring config directories exist] *****************************",
                                "Tuesday 24 February 2026  00:56:19 +0000 (0:00:00.972)       0:00:05.813 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/mariadb-server:2024.2', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.10', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/mariadb-server:2024.2', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.12', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/mariadb-server:2024.2', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.11', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}})\u001b[0m",
                                "",
                                "TASK [mariadb : Ensuring database backup config directory exists] **************",
                                "Tuesday 24 February 2026  00:56:24 +0000 (0:00:05.059)       0:00:10.873 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [mariadb : Copying over my.cnf for mariabackup] ***************************",
                                "Tuesday 24 February 2026  00:56:25 +0000 (0:00:01.083)       0:00:11.957 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [mariadb : Copying over config.json files for services] *******************",
                                "Tuesday 24 February 2026  00:56:27 +0000 (0:00:02.056)       0:00:14.013 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/mariadb-server:2024.2', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.11', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/mariadb-server:2024.2', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.10', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/mariadb-server:2024.2', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.12', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}})\u001b[0m",
                                "",
                                "TASK [mariadb : Copying over config.json files for mariabackup] ****************",
                                "Tuesday 24 February 2026  00:56:32 +0000 (0:00:04.539)       0:00:18.553 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [mariadb : Copying over galera.cnf] ***************************************",
                                "Tuesday 24 February 2026  00:56:33 +0000 (0:00:01.322)       0:00:19.875 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [mariadb : include_tasks] *************************************************",
                                "Tuesday 24 February 2026  00:56:39 +0000 (0:00:05.462)       0:00:25.338 ****** ",
                                "\u001b[0;36mincluded: /ansible/roles/mariadb/tasks/copy-certs.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [service-cert-copy : mariadb | Copying over extra CA certificates] ********",
                                "Tuesday 24 February 2026  00:56:39 +0000 (0:00:00.627)       0:00:25.965 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/mariadb-server:2024.2', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.11', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/mariadb-server:2024.2', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.12', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/mariadb-server:2024.2', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.10', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [service-cert-copy : mariadb | Copying over backend internal TLS certificate] ***",
                                "Tuesday 24 February 2026  00:56:44 +0000 (0:00:04.403)       0:00:30.368 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/mariadb-server:2024.2', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.10', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/mariadb-server:2024.2', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.12', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/mariadb-server:2024.2', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.11', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [service-cert-copy : mariadb | Copying over backend internal TLS key] *****",
                                "Tuesday 24 February 2026  00:56:47 +0000 (0:00:03.636)       0:00:34.005 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/mariadb-server:2024.2', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.11', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/mariadb-server:2024.2', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.10', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/mariadb-server:2024.2', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.12', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [mariadb : Check mariadb containers] **************************************",
                                "Tuesday 24 February 2026  00:56:51 +0000 (0:00:03.614)       0:00:37.620 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/mariadb-server:2024.2', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.10', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/mariadb-server:2024.2', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.11', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'mariadb', 'value': {'container_name': 'mariadb', 'group': 'mariadb_shard_0', 'enabled': True, 'image': 'registry.osism.tech/kolla/mariadb-server:2024.2', 'volumes': ['/etc/kolla/mariadb/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/hosts:/etc/hosts:ro', '/etc/timezone:/etc/timezone:ro', 'mariadb:/var/lib/mysql', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/clustercheck'], 'timeout': '30'}, 'environment': {'MYSQL_USERNAME': 'monitor', 'MYSQL_PASSWORD': 'iek7ooth9miesodoh2ongohcaachah0I', 'MYSQL_HOST': '192.168.16.12', 'AVAILABLE_WHEN_DONOR': '1'}, 'haproxy': {'mariadb': {'enabled': True, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s', ''], 'custom_member_list': [' server testbed-node-0 192.168.16.10:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 192.168.16.11:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 192.168.16.12:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}, 'mariadb_external_lb': {'enabled': False, 'mode': 'tcp', 'port': '3306', 'listen_port': '3306', 'frontend_tcp_extra': ['option clitcpka', 'timeout client 3600s'], 'backend_tcp_extra': ['option srvtcpka', 'timeout server 3600s'], 'custom_member_list': [' server testbed-node-0 testbed-node-0:3306 check port 3306 inter 2000 rise 2 fall 5', ' server testbed-node-1 testbed-node-1:3306 check port 3306 inter 2000 rise 2 fall 5 backup', ' server testbed-node-2 testbed-node-2:3306 check port 3306 inter 2000 rise 2 fall 5 backup', '']}}}})\u001b[0m",
                                "",
                                "TASK [mariadb : Create MariaDB volume] *****************************************",
                                "Tuesday 24 February 2026  00:56:55 +0000 (0:00:04.441)       0:00:42.062 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [mariadb : Divide hosts by their MariaDB volume availability] *************",
                                "Tuesday 24 February 2026  00:56:56 +0000 (0:00:00.816)       0:00:42.878 ****** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [mariadb : Establish whether the cluster has already existed] *************",
                                "Tuesday 24 February 2026  00:56:57 +0000 (0:00:00.608)       0:00:43.487 ****** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [mariadb : Check MariaDB service port liveness] ***************************",
                                "Tuesday 24 February 2026  00:56:57 +0000 (0:00:00.350)       0:00:43.837 ****** ",
                                "\u001b[0;31mfatal: [testbed-node-1]: FAILED! => {\"changed\": false, \"elapsed\": 10, \"msg\": \"Timeout when waiting for search string MariaDB in 192.168.16.11:3306\"}\u001b[0m",
                                "\u001b[0;36m...ignoring\u001b[0m",
                                "\u001b[0;31mfatal: [testbed-node-0]: FAILED! => {\"changed\": false, \"elapsed\": 10, \"msg\": \"Timeout when waiting for search string MariaDB in 192.168.16.10:3306\"}\u001b[0m",
                                "\u001b[0;36m...ignoring\u001b[0m",
                                "\u001b[0;31mfatal: [testbed-node-2]: FAILED! => {\"changed\": false, \"elapsed\": 10, \"msg\": \"Timeout when waiting for search string MariaDB in 192.168.16.12:3306\"}\u001b[0m",
                                "\u001b[0;36m...ignoring\u001b[0m",
                                "",
                                "TASK [mariadb : Divide hosts by their MariaDB service port liveness] ***********",
                                "Tuesday 24 February 2026  00:57:08 +0000 (0:00:10.988)       0:00:54.825 ****** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [mariadb : Fail on existing but stopped cluster] **************************",
                                "Tuesday 24 February 2026  00:57:09 +0000 (0:00:00.470)       0:00:55.296 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [mariadb : Check MariaDB service WSREP sync status] ***********************",
                                "Tuesday 24 February 2026  00:57:09 +0000 (0:00:00.760)       0:00:56.056 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [mariadb : Extract MariaDB service WSREP sync status] *********************",
                                "Tuesday 24 February 2026  00:57:10 +0000 (0:00:00.641)       0:00:56.698 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [mariadb : Divide hosts by their MariaDB service WSREP sync status] *******",
                                "Tuesday 24 February 2026  00:57:10 +0000 (0:00:00.511)       0:00:57.210 ****** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [mariadb : Fail when MariaDB services are not synced across the whole cluster] ***",
                                "Tuesday 24 February 2026  00:57:11 +0000 (0:00:00.475)       0:00:57.686 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [mariadb : include_tasks] *************************************************",
                                "Tuesday 24 February 2026  00:57:12 +0000 (0:00:00.881)       0:00:58.567 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/roles/mariadb/tasks/bootstrap_cluster.yml for testbed-node-0\u001b[0m",
                                "",
                                "TASK [mariadb : Running MariaDB bootstrap container] ***************************",
                                "Tuesday 24 February 2026  00:57:12 +0000 (0:00:00.451)       0:00:59.018 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [mariadb : Store bootstrap host name into facts] **************************",
                                "Tuesday 24 February 2026  00:57:23 +0000 (0:00:10.911)       0:01:09.930 ****** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [mariadb : include_tasks] *************************************************",
                                "Tuesday 24 February 2026  00:57:23 +0000 (0:00:00.147)       0:01:10.077 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [mariadb : Starting first MariaDB container] *******************",
                                "Tuesday 24 February 2026  00:57:24 +0000 (0:00:01.184)       0:01:11.262 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "RUNNING HANDLER [mariadb : Wait for first MariaDB service port liveness] *******",
                                "Tuesday 24 February 2026  00:57:33 +0000 (0:00:08.946)       0:01:20.208 ****** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "RUNNING HANDLER [mariadb : Wait for first MariaDB service to sync WSREP] *******",
                                "Tuesday 24 February 2026  00:57:35 +0000 (0:00:01.641)       0:01:21.850 ****** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "RUNNING HANDLER [mariadb : Ensure MariaDB is running normally on bootstrap host] ***",
                                "Tuesday 24 February 2026  00:57:38 +0000 (0:00:02.889)       0:01:24.739 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "RUNNING HANDLER [mariadb : Restart MariaDB on existing cluster members] ********",
                                "Tuesday 24 February 2026  00:57:38 +0000 (0:00:00.124)       0:01:24.863 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [mariadb : Start MariaDB on new nodes] *************************",
                                "Tuesday 24 February 2026  00:57:38 +0000 (0:00:00.342)       0:01:25.206 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[1;35m[WARNING]: Could not match supplied host pattern, ignoring: mariadb_restart\u001b[0m",
                                "",
                                "PLAY [Restart mariadb services] ************************************************",
                                "\u001b[0;36mskipping: no hosts matched\u001b[0m",
                                "",
                                "PLAY [Start mariadb services] **************************************************",
                                "",
                                "TASK [mariadb : Restart MariaDB container] *************************************",
                                "Tuesday 24 February 2026  00:57:39 +0000 (0:00:00.603)       0:01:25.810 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [mariadb : Wait for MariaDB service port liveness] ************************",
                                "Tuesday 24 February 2026  00:57:58 +0000 (0:00:18.602)       0:01:44.413 ****** ",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [mariadb : Wait for MariaDB service to sync WSREP] ************************",
                                "Tuesday 24 February 2026  00:58:13 +0000 (0:00:15.600)       0:02:00.013 ****** ",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "",
                                "PLAY [Start mariadb services] **************************************************",
                                "",
                                "TASK [mariadb : Restart MariaDB container] *************************************",
                                "Tuesday 24 February 2026  00:58:16 +0000 (0:00:02.675)       0:02:02.689 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [mariadb : Wait for MariaDB service port liveness] ************************",
                                "Tuesday 24 February 2026  00:58:35 +0000 (0:00:19.127)       0:02:21.817 ****** ",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [mariadb : Wait for MariaDB service to sync WSREP] ************************",
                                "Tuesday 24 February 2026  00:58:51 +0000 (0:00:15.622)       0:02:37.439 ****** ",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "PLAY [Restart bootstrap mariadb service] ***************************************",
                                "",
                                "TASK [mariadb : Restart MariaDB container] *************************************",
                                "Tuesday 24 February 2026  00:58:53 +0000 (0:00:02.735)       0:02:40.175 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [mariadb : Wait for MariaDB service port liveness] ************************",
                                "Tuesday 24 February 2026  00:59:06 +0000 (0:00:12.798)       0:02:52.973 ****** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [mariadb : Wait for MariaDB service to sync WSREP] ************************",
                                "Tuesday 24 February 2026  00:59:11 +0000 (0:00:04.639)       0:02:57.613 ****** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "PLAY [Apply mariadb post-configuration] ****************************************",
                                "",
                                "TASK [Include mariadb post-deploy.yml] *****************************************",
                                "Tuesday 24 February 2026  00:59:14 +0000 (0:00:02.977)       0:03:00.590 ****** ",
                                "\u001b[0;36mincluded: mariadb for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [mariadb : Creating shard root mysql user] ********************************",
                                "Tuesday 24 February 2026  00:59:14 +0000 (0:00:00.492)       0:03:01.083 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [mariadb : Creating mysql monitor user] ***********************************",
                                "Tuesday 24 February 2026  00:59:17 +0000 (0:00:02.868)       0:03:03.951 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [mariadb : Creating database backup user and setting permissions] *********",
                                "Tuesday 24 February 2026  00:59:20 +0000 (0:00:03.083)       0:03:07.034 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [mariadb : Granting permissions on Mariabackup database to backup user] ***",
                                "Tuesday 24 February 2026  00:59:23 +0000 (0:00:02.893)       0:03:09.928 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [mariadb : Wait for MariaDB service to be ready through VIP] **************",
                                "Tuesday 24 February 2026  00:59:26 +0000 (0:00:02.522)       0:03:12.450 ****** ",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [Include mariadb post-upgrade.yml] ****************************************",
                                "Tuesday 24 February 2026  00:59:29 +0000 (0:00:03.582)       0:03:16.032 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;32mlocalhost\u001b[0m                  : \u001b[0;32mok=3   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=1   \u001b[0m rescued=0    \u001b[1;35mignored=1   \u001b[0m",
                                "\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=34  \u001b[0m \u001b[0;33mchanged=16  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=11  \u001b[0m rescued=0    \u001b[1;35mignored=1   \u001b[0m",
                                "\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=20  \u001b[0m \u001b[0;33mchanged=7   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=18  \u001b[0m rescued=0    \u001b[1;35mignored=1   \u001b[0m",
                                "\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=20  \u001b[0m \u001b[0;33mchanged=7   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=18  \u001b[0m rescued=0    \u001b[1;35mignored=1   \u001b[0m",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Tuesday 24 February 2026  00:59:30 +0000 (0:00:00.274)       0:03:16.306 ****** ",
                                "=============================================================================== ",
                                "mariadb : Restart MariaDB container ------------------------------------ 37.73s",
                                "mariadb : Wait for MariaDB service port liveness ----------------------- 31.22s",
                                "mariadb : Restart MariaDB container ------------------------------------ 12.80s",
                                "mariadb : Check MariaDB service port liveness -------------------------- 10.99s",
                                "mariadb : Running MariaDB bootstrap container -------------------------- 10.91s",
                                "mariadb : Starting first MariaDB container ------------------------------ 8.95s",
                                "mariadb : Copying over galera.cnf --------------------------------------- 5.46s",
                                "mariadb : Wait for MariaDB service to sync WSREP ------------------------ 5.41s",
                                "mariadb : Ensuring config directories exist ----------------------------- 5.06s",
                                "mariadb : Wait for MariaDB service port liveness ------------------------ 4.64s",
                                "mariadb : Copying over config.json files for services ------------------- 4.54s",
                                "mariadb : Check mariadb containers -------------------------------------- 4.44s",
                                "service-cert-copy : mariadb | Copying over extra CA certificates -------- 4.40s",
                                "service-cert-copy : mariadb | Copying over backend internal TLS certificate --- 3.64s",
                                "service-cert-copy : mariadb | Copying over backend internal TLS key ----- 3.61s",
                                "mariadb : Wait for MariaDB service to be ready through VIP -------------- 3.58s",
                                "mariadb : Creating mysql monitor user ----------------------------------- 3.08s",
                                "mariadb : Wait for MariaDB service to sync WSREP ------------------------ 2.98s",
                                "Check MariaDB service --------------------------------------------------- 2.97s",
                                "mariadb : Creating database backup user and setting permissions --------- 2.89s",
                                "\u001b[32m2026-02-24 00:59:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 21cb4ba5-3e33-4fe4-8954-1e96067dfc1b is in state SUCCESS\u001b[0m",
                                "\u001b[32m2026-02-24 00:59:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 00:59:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c587e61e-d18f-4e39-9404-027cd3697185 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:59:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9cb83792-8f4e-4574-a78c-9b0e654d6fd6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:59:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 460230b5-3593-4a25-becb-4c4727892a2e is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:59:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 00:59:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c587e61e-d18f-4e39-9404-027cd3697185 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:59:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9cb83792-8f4e-4574-a78c-9b0e654d6fd6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:59:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 460230b5-3593-4a25-becb-4c4727892a2e is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:59:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 00:59:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c587e61e-d18f-4e39-9404-027cd3697185 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:59:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9cb83792-8f4e-4574-a78c-9b0e654d6fd6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:59:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 460230b5-3593-4a25-becb-4c4727892a2e is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:59:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 00:59:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c587e61e-d18f-4e39-9404-027cd3697185 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:59:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9cb83792-8f4e-4574-a78c-9b0e654d6fd6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:59:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 460230b5-3593-4a25-becb-4c4727892a2e is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:59:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 00:59:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c587e61e-d18f-4e39-9404-027cd3697185 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:59:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9cb83792-8f4e-4574-a78c-9b0e654d6fd6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:59:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 460230b5-3593-4a25-becb-4c4727892a2e is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:59:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 00:59:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c587e61e-d18f-4e39-9404-027cd3697185 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:59:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9cb83792-8f4e-4574-a78c-9b0e654d6fd6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:59:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 460230b5-3593-4a25-becb-4c4727892a2e is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:59:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 00:59:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c587e61e-d18f-4e39-9404-027cd3697185 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:59:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9cb83792-8f4e-4574-a78c-9b0e654d6fd6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:59:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 460230b5-3593-4a25-becb-4c4727892a2e is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:59:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 00:59:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c587e61e-d18f-4e39-9404-027cd3697185 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:59:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9cb83792-8f4e-4574-a78c-9b0e654d6fd6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:59:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 460230b5-3593-4a25-becb-4c4727892a2e is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:59:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 00:59:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c587e61e-d18f-4e39-9404-027cd3697185 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:59:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9cb83792-8f4e-4574-a78c-9b0e654d6fd6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:59:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 460230b5-3593-4a25-becb-4c4727892a2e is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 00:59:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 01:00:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c587e61e-d18f-4e39-9404-027cd3697185 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:00:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9cb83792-8f4e-4574-a78c-9b0e654d6fd6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:00:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 460230b5-3593-4a25-becb-4c4727892a2e is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:00:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 01:00:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c587e61e-d18f-4e39-9404-027cd3697185 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:00:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9cb83792-8f4e-4574-a78c-9b0e654d6fd6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:00:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 460230b5-3593-4a25-becb-4c4727892a2e is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:00:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 01:00:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c587e61e-d18f-4e39-9404-027cd3697185 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:00:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9cb83792-8f4e-4574-a78c-9b0e654d6fd6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:00:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 460230b5-3593-4a25-becb-4c4727892a2e is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:00:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 01:00:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c587e61e-d18f-4e39-9404-027cd3697185 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:00:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9cb83792-8f4e-4574-a78c-9b0e654d6fd6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:00:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 460230b5-3593-4a25-becb-4c4727892a2e is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:00:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 01:00:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c587e61e-d18f-4e39-9404-027cd3697185 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:00:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9cb83792-8f4e-4574-a78c-9b0e654d6fd6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:00:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 460230b5-3593-4a25-becb-4c4727892a2e is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:00:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 01:00:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c587e61e-d18f-4e39-9404-027cd3697185 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:00:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9cb83792-8f4e-4574-a78c-9b0e654d6fd6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:00:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 460230b5-3593-4a25-becb-4c4727892a2e is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:00:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 01:00:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c587e61e-d18f-4e39-9404-027cd3697185 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:00:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9cb83792-8f4e-4574-a78c-9b0e654d6fd6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:00:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 460230b5-3593-4a25-becb-4c4727892a2e is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:00:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 01:00:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c587e61e-d18f-4e39-9404-027cd3697185 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:00:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9cb83792-8f4e-4574-a78c-9b0e654d6fd6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:00:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 460230b5-3593-4a25-becb-4c4727892a2e is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:00:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 01:00:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c587e61e-d18f-4e39-9404-027cd3697185 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:00:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9cb83792-8f4e-4574-a78c-9b0e654d6fd6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:00:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 460230b5-3593-4a25-becb-4c4727892a2e is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:00:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 01:00:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c587e61e-d18f-4e39-9404-027cd3697185 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:00:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9cb83792-8f4e-4574-a78c-9b0e654d6fd6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:00:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 460230b5-3593-4a25-becb-4c4727892a2e is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:00:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 01:00:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c587e61e-d18f-4e39-9404-027cd3697185 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:00:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9cb83792-8f4e-4574-a78c-9b0e654d6fd6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:00:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 460230b5-3593-4a25-becb-4c4727892a2e is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:00:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 01:00:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c587e61e-d18f-4e39-9404-027cd3697185 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:00:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9cb83792-8f4e-4574-a78c-9b0e654d6fd6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:00:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 460230b5-3593-4a25-becb-4c4727892a2e is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:00:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 01:00:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c587e61e-d18f-4e39-9404-027cd3697185 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:00:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9cb83792-8f4e-4574-a78c-9b0e654d6fd6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:00:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 460230b5-3593-4a25-becb-4c4727892a2e is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:00:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 01:00:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c587e61e-d18f-4e39-9404-027cd3697185 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:00:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9cb83792-8f4e-4574-a78c-9b0e654d6fd6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:00:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 460230b5-3593-4a25-becb-4c4727892a2e is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:00:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 01:00:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c587e61e-d18f-4e39-9404-027cd3697185 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:00:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9cb83792-8f4e-4574-a78c-9b0e654d6fd6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:00:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 460230b5-3593-4a25-becb-4c4727892a2e is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:00:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 01:00:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c587e61e-d18f-4e39-9404-027cd3697185 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:00:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9cb83792-8f4e-4574-a78c-9b0e654d6fd6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:00:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 460230b5-3593-4a25-becb-4c4727892a2e is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:00:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 01:00:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c587e61e-d18f-4e39-9404-027cd3697185 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:00:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9cb83792-8f4e-4574-a78c-9b0e654d6fd6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:00:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 460230b5-3593-4a25-becb-4c4727892a2e is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:00:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 01:00:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c587e61e-d18f-4e39-9404-027cd3697185 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:00:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9cb83792-8f4e-4574-a78c-9b0e654d6fd6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:00:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 460230b5-3593-4a25-becb-4c4727892a2e is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:00:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 01:00:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c587e61e-d18f-4e39-9404-027cd3697185 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:00:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9cb83792-8f4e-4574-a78c-9b0e654d6fd6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:00:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 460230b5-3593-4a25-becb-4c4727892a2e is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:00:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 01:00:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c587e61e-d18f-4e39-9404-027cd3697185 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:00:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9cb83792-8f4e-4574-a78c-9b0e654d6fd6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:00:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 460230b5-3593-4a25-becb-4c4727892a2e is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:00:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 01:01:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c587e61e-d18f-4e39-9404-027cd3697185 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:01:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9cb83792-8f4e-4574-a78c-9b0e654d6fd6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:01:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 460230b5-3593-4a25-becb-4c4727892a2e is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:01:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 01:01:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c587e61e-d18f-4e39-9404-027cd3697185 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:01:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9cb83792-8f4e-4574-a78c-9b0e654d6fd6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:01:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 460230b5-3593-4a25-becb-4c4727892a2e is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:01:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 01:01:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c5f0c529-69aa-4268-97af-bde1796abaa2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:01:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c587e61e-d18f-4e39-9404-027cd3697185 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:01:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9cb83792-8f4e-4574-a78c-9b0e654d6fd6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:01:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 460230b5-3593-4a25-becb-4c4727892a2e is in state SUCCESS\u001b[0m",
                                "",
                                "\u001b[1;35m[WARNING]: Collection community.general does not support Ansible version\u001b[0m",
                                "\u001b[1;35m2.16.14\u001b[0m",
                                "",
                                "PLAY [Create ceph pools] *******************************************************",
                                "",
                                "TASK [ceph-facts : Include facts.yml] ******************************************",
                                "Tuesday 24 February 2026  00:58:46 +0000 (0:00:00.613)       0:00:00.613 ****** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-facts/tasks/facts.yml for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m",
                                "",
                                "TASK [ceph-facts : Check if it is atomic host] *********************************",
                                "Tuesday 24 February 2026  00:58:47 +0000 (0:00:00.684)       0:00:01.297 ****** ",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact is_atomic] *****************************************",
                                "Tuesday 24 February 2026  00:58:47 +0000 (0:00:00.646)       0:00:01.944 ****** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Check if podman binary is present] **************************",
                                "Tuesday 24 February 2026  00:58:48 +0000 (0:00:00.358)       0:00:02.303 ****** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact container_binary] **********************************",
                                "Tuesday 24 February 2026  00:58:49 +0000 (0:00:00.988)       0:00:03.292 ****** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact ceph_cmd] ******************************************",
                                "Tuesday 24 February 2026  00:58:49 +0000 (0:00:00.362)       0:00:03.655 ****** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact discovered_interpreter_python] *********************",
                                "Tuesday 24 February 2026  00:58:50 +0000 (0:00:00.354)       0:00:04.009 ****** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact discovered_interpreter_python if not previously set] ***",
                                "Tuesday 24 February 2026  00:58:50 +0000 (0:00:00.374)       0:00:04.384 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact ceph_release ceph_stable_release] ******************",
                                "Tuesday 24 February 2026  00:58:51 +0000 (0:00:00.604)       0:00:04.988 ****** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact monitor_name ansible_facts['hostname']] ************",
                                "Tuesday 24 February 2026  00:58:51 +0000 (0:00:00.348)       0:00:05.337 ****** ",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item=testbed-node-0)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-1(192.168.16.11)] => (item=testbed-node-1)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-2(192.168.16.12)] => (item=testbed-node-2)\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact container_exec_cmd] ********************************",
                                "Tuesday 24 February 2026  00:58:52 +0000 (0:00:00.714)       0:00:06.051 ****** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Find a running mon container] *******************************",
                                "Tuesday 24 February 2026  00:58:52 +0000 (0:00:00.494)       0:00:06.546 ****** ",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item=testbed-node-0)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-1(192.168.16.11)] => (item=testbed-node-1)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-2(192.168.16.12)] => (item=testbed-node-2)\u001b[0m",
                                "",
                                "TASK [ceph-facts : Check for a ceph mon socket] ********************************",
                                "Tuesday 24 February 2026  00:58:54 +0000 (0:00:02.319)       0:00:08.866 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-0) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-1) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Check if the ceph mon socket is in-use] *********************",
                                "Tuesday 24 February 2026  00:58:55 +0000 (0:00:00.895)       0:00:09.762 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-0', 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-1', 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-2', 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact running_mon - non_container] ***********************",
                                "Tuesday 24 February 2026  00:58:56 +0000 (0:00:00.874)       0:00:10.637 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': {'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-0', 'ansible_loop_var': 'item'}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': {'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-1', 'ansible_loop_var': 'item'}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': {'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'not containerized_deployment | bool', 'item': 'testbed-node-2', 'ansible_loop_var': 'item'}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact running_mon - container] ***************************",
                                "Tuesday 24 February 2026  00:58:57 +0000 (0:00:00.402)       0:00:11.039 ****** ",
                                "\u001b[0;32mok: [testbed-node-3] => (item={'changed': False, 'stdout': '1cb145726d67', 'stderr': '', 'rc': 0, 'cmd': ['docker', 'ps', '-q', '--filter', 'name=ceph-mon-testbed-node-0'], 'start': '2026-02-24 00:58:53.303404', 'end': '2026-02-24 00:58:53.343080', 'delta': '0:00:00.039676', 'msg': '', 'invocation': {'module_args': {'_raw_params': 'docker ps -q --filter name=ceph-mon-testbed-node-0', '_uses_shell': False, 'expand_argument_vars': True, 'stdin_add_newline': True, 'strip_empty_ends': True, 'argv': None, 'chdir': None, 'executable': None, 'creates': None, 'removes': None, 'stdin': None}}, 'stdout_lines': ['1cb145726d67'], 'stderr_lines': [], 'failed': False, 'failed_when_result': False, 'item': 'testbed-node-0', 'ansible_loop_var': 'item'})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item={'changed': False, 'stdout': '65c197c778ba', 'stderr': '', 'rc': 0, 'cmd': ['docker', 'ps', '-q', '--filter', 'name=ceph-mon-testbed-node-1'], 'start': '2026-02-24 00:58:54.123679', 'end': '2026-02-24 00:58:54.173430', 'delta': '0:00:00.049751', 'msg': '', 'invocation': {'module_args': {'_raw_params': 'docker ps -q --filter name=ceph-mon-testbed-node-1', '_uses_shell': False, 'expand_argument_vars': True, 'stdin_add_newline': True, 'strip_empty_ends': True, 'argv': None, 'chdir': None, 'executable': None, 'creates': None, 'removes': None, 'stdin': None}}, 'stdout_lines': ['65c197c778ba'], 'stderr_lines': [], 'failed': False, 'failed_when_result': False, 'item': 'testbed-node-1', 'ansible_loop_var': 'item'})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item={'changed': False, 'stdout': '839d7a58ab74', 'stderr': '', 'rc': 0, 'cmd': ['docker', 'ps', '-q', '--filter', 'name=ceph-mon-testbed-node-2'], 'start': '2026-02-24 00:58:54.681655', 'end': '2026-02-24 00:58:54.723543', 'delta': '0:00:00.041888', 'msg': '', 'invocation': {'module_args': {'_raw_params': 'docker ps -q --filter name=ceph-mon-testbed-node-2', '_uses_shell': False, 'expand_argument_vars': True, 'stdin_add_newline': True, 'strip_empty_ends': True, 'argv': None, 'chdir': None, 'executable': None, 'creates': None, 'removes': None, 'stdin': None}}, 'stdout_lines': ['839d7a58ab74'], 'stderr_lines': [], 'failed': False, 'failed_when_result': False, 'item': 'testbed-node-2', 'ansible_loop_var': 'item'})\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _container_exec_cmd] *******************************",
                                "Tuesday 24 February 2026  00:58:57 +0000 (0:00:00.199)       0:00:11.239 ****** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Get current fsid if cluster is already running] *************",
                                "Tuesday 24 February 2026  00:58:57 +0000 (0:00:00.490)       0:00:11.729 ****** ",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-2(192.168.16.12)]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact current_fsid rc 1] *********************************",
                                "Tuesday 24 February 2026  00:58:59 +0000 (0:00:01.891)       0:00:13.621 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Get current fsid] *******************************************",
                                "Tuesday 24 February 2026  00:58:59 +0000 (0:00:00.310)       0:00:13.931 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact fsid] **********************************************",
                                "Tuesday 24 February 2026  00:59:00 +0000 (0:00:00.379)       0:00:14.310 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact fsid from current_fsid] ****************************",
                                "Tuesday 24 February 2026  00:59:00 +0000 (0:00:00.442)       0:00:14.753 ****** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Generate cluster fsid] **************************************",
                                "Tuesday 24 February 2026  00:59:00 +0000 (0:00:00.126)       0:00:14.880 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact fsid] **********************************************",
                                "Tuesday 24 February 2026  00:59:01 +0000 (0:00:00.218)       0:00:15.098 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Resolve device link(s)] *************************************",
                                "Tuesday 24 February 2026  00:59:01 +0000 (0:00:00.264)       0:00:15.363 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact build devices from resolved symlinks] **************",
                                "Tuesday 24 February 2026  00:59:01 +0000 (0:00:00.283)       0:00:15.646 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Resolve dedicated_device link(s)] ***************************",
                                "Tuesday 24 February 2026  00:59:02 +0000 (0:00:00.444)       0:00:16.091 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact build dedicated_devices from resolved symlinks] ****",
                                "Tuesday 24 February 2026  00:59:02 +0000 (0:00:00.328)       0:00:16.419 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Resolve bluestore_wal_device link(s)] ***********************",
                                "Tuesday 24 February 2026  00:59:02 +0000 (0:00:00.355)       0:00:16.774 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact build bluestore_wal_devices from resolved symlinks] ***",
                                "Tuesday 24 February 2026  00:59:03 +0000 (0:00:00.331)       0:00:17.106 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Collect existed devices] ************************************",
                                "Tuesday 24 February 2026  00:59:03 +0000 (0:00:00.438)       0:00:17.544 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'dm-0', 'value': {'holders': [], 'host': '', 'links': {'ids': ['dm-name-ceph--9b92f24c--dac1--5bff--a390--34d9ff9e0f4f-osd--block--9b92f24c--dac1--5bff--a390--34d9ff9e0f4f', 'dm-uuid-LVM-h1BZIKSwq0ZvfBWtXhB9IpbOvYH7db3oH2r78k3xPmfToT2WBghjfHVanEend8Zd'], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': '', 'sectors': 41934848, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'dm-1', 'value': {'holders': [], 'host': '', 'links': {'ids': ['dm-name-ceph--7f6a8fe5--6adc--52b4--b256--edcdbfd9bf4e-osd--block--7f6a8fe5--6adc--52b4--b256--edcdbfd9bf4e', 'dm-uuid-LVM-xRh1zTNtAhB8OweVwVFtzaksaCJoA7fP27eNW4gBNJ6VNIfew4kSrXXLjM4ybDHa'], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': '', 'sectors': 41934848, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'loop0', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'loop1', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'loop2', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'loop3', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'loop4', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'loop5', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'loop6', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'loop7', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'sda', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_f90a4e80-1f36-4d14-b7da-de7b20e9fbca', 'scsi-SQEMU_QEMU_HARDDISK_f90a4e80-1f36-4d14-b7da-de7b20e9fbca'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {'sda1': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_f90a4e80-1f36-4d14-b7da-de7b20e9fbca-part1', 'scsi-SQEMU_QEMU_HARDDISK_f90a4e80-1f36-4d14-b7da-de7b20e9fbca-part1'], 'labels': ['cloudimg-rootfs'], 'masters': [], 'uuids': ['b852d8d2-8460-44aa-8998-23e4f04d73cf']}, 'sectors': 165672927, 'sectorsize': 512, 'size': '79.00 GB', 'start': '2099200', 'uuid': 'b852d8d2-8460-44aa-8998-23e4f04d73cf'}, 'sda14': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_f90a4e80-1f36-4d14-b7da-de7b20e9fbca-part14', 'scsi-SQEMU_QEMU_HARDDISK_f90a4e80-1f36-4d14-b7da-de7b20e9fbca-part14'], 'labels': [], 'masters': [], 'uuids': []}, 'sectors': 8192, 'sectorsize': 512, 'size': '4.00 MB', 'start': '2048', 'uuid': None}, 'sda15': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_f90a4e80-1f36-4d14-b7da-de7b20e9fbca-part15', 'scsi-SQEMU_QEMU_HARDDISK_f90a4e80-1f36-4d14-b7da-de7b20e9fbca-part15'], 'labels': ['UEFI'], 'masters': [], 'uuids': ['5C78-612A']}, 'sectors': 217088, 'sectorsize': 512, 'size': '106.00 MB', 'start': '10240', 'uuid': '5C78-612A'}, 'sda16': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_f90a4e80-1f36-4d14-b7da-de7b20e9fbca-part16', 'scsi-SQEMU_QEMU_HARDDISK_f90a4e80-1f36-4d14-b7da-de7b20e9fbca-part16'], 'labels': ['BOOT'], 'masters': [], 'uuids': ['09d53dc1-1e03-4286-bbb8-2b1796cf92ec']}, 'sectors': 1869825, 'sectorsize': 512, 'size': '913.00 MB', 'start': '227328', 'uuid': '09d53dc1-1e03-4286-bbb8-2b1796cf92ec'}}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 167772160, 'sectorsize': '512', 'size': '80.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'sdb', 'value': {'holders': ['ceph--9b92f24c--dac1--5bff--a390--34d9ff9e0f4f-osd--block--9b92f24c--dac1--5bff--a390--34d9ff9e0f4f'], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['lvm-pv-uuid-mec8nJ-xUip-qiWP-sj8w-uf5B-1Lt0-v8yfvH', 'scsi-0QEMU_QEMU_HARDDISK_9c6b6763-9704-41ec-9b6d-97eb80a5145c', 'scsi-SQEMU_QEMU_HARDDISK_9c6b6763-9704-41ec-9b6d-97eb80a5145c'], 'labels': [], 'masters': ['dm-0'], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'sdc', 'value': {'holders': ['ceph--7f6a8fe5--6adc--52b4--b256--edcdbfd9bf4e-osd--block--7f6a8fe5--6adc--52b4--b256--edcdbfd9bf4e'], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['lvm-pv-uuid-HdfFdZ-81SJ-0Umt-qAzb-Dao2-7tUe-tHuhnG', 'scsi-0QEMU_QEMU_HARDDISK_6644f191-69e9-4f3a-baff-cffe4152595b', 'scsi-SQEMU_QEMU_HARDDISK_6644f191-69e9-4f3a-baff-cffe4152595b'], 'labels': [], 'masters': ['dm-1'], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'sdd', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_4534650b-23a0-46c8-8bf1-44520caee637', 'scsi-SQEMU_QEMU_HARDDISK_4534650b-23a0-46c8-8bf1-44520caee637'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'dm-0', 'value': {'holders': [], 'host': '', 'links': {'ids': ['dm-name-ceph--294dd48f--9d6d--5079--8214--59e60af9e526-osd--block--294dd48f--9d6d--5079--8214--59e60af9e526', 'dm-uuid-LVM-wjENyDIL1ovSEUDphRiOfALWwweDrSZWE8niuN9gyF3lbbybFqRYsEaNkOnZAeUN'], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': '', 'sectors': 41934848, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'sr0', 'value': {'holders': [], 'host': 'IDE interface: Intel Corporation 82371SB PIIX3 IDE [Natoma/Triton II]', 'links': {'ids': ['ata-QEMU_DVD-ROM_QM00001'], 'labels': ['config-2'], 'masters': [], 'uuids': ['2026-02-24-00-01-58-00']}, 'model': 'QEMU DVD-ROM', 'partitions': {}, 'removable': '1', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'mq-deadline', 'sectors': 253, 'sectorsize': '2048', 'size': '506.00 KB', 'support_discard': '0', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'dm-1', 'value': {'holders': [], 'host': '', 'links': {'ids': ['dm-name-ceph--7fbf4d63--7c6f--5c70--887c--2b73c978775b-osd--block--7fbf4d63--7c6f--5c70--887c--2b73c978775b', 'dm-uuid-LVM-aRcKlwv7ZYsrJYaz5OO903mOAEBnwLXI61utfRQTUZQKHwm4ceJ3NApz8UrwnN8y'], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': '', 'sectors': 41934848, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'loop0', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'loop1', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'loop2', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'loop3', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'loop4', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'loop5', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'loop6', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'loop7', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'sda', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_47e600ec-05ca-44a6-93d0-5ca3104e3639', 'scsi-SQEMU_QEMU_HARDDISK_47e600ec-05ca-44a6-93d0-5ca3104e3639'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {'sda1': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_47e600ec-05ca-44a6-93d0-5ca3104e3639-part1', 'scsi-SQEMU_QEMU_HARDDISK_47e600ec-05ca-44a6-93d0-5ca3104e3639-part1'], 'labels': ['cloudimg-rootfs'], 'masters': [], 'uuids': ['b852d8d2-8460-44aa-8998-23e4f04d73cf']}, 'sectors': 165672927, 'sectorsize': 512, 'size': '79.00 GB', 'start': '2099200', 'uuid': 'b852d8d2-8460-44aa-8998-23e4f04d73cf'}, 'sda14': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_47e600ec-05ca-44a6-93d0-5ca3104e3639-part14', 'scsi-SQEMU_QEMU_HARDDISK_47e600ec-05ca-44a6-93d0-5ca3104e3639-part14'], 'labels': [], 'masters': [], 'uuids': []}, 'sectors': 8192, 'sectorsize': 512, 'size': '4.00 MB', 'start': '2048', 'uuid': None}, 'sda15': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_47e600ec-05ca-44a6-93d0-5ca3104e3639-part15', 'scsi-SQEMU_QEMU_HARDDISK_47e600ec-05ca-44a6-93d0-5ca3104e3639-part15'], 'labels': ['UEFI'], 'masters': [], 'uuids': ['5C78-612A']}, 'sectors': 217088, 'sectorsize': 512, 'size': '106.00 MB', 'start': '10240', 'uuid': '5C78-612A'}, 'sda16': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_47e600ec-05ca-44a6-93d0-5ca3104e3639-part16', 'scsi-SQEMU_QEMU_HARDDISK_47e600ec-05ca-44a6-93d0-5ca3104e3639-part16'], 'labels': ['BOOT'], 'masters': [], 'uuids': ['09d53dc1-1e03-4286-bbb8-2b1796cf92ec']}, 'sectors': 1869825, 'sectorsize': 512, 'size': '913.00 MB', 'start': '227328', 'uuid': '09d53dc1-1e03-4286-bbb8-2b1796cf92ec'}}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 167772160, 'sectorsize': '512', 'size': '80.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'sdb', 'value': {'holders': ['ceph--294dd48f--9d6d--5079--8214--59e60af9e526-osd--block--294dd48f--9d6d--5079--8214--59e60af9e526'], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['lvm-pv-uuid-1V6jIs-3p8q-l9Rl-OJ2n-8L4P-5xCz-K0YqYS', 'scsi-0QEMU_QEMU_HARDDISK_f659b61f-c39f-426f-a640-6a753951661d', 'scsi-SQEMU_QEMU_HARDDISK_f659b61f-c39f-426f-a640-6a753951661d'], 'labels': [], 'masters': ['dm-0'], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'sdc', 'value': {'holders': ['ceph--7fbf4d63--7c6f--5c70--887c--2b73c978775b-osd--block--7fbf4d63--7c6f--5c70--887c--2b73c978775b'], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['lvm-pv-uuid-aHCB2h-b0pT-WCWd-loOK-A09H-yFYp-FH2wIv', 'scsi-0QEMU_QEMU_HARDDISK_fb58bf16-a1c6-4abb-925b-29c6a6a6f100', 'scsi-SQEMU_QEMU_HARDDISK_fb58bf16-a1c6-4abb-925b-29c6a6a6f100'], 'labels': [], 'masters': ['dm-1'], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'sdd', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_5660bb62-e2c9-45b2-ae11-8ab0f829b194', 'scsi-SQEMU_QEMU_HARDDISK_5660bb62-e2c9-45b2-ae11-8ab0f829b194'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'sr0', 'value': {'holders': [], 'host': 'IDE interface: Intel Corporation 82371SB PIIX3 IDE [Natoma/Triton II]', 'links': {'ids': ['ata-QEMU_DVD-ROM_QM00001'], 'labels': ['config-2'], 'masters': [], 'uuids': ['2026-02-24-00-01-59-00']}, 'model': 'QEMU DVD-ROM', 'partitions': {}, 'removable': '1', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'mq-deadline', 'sectors': 253, 'sectorsize': '2048', 'size': '506.00 KB', 'support_discard': '0', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'dm-0', 'value': {'holders': [], 'host': '', 'links': {'ids': ['dm-name-ceph--752a9ec4--4bfc--5551--82cd--92b802bdad97-osd--block--752a9ec4--4bfc--5551--82cd--92b802bdad97', 'dm-uuid-LVM-jTt9fg6XlhiXbWhW0I9T02tzYsIeoKIoaf5BxtnvIAUk6EWNKNJMqSjVsZqTXHAa'], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': '', 'sectors': 41934848, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'dm-1', 'value': {'holders': [], 'host': '', 'links': {'ids': ['dm-name-ceph--ba85a0e3--2f57--51d5--a9f1--072f0e9a8deb-osd--block--ba85a0e3--2f57--51d5--a9f1--072f0e9a8deb', 'dm-uuid-LVM-YjmBzAWpiQhdCbdayUM0w3sqNqanvbLDieybFDy5wWcdDcB5LPRFHjaRE0WCOjBW'], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational'\u001b[32m2026-02-24 01:01:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                ": '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': '', 'sectors': 41934848, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'loop0', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'loop1', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'loop2', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'loop3', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'loop4', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'loop5', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'loop6', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'loop7', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'sda', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_7a9487cb-ec9e-4508-9c5a-f692700fc5a4', 'scsi-SQEMU_QEMU_HARDDISK_7a9487cb-ec9e-4508-9c5a-f692700fc5a4'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {'sda1': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_7a9487cb-ec9e-4508-9c5a-f692700fc5a4-part1', 'scsi-SQEMU_QEMU_HARDDISK_7a9487cb-ec9e-4508-9c5a-f692700fc5a4-part1'], 'labels': ['cloudimg-rootfs'], 'masters': [], 'uuids': ['b852d8d2-8460-44aa-8998-23e4f04d73cf']}, 'sectors': 165672927, 'sectorsize': 512, 'size': '79.00 GB', 'start': '2099200', 'uuid': 'b852d8d2-8460-44aa-8998-23e4f04d73cf'}, 'sda14': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_7a9487cb-ec9e-4508-9c5a-f692700fc5a4-part14', 'scsi-SQEMU_QEMU_HARDDISK_7a9487cb-ec9e-4508-9c5a-f692700fc5a4-part14'], 'labels': [], 'masters': [], 'uuids': []}, 'sectors': 8192, 'sectorsize': 512, 'size': '4.00 MB', 'start': '2048', 'uuid': None}, 'sda15': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_7a9487cb-ec9e-4508-9c5a-f692700fc5a4-part15', 'scsi-SQEMU_QEMU_HARDDISK_7a9487cb-ec9e-4508-9c5a-f692700fc5a4-part15'], 'labels': ['UEFI'], 'masters': [], 'uuids': ['5C78-612A']}, 'sectors': 217088, 'sectorsize': 512, 'size': '106.00 MB', 'start': '10240', 'uuid': '5C78-612A'}, 'sda16': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_7a9487cb-ec9e-4508-9c5a-f692700fc5a4-part16', 'scsi-SQEMU_QEMU_HARDDISK_7a9487cb-ec9e-4508-9c5a-f692700fc5a4-part16'], 'labels': ['BOOT'], 'masters': [], 'uuids': ['09d53dc1-1e03-4286-bbb8-2b1796cf92ec']}, 'sectors': 1869825, 'sectorsize': 512, 'size': '913.00 MB', 'start': '227328', 'uuid': '09d53dc1-1e03-4286-bbb8-2b1796cf92ec'}}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 167772160, 'sectorsize': '512', 'size': '80.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'sdb', 'value': {'holders': ['ceph--752a9ec4--4bfc--5551--82cd--92b802bdad97-osd--block--752a9ec4--4bfc--5551--82cd--92b802bdad97'], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['lvm-pv-uuid-vfwhPS-pgoq-FNQw-W72Z-epjC-Ziiv-qWNph9', 'scsi-0QEMU_QEMU_HARDDISK_1558935b-4c66-4614-97e5-cda14abd2641', 'scsi-SQEMU_QEMU_HARDDISK_1558935b-4c66-4614-97e5-cda14abd2641'], 'labels': [], 'masters': ['dm-0'], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'sdc', 'value': {'holders': ['ceph--ba85a0e3--2f57--51d5--a9f1--072f0e9a8deb-osd--block--ba85a0e3--2f57--51d5--a9f1--072f0e9a8deb'], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['lvm-pv-uuid-XuBysF-MTWF-hP8t-3CGS-0PEc-QL0q-u9CYq2', 'scsi-0QEMU_QEMU_HARDDISK_6f4643f0-329b-4ae2-97c7-d348648fd970', 'scsi-SQEMU_QEMU_HARDDISK_6f4643f0-329b-4ae2-97c7-d348648fd970'], 'labels': [], 'masters': ['dm-1'], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'sdd', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_5f47789f-2a1a-442a-8a3c-644a4f92e5a8', 'scsi-SQEMU_QEMU_HARDDISK_5f47789f-2a1a-442a-8a3c-644a4f92e5a8'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'sr0', 'value': {'holders': [], 'host': 'IDE interface: Intel Corporation 82371SB PIIX3 IDE [Natoma/Triton II]', 'links': {'ids': ['ata-QEMU_DVD-ROM_QM00001'], 'labels': ['config-2'], 'masters': [], 'uuids': ['2026-02-24-00-02-03-00']}, 'model': 'QEMU DVD-ROM', 'partitions': {}, 'removable': '1', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'mq-deadline', 'sectors': 253, 'sectorsize': '2048', 'size': '506.00 KB', 'support_discard': '0', 'vendor': 'QEMU', 'virtual': 1}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact devices generate device list when osd_auto_discovery] ***",
                                "Tuesday 24 February 2026  00:59:04 +0000 (0:00:00.557)       0:00:18.102 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'dm-0', 'value': {'holders': [], 'host': '', 'links': {'ids': ['dm-name-ceph--9b92f24c--dac1--5bff--a390--34d9ff9e0f4f-osd--block--9b92f24c--dac1--5bff--a390--34d9ff9e0f4f', 'dm-uuid-LVM-h1BZIKSwq0ZvfBWtXhB9IpbOvYH7db3oH2r78k3xPmfToT2WBghjfHVanEend8Zd'], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': '', 'sectors': 41934848, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'dm-1', 'value': {'holders': [], 'host': '', 'links': {'ids': ['dm-name-ceph--7f6a8fe5--6adc--52b4--b256--edcdbfd9bf4e-osd--block--7f6a8fe5--6adc--52b4--b256--edcdbfd9bf4e', 'dm-uuid-LVM-xRh1zTNtAhB8OweVwVFtzaksaCJoA7fP27eNW4gBNJ6VNIfew4kSrXXLjM4ybDHa'], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': '', 'sectors': 41934848, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop0', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop1', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop2', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop3', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop4', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop5', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop6', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'dm-0', 'value': {'holders': [], 'host': '', 'links': {'ids': ['dm-name-ceph--294dd48f--9d6d--5079--8214--59e60af9e526-osd--block--294dd48f--9d6d--5079--8214--59e60af9e526', 'dm-uuid-LVM-wjENyDIL1ovSEUDphRiOfALWwweDrSZWE8niuN9gyF3lbbybFqRYsEaNkOnZAeUN'], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': '', 'sectors': 41934848, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop7', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'dm-1', 'value': {'holders': [], 'host': '', 'links': {'ids': ['dm-name-ceph--7fbf4d63--7c6f--5c70--887c--2b73c978775b-osd--block--7fbf4d63--7c6f--5c70--887c--2b73c978775b', 'dm-uuid-LVM-aRcKlwv7ZYsrJYaz5OO903mOAEBnwLXI61utfRQTUZQKHwm4ceJ3NApz8UrwnN8y'], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': '', 'sectors': 41934848, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sda', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_f90a4e80-1f36-4d14-b7da-de7b20e9fbca', 'scsi-SQEMU_QEMU_HARDDISK_f90a4e80-1f36-4d14-b7da-de7b20e9fbca'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {'sda1': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_f90a4e80-1f36-4d14-b7da-de7b20e9fbca-part1', 'scsi-SQEMU_QEMU_HARDDISK_f90a4e80-1f36-4d14-b7da-de7b20e9fbca-part1'], 'labels': ['cloudimg-rootfs'], 'masters': [], 'uuids': ['b852d8d2-8460-44aa-8998-23e4f04d73cf']}, 'sectors': 165672927, 'sectorsize': 512, 'size': '79.00 GB', 'start': '2099200', 'uuid': 'b852d8d2-8460-44aa-8998-23e4f04d73cf'}, 'sda14': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_f90a4e80-1f36-4d14-b7da-de7b20e9fbca-part14', 'scsi-SQEMU_QEMU_HARDDISK_f90a4e80-1f36-4d14-b7da-de7b20e9fbca-part14'], 'labels': [], 'masters': [], 'uuids': []}, 'sectors': 8192, 'sectorsize': 512, 'size': '4.00 MB', 'start': '2048', 'uuid': None}, 'sda15': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_f90a4e80-1f36-4d14-b7da-de7b20e9fbca-part15', 'scsi-SQEMU_QEMU_HARDDISK_f90a4e80-1f36-4d14-b7da-de7b20e9fbca-part15'], 'labels': ['UEFI'], 'masters': [], 'uuids': ['5C78-612A']}, 'sectors': 217088, 'sectorsize': 512, 'size': '106.00 MB', 'start': '10240', 'uuid': '5C78-612A'}, 'sda16': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_f90a4e80-1f36-4d14-b7da-de7b20e9fbca-part16', 'scsi-SQEMU_QEMU_HARDDISK_f90a4e80-1f36-4d14-b7da-de7b20e9fbca-part16'], 'labels': ['BOOT'], 'masters': [], 'uuids': ['09d53dc1-1e03-4286-bbb8-2b1796cf92ec']}, 'sectors': 1869825, 'sectorsize': 512, 'size': '913.00 MB', 'start': '227328', 'uuid': '09d53dc1-1e03-4286-bbb8-2b1796cf92ec'}}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 167772160, 'sectorsize': '512', 'size': '80.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop0', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sdb', 'value': {'holders': ['ceph--9b92f24c--dac1--5bff--a390--34d9ff9e0f4f-osd--block--9b92f24c--dac1--5bff--a390--34d9ff9e0f4f'], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['lvm-pv-uuid-mec8nJ-xUip-qiWP-sj8w-uf5B-1Lt0-v8yfvH', 'scsi-0QEMU_QEMU_HARDDISK_9c6b6763-9704-41ec-9b6d-97eb80a5145c', 'scsi-SQEMU_QEMU_HARDDISK_9c6b6763-9704-41ec-9b6d-97eb80a5145c'], 'labels': [], 'masters': ['dm-0'], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop1', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sdc', 'value': {'holders': ['ceph--7f6a8fe5--6adc--52b4--b256--edcdbfd9bf4e-osd--block--7f6a8fe5--6adc--52b4--b256--edcdbfd9bf4e'], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['lvm-pv-uuid-HdfFdZ-81SJ-0Umt-qAzb-Dao2-7tUe-tHuhnG', 'scsi-0QEMU_QEMU_HARDDISK_6644f191-69e9-4f3a-baff-cffe4152595b', 'scsi-SQEMU_QEMU_HARDDISK_6644f191-69e9-4f3a-baff-cffe4152595b'], 'labels': [], 'masters': ['dm-1'], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sdd', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_4534650b-23a0-46c8-8bf1-44520caee637', 'scsi-SQEMU_QEMU_HARDDISK_4534650b-23a0-46c8-8bf1-44520caee637'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop2', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sr0', 'value': {'holders': [], 'host': 'IDE interface: Intel Corporation 82371SB PIIX3 IDE [Natoma/Triton II]', 'links': {'ids': ['ata-QEMU_DVD-ROM_QM00001'], 'labels': ['config-2'], 'masters': [], 'uuids': ['2026-02-24-00-01-58-00']}, 'model': 'QEMU DVD-ROM', 'partitions': {}, 'removable': '1', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'mq-deadline', 'sectors': 253, 'sectorsize': '2048', 'size': '506.00 KB', 'support_discard': '0', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop3', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop4', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop5', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'dm-0', 'value': {'holders': [], 'host': '', 'links': {'ids': ['dm-name-ceph--752a9ec4--4bfc--5551--82cd--92b802bdad97-osd--block--752a9ec4--4bfc--5551--82cd--92b802bdad97', 'dm-uuid-LVM-jTt9fg6XlhiXbWhW0I9T02tzYsIeoKIoaf5BxtnvIAUk6EWNKNJMqSjVsZqTXHAa'], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': '', 'sectors': 41934848, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop6', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'dm-1', 'value': {'holders': [], 'host': '', 'links': {'ids': ['dm-name-ceph--ba85a0e3--2f57--51d5--a9f1--072f0e9a8deb-osd--block--ba85a0e3--2f57--51d5--a9f1--072f0e9a8deb', 'dm-uuid-LVM-YjmBzAWpiQhdCbdayUM0w3sqNqanvbLDieybFDy5wWcdDcB5LPRFHjaRE0WCOjBW'], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': '', 'sectors': 41934848, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop7', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop0', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sda', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_47e600ec-05ca-44a6-93d0-5ca3104e3639', 'scsi-SQEMU_QEMU_HARDDISK_47e600ec-05ca-44a6-93d0-5ca3104e3639'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {'sda1': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_47e600ec-05ca-44a6-93d0-5ca3104e3639-part1', 'scsi-SQEMU_QEMU_HARDDISK_47e600ec-05ca-44a6-93d0-5ca3104e3639-part1'], 'labels': ['cloudimg-rootfs'], 'masters': [], 'uuids': ['b852d8d2-8460-44aa-8998-23e4f04d73cf']}, 'sectors': 165672927, 'sectorsize': 512, 'size': '79.00 GB', 'start': '2099200', 'uuid': 'b852d8d2-8460-44aa-8998-23e4f04d73cf'}, 'sda14': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_47e600ec-05ca-44a6-93d0-5ca3104e3639-part14', 'scsi-SQEMU_QEMU_HARDDISK_47e600ec-05ca-44a6-93d0-5ca3104e3639-part14'], 'labels': [], 'masters': [], 'uuids': []}, 'sectors': 8192, 'sectorsize': 512, 'size': '4.00 MB', 'start': '2048', 'uuid': None}, 'sda15': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_47e600ec-05ca-44a6-93d0-5ca3104e3639-part15', 'scsi-SQEMU_QEMU_HARDDISK_47e600ec-05ca-44a6-93d0-5ca3104e3639-part15'], 'labels': ['UEFI'], 'masters': [], 'uuids': ['5C78-612A']}, 'sectors': 217088, 'sectorsize': 512, 'size': '106.00 MB', 'start': '10240', 'uuid': '5C78-612A'}, 'sda16': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_47e600ec-05ca-44a6-93d0-5ca3104e3639-part16', 'scsi-SQEMU_QEMU_HARDDISK_47e600ec-05ca-44a6-93d0-5ca3104e3639-part16'], 'labels': ['BOOT'], 'masters': [], 'uuids': ['09d53dc1-1e03-4286-bbb8-2b1796cf92ec']}, 'sectors': 1869825, 'sectorsize': 512, 'size': '913.00 MB', 'start': '227328', 'uuid': '09d53dc1-1e03-4286-bbb8-2b1796cf92ec'}}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 167772160, 'sectorsize': '512', 'size': '80.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop1', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sdb', 'value': {'holders': ['ceph--294dd48f--9d6d--5079--8214--59e60af9e526-osd--block--294dd48f--9d6d--5079--8214--59e60af9e526'], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['lvm-pv-uuid-1V6jIs-3p8q-l9Rl-OJ2n-8L4P-5xCz-K0YqYS', 'scsi-0QEMU_QEMU_HARDDISK_f659b61f-c39f-426f-a640-6a753951661d', 'scsi-SQEMU_QEMU_HARDDISK_f659b61f-c39f-426f-a640-6a753951661d'], 'labels': [], 'masters': ['dm-0'], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop2', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sdc', 'value': {'holders': ['ceph--7fbf4d63--7c6f--5c70--887c--2b73c978775b-osd--block--7fbf4d63--7c6f--5c70--887c--2b73c978775b'], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['lvm-pv-uuid-aHCB2h-b0pT-WCWd-loOK-A09H-yFYp-FH2wIv', 'scsi-0QEMU_QEMU_HARDDISK_fb58bf16-a1c6-4abb-925b-29c6a6a6f100', 'scsi-SQEMU_QEMU_HARDDISK_fb58bf16-a1c6-4abb-925b-29c6a6a6f100'], 'labels': [], 'masters': ['dm-1'], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop3', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop4', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sdd', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_5660bb62-e2c9-45b2-ae11-8ab0f829b194', 'scsi-SQEMU_QEMU_HARDDISK_5660bb62-e2c9-45b2-ae11-8ab0f829b194'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop5', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sr0', 'value': {'holders': [], 'host': 'IDE interface: Intel Corporation 82371SB PIIX3 IDE [Natoma/Triton II]', 'links': {'ids': ['ata-QEMU_DVD-ROM_QM00001'], 'labels': ['config-2'], 'masters': [], 'uuids': ['2026-02-24-00-01-59-00']}, 'model': 'QEMU DVD-ROM', 'partitions': {}, 'removable': '1', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'mq-deadline', 'sectors': 253, 'sectorsize': '2048', 'size': '506.00 KB', 'support_discard': '0', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop6', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'loop7', 'value': {'holders': [], 'host': '', 'links': {'ids': [], 'labels': [], 'masters': [], 'uuids': []}, 'model': None, 'partitions': {}, 'removable': '0', 'rotational': '0', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 0, 'sectorsize': '512', 'size': '0.00 Bytes', 'support_discard': '0', 'vendor': None, 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sda', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_7a9487cb-ec9e-4508-9c5a-f692700fc5a4', 'scsi-SQEMU_QEMU_HARDDISK_7a9487cb-ec9e-4508-9c5a-f692700fc5a4'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {'sda1': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_7a9487cb-ec9e-4508-9c5a-f692700fc5a4-part1', 'scsi-SQEMU_QEMU_HARDDISK_7a9487cb-ec9e-4508-9c5a-f692700fc5a4-part1'], 'labels': ['cloudimg-rootfs'], 'masters': [], 'uuids': ['b852d8d2-8460-44aa-8998-23e4f04d73cf']}, 'sectors': 165672927, 'sectorsize': 512, 'size': '79.00 GB', 'start': '2099200', 'uuid': 'b852d8d2-8460-44aa-8998-23e4f04d73cf'}, 'sda14': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_7a9487cb-ec9e-4508-9c5a-f692700fc5a4-part14', 'scsi-SQEMU_QEMU_HARDDISK_7a9487cb-ec9e-4508-9c5a-f692700fc5a4-part14'], 'labels': [], 'masters': [], 'uuids': []}, 'sectors': 8192, 'sectorsize': 512, 'size': '4.00 MB', 'start': '2048', 'uuid': None}, 'sda15': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_7a9487cb-ec9e-4508-9c5a-f692700fc5a4-part15', 'scsi-SQEMU_QEMU_HARDDISK_7a9487cb-ec9e-4508-9c5a-f692700fc5a4-part15'], 'labels': ['UEFI'], 'masters': [], 'uuids': ['5C78-612A']}, 'sectors': 217088, 'sectorsize': 512, 'size': '106.00 MB', 'start': '10240', 'uuid': '5C78-612A'}, 'sda16': {'holders': [], 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_7a9487cb-ec9e-4508-9c5a-f692700fc5a4-part16', 'scsi-SQEMU_QEMU_HARDDISK_7a9487cb-ec9e-4508-9c5a-f692700fc5a4-part16'], 'labels': ['BOOT'], 'masters': [], 'uuids': ['09d53dc1-1e03-4286-bbb8-2b1796cf92ec']}, 'sectors': 1869825, 'sectorsize': 512, 'size': '913.00 MB', 'start': '227328', 'uuid': '09d53dc1-1e03-4286-bbb8-2b1796cf92ec'}}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 167772160, 'sectorsize': '512', 'size': '80.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sdb', 'value': {'holders': ['ceph--752a9ec4--4bfc--5551--82cd--92b802bdad97-osd--block--752a9ec4--4bfc--5551--82cd--92b802bdad97'], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['lvm-pv-uuid-vfwhPS-pgoq-FNQw-W72Z-epjC-Ziiv-qWNph9', 'scsi-0QEMU_QEMU_HARDDISK_1558935b-4c66-4614-97e5-cda14abd2641', 'scsi-SQEMU_QEMU_HARDDISK_1558935b-4c66-4614-97e5-cda14abd2641'], 'labels': [], 'masters': ['dm-0'], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sdc', 'value': {'holders': ['ceph--ba85a0e3--2f57--51d5--a9f1--072f0e9a8deb-osd--block--ba85a0e3--2f57--51d5--a9f1--072f0e9a8deb'], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['lvm-pv-uuid-XuBysF-MTWF-hP8t-3CGS-0PEc-QL0q-u9CYq2', 'scsi-0QEMU_QEMU_HARDDISK_6f4643f0-329b-4ae2-97c7-d348648fd970', 'scsi-SQEMU_QEMU_HARDDISK_6f4643f0-329b-4ae2-97c7-d348648fd970'], 'labels': [], 'masters': ['dm-1'], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sdd', 'value': {'holders': [], 'host': 'SCSI storage controller: Red Hat, Inc. Virtio SCSI', 'links': {'ids': ['scsi-0QEMU_QEMU_HARDDISK_5f47789f-2a1a-442a-8a3c-644a4f92e5a8', 'scsi-SQEMU_QEMU_HARDDISK_5f47789f-2a1a-442a-8a3c-644a4f92e5a8'], 'labels': [], 'masters': [], 'uuids': []}, 'model': 'QEMU HARDDISK', 'partitions': {}, 'removable': '0', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'none', 'sectors': 41943040, 'sectorsize': '512', 'size': '20.00 GB', 'support_discard': '4096', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'changed': False, 'skipped': True, 'skip_reason': 'Conditional result was False', 'false_condition': 'osd_auto_discovery | default(False) | bool', 'item': {'key': 'sr0', 'value': {'holders': [], 'host': 'IDE interface: Intel Corporation 82371SB PIIX3 IDE [Natoma/Triton II]', 'links': {'ids': ['ata-QEMU_DVD-ROM_QM00001'], 'labels': ['config-2'], 'masters': [], 'uuids': ['2026-02-24-00-02-03-00']}, 'model': 'QEMU DVD-ROM', 'partitions': {}, 'removable': '1', 'rotational': '1', 'sas_address': None, 'sas_device_handle': None, 'scheduler_mode': 'mq-deadline', 'sectors': 253, 'sectorsize': '2048', 'size': '506.00 KB', 'support_discard': '0', 'vendor': 'QEMU', 'virtual': 1}}, 'ansible_loop_var': 'item'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Check if the ceph conf exists] ******************************",
                                "Tuesday 24 February 2026  00:59:04 +0000 (0:00:00.571)       0:00:18.674 ****** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set default osd_pool_default_crush_rule fact] ***************",
                                "Tuesday 24 February 2026  00:59:05 +0000 (0:00:00.756)       0:00:19.430 ****** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Read osd pool default crush rule] ***************************",
                                "Tuesday 24 February 2026  00:59:05 +0000 (0:00:00.436)       0:00:19.866 ****** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set osd_pool_default_crush_rule fact] ***********************",
                                "Tuesday 24 February 2026  00:59:06 +0000 (0:00:00.652)       0:00:20.518 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Read osd pool default crush rule] ***************************",
                                "Tuesday 24 February 2026  00:59:06 +0000 (0:00:00.293)       0:00:20.812 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set osd_pool_default_crush_rule fact] ***********************",
                                "Tuesday 24 February 2026  00:59:07 +0000 (0:00:00.391)       0:00:21.204 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _monitor_addresses - ipv4] *************************",
                                "Tuesday 24 February 2026  00:59:07 +0000 (0:00:00.490)       0:00:21.694 ****** ",
                                "\u001b[0;32mok: [testbed-node-3] => (item=testbed-node-0)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item=testbed-node-0)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item=testbed-node-1)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item=testbed-node-0)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item=testbed-node-1)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item=testbed-node-2)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item=testbed-node-1)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item=testbed-node-2)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item=testbed-node-2)\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _monitor_addresses - ipv6] *************************",
                                "Tuesday 24 February 2026  00:59:08 +0000 (0:00:00.854)       0:00:22.549 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-0) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-1) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=testbed-node-0) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=testbed-node-1) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=testbed-node-2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=testbed-node-0) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=testbed-node-1) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=testbed-node-2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Import_tasks set_radosgw_address.yml] ***********************",
                                "Tuesday 24 February 2026  00:59:08 +0000 (0:00:00.378)       0:00:22.927 ****** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-facts/tasks/set_radosgw_address.yml for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set current radosgw_address_block, radosgw_address, radosgw_interface  from node \"{{ ceph_dashboard_call_item }}\"] ***",
                                "Tuesday 24 February 2026  00:59:09 +0000 (0:00:00.672)       0:00:23.599 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _radosgw_address to radosgw_address_block ipv4] ****",
                                "Tuesday 24 February 2026  00:59:09 +0000 (0:00:00.322)       0:00:23.922 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _radosgw_address to radosgw_address_block ipv6] ****",
                                "Tuesday 24 February 2026  00:59:10 +0000 (0:00:00.365)       0:00:24.287 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _radosgw_address to radosgw_address] ***************",
                                "Tuesday 24 February 2026  00:59:10 +0000 (0:00:00.336)       0:00:24.623 ****** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _interface] ****************************************",
                                "Tuesday 24 February 2026  00:59:11 +0000 (0:00:00.845)       0:00:25.468 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-3) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-4) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-5) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _radosgw_address to radosgw_interface - ipv4] ******",
                                "Tuesday 24 February 2026  00:59:11 +0000 (0:00:00.375)       0:00:25.844 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-3) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-4) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-5) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact _radosgw_address to radosgw_interface - ipv6] ******",
                                "Tuesday 24 February 2026  00:59:12 +0000 (0:00:00.427)       0:00:26.272 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-3) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-4) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=testbed-node-5) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Reset rgw_instances (workaround)] ***************************",
                                "Tuesday 24 February 2026  00:59:12 +0000 (0:00:00.369)       0:00:26.642 ****** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact rgw_instances] *************************************",
                                "Tuesday 24 February 2026  00:59:13 +0000 (0:00:00.332)       0:00:26.975 ****** ",
                                "\u001b[0;32mok: [testbed-node-3] => (item=0)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item=0)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item=0)\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact ceph_run_cmd] **************************************",
                                "Tuesday 24 February 2026  00:59:13 +0000 (0:00:00.465)       0:00:27.440 ****** ",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item=testbed-node-0)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-1(192.168.16.11)] => (item=testbed-node-1)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-2(192.168.16.12)] => (item=testbed-node-2)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item=testbed-node-3)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-4(192.168.16.14)] => (item=testbed-node-4)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-5(192.168.16.15)] => (item=testbed-node-5)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-manager(192.168.16.5)] => (item=testbed-manager)\u001b[0m",
                                "",
                                "TASK [ceph-facts : Set_fact ceph_admin_command] ********************************",
                                "Tuesday 24 February 2026  00:59:14 +0000 (0:00:01.084)       0:00:28.525 ****** ",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-0(192.168.16.10)] => (item=testbed-node-0)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-1(192.168.16.11)] => (item=testbed-node-1)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-2(192.168.16.12)] => (item=testbed-node-2)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item=testbed-node-3)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-4(192.168.16.14)] => (item=testbed-node-4)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-5(192.168.16.15)] => (item=testbed-node-5)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-manager(192.168.16.5)] => (item=testbed-manager)\u001b[0m",
                                "",
                                "TASK [Include tasks from the ceph-osd role] ************************************",
                                "Tuesday 24 February 2026  00:59:16 +0000 (0:00:02.041)       0:00:30.567 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/tasks/openstack_config.yml for testbed-node-5\u001b[0m",
                                "",
                                "TASK [create openstack pool(s)] ************************************************",
                                "Tuesday 24 February 2026  00:59:16 +0000 (0:00:00.368)       0:00:30.936 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item={'application': 'rbd', 'erasure_profile': '', 'expected_num_objects': '', 'min_size': 0, 'name': 'backups', 'pg_autoscale_mode': False, 'pg_num': 32, 'pgp_num': 32, 'rule_name': 'replicated_rule', 'size': 3, 'type': 1})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item={'application': 'rbd', 'erasure_profile': '', 'expected_num_objects': '', 'min_size': 0, 'name': 'volumes', 'pg_autoscale_mode': False, 'pg_num': 32, 'pgp_num': 32, 'rule_name': 'replicated_rule', 'size': 3, 'type': 1})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item={'application': 'rbd', 'erasure_profile': '', 'expected_num_objects': '', 'min_size': 0, 'name': 'images', 'pg_autoscale_mode': False, 'pg_num': 32, 'pgp_num': 32, 'rule_name': 'replicated_rule', 'size': 3, 'type': 1})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item={'application': 'rbd', 'erasure_profile': '', 'expected_num_objects': '', 'min_size': 0, 'name': 'metrics', 'pg_autoscale_mode': False, 'pg_num': 32, 'pgp_num': 32, 'rule_name': 'replicated_rule', 'size': 3, 'type': 1})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item={'application': 'rbd', 'erasure_profile': '', 'expected_num_objects': '', 'min_size': 0, 'name': 'vms', 'pg_autoscale_mode': False, 'pg_num': 32, 'pgp_num': 32, 'rule_name': 'replicated_rule', 'size': 3, 'type': 1})\u001b[0m",
                                "",
                                "TASK [generate keys] ***********************************************************",
                                "Tuesday 24 February 2026  01:00:05 +0000 (0:00:48.587)       0:01:19.524 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5 -> {{ groups[mon_group_name][0] }}]\u001b[0m",
                                "",
                                "TASK [get keys from monitors] **************************************************",
                                "Tuesday 24 February 2026  01:00:32 +0000 (0:00:27.173)       0:01:46.698 ****** ",
                                "\u001b[0;32mok: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5 -> {{ groups.get(mon_group_name)[0] }}]\u001b[0m",
                                "",
                                "TASK [copy ceph key(s) if needed] **********************************************",
                                "Tuesday 24 February 2026  01:00:46 +0000 (0:00:13.807)       0:02:00.505 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-1(192.168.16.11)] => (item=None)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-2(192.168.16.12)] => (item=None)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-1(192.168.16.11)] => (item=None)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-2(192.168.16.12)] => (item=None)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-1(192.168.16.11)] => (item=None)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-2(192.168.16.12)] => (item=None)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-1(192.168.16.11)] => (item=None)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-2(192.168.16.12)] => (item=None)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-1(192.168.16.11)] => (item=None)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-2(192.168.16.12)] => (item=None)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-0(192.168.16.10)] => (item=None)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-1(192.168.16.11)] => (item=None)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5 -> testbed-node-2(192.168.16.12)] => (item=None)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5 -> {{ item.1 }}]\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;32mtestbed-node-3\u001b[0m             : \u001b[0;32mok=25  \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=28  \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;32mtestbed-node-4\u001b[0m             : \u001b[0;32mok=18  \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=21  \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-5\u001b[0m             : \u001b[0;32mok=23  \u001b[0m \u001b[0;33mchanged=3   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=20  \u001b[0m rescued=0    ignored=0   ",
                                "",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Tuesday 24 February 2026  01:01:05 +0000 (0:00:19.375)       0:02:19.881 ****** ",
                                "=============================================================================== ",
                                "create openstack pool(s) ----------------------------------------------- 48.59s",
                                "generate keys ---------------------------------------------------------- 27.17s",
                                "copy ceph key(s) if needed --------------------------------------------- 19.38s",
                                "get keys from monitors ------------------------------------------------- 13.81s",
                                "ceph-facts : Find a running mon container ------------------------------- 2.32s",
                                "ceph-facts : Set_fact ceph_admin_command -------------------------------- 2.04s",
                                "ceph-facts : Get current fsid if cluster is already running ------------- 1.89s",
                                "ceph-facts : Set_fact ceph_run_cmd -------------------------------------- 1.08s",
                                "ceph-facts : Check if podman binary is present -------------------------- 0.99s",
                                "ceph-facts : Check for a ceph mon socket -------------------------------- 0.90s",
                                "ceph-facts : Check if the ceph mon socket is in-use --------------------- 0.87s",
                                "ceph-facts : Set_fact _monitor_addresses - ipv4 ------------------------- 0.86s",
                                "ceph-facts : Set_fact _radosgw_address to radosgw_address --------------- 0.85s",
                                "ceph-facts : Check if the ceph conf exists ------------------------------ 0.76s",
                                "ceph-facts : Set_fact monitor_name ansible_facts['hostname'] ------------ 0.71s",
                                "ceph-facts : Include facts.yml ------------------------------------------ 0.68s",
                                "ceph-facts : Import_tasks set_radosgw_address.yml ----------------------- 0.67s",
                                "ceph-facts : Read osd pool default crush rule --------------------------- 0.65s",
                                "ceph-facts : Check if it is atomic host --------------------------------- 0.65s",
                                "ceph-facts : Set_fact discovered_interpreter_python if not previously set --- 0.60s",
                                "\u001b[32m2026-02-24 01:01:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c5f0c529-69aa-4268-97af-bde1796abaa2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:01:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c587e61e-d18f-4e39-9404-027cd3697185 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:01:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9cb83792-8f4e-4574-a78c-9b0e654d6fd6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:01:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 01:01:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c5f0c529-69aa-4268-97af-bde1796abaa2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:01:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c587e61e-d18f-4e39-9404-027cd3697185 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:01:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9cb83792-8f4e-4574-a78c-9b0e654d6fd6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:01:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 01:01:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c5f0c529-69aa-4268-97af-bde1796abaa2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:01:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c587e61e-d18f-4e39-9404-027cd3697185 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:01:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9cb83792-8f4e-4574-a78c-9b0e654d6fd6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:01:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 01:01:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c5f0c529-69aa-4268-97af-bde1796abaa2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:01:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c587e61e-d18f-4e39-9404-027cd3697185 is in state SUCCESS\u001b[0m",
                                "",
                                "",
                                "PLAY [Group hosts based on configuration] **************************************",
                                "",
                                "TASK [Group hosts based on Kolla action] ***************************************",
                                "Tuesday 24 February 2026  00:59:34 +0000 (0:00:00.275)       0:00:00.275 ****** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [Group hosts based on enabled services] ***********************************",
                                "Tuesday 24 February 2026  00:59:35 +0000 (0:00:00.303)       0:00:00.578 ****** ",
                                "\u001b[0;32mok: [testbed-node-0] => (item=enable_horizon_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item=enable_horizon_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item=enable_horizon_True)\u001b[0m",
                                "",
                                "PLAY [Apply role horizon] ******************************************************",
                                "",
                                "TASK [horizon : include_tasks] *************************************************",
                                "Tuesday 24 February 2026  00:59:35 +0000 (0:00:00.399)       0:00:00.978 ****** ",
                                "\u001b[0;36mincluded: /ansible/roles/horizon/tasks/deploy.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [horizon : Ensuring config directories exist] *****************************",
                                "Tuesday 24 February 2026  00:59:36 +0000 (0:00:00.503)       0:00:01.481 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'horizon', 'value': {'container_name': 'horizon', 'group': 'horizon', 'enabled': True, 'image': 'registry.osism.tech/kolla/horizon:2024.2', 'environment': {'ENABLE_BLAZAR': 'no', 'ENABLE_CLOUDKITTY': 'no', 'ENABLE_DESIGNATE': 'yes', 'ENABLE_FWAAS': 'no', 'ENABLE_HEAT': 'no', 'ENABLE_IRONIC': 'no', 'ENABLE_MAGNUM': 'yes', 'ENABLE_MANILA': 'yes', 'ENABLE_MASAKARI': 'no', 'ENABLE_MISTRAL': 'no', 'ENABLE_NEUTRON_VPNAAS': 'no', 'ENABLE_OCTAVIA': 'yes', 'ENABLE_TACKER': 'no', 'ENABLE_TROVE': 'no', 'ENABLE_WATCHER': 'no', 'ENABLE_ZUN': 'no', 'FORCE_GENERATE': 'no'}, 'volumes': ['/etc/kolla/horizon/:/var/lib/kolla/config_files/:ro', '', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:80'], 'timeout': '30'}, 'haproxy': {'horizon': {'enabled': True, 'mode': 'http', 'external': False, 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_redirect': {'enabled': True, 'mode': 'redirect', 'external': False, 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'horizon_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_external_redirect': {'enabled': True, 'mode': 'redirect', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'acme_client': {'enabled': True, 'with_frontend': False, 'custom_member_list': []}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'horizon', 'value': {'container_name': 'horizon', 'group': 'horizon', 'enabled': True, 'image': 'registry.osism.tech/kolla/horizon:2024.2', 'environment': {'ENABLE_BLAZAR': 'no', 'ENABLE_CLOUDKITTY': 'no', 'ENABLE_DESIGNATE': 'yes', 'ENABLE_FWAAS': 'no', 'ENABLE_HEAT': 'no', 'ENABLE_IRONIC': 'no', 'ENABLE_MAGNUM': 'yes', 'ENABLE_MANILA': 'yes', 'ENABLE_MASAKARI': 'no', 'ENABLE_MISTRAL': 'no', 'ENABLE_NEUTRON_VPNAAS': 'no', 'ENABLE_OCTAVIA': 'yes', 'ENABLE_TACKER': 'no', 'ENABLE_TROVE': 'no', 'ENABLE_WATCHER': 'no', 'ENABLE_ZUN': 'no', 'FORCE_GENERATE': 'no'}, 'volumes': ['/etc/kolla/horizon/:/var/lib/kolla/config_files/:ro', '', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:80'], 'timeout': '30'}, 'haproxy': {'horizon': {'enabled': True, 'mode': 'http', 'external': False, 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_redirect': {'enabled': True, 'mode': 'redirect', 'external': False, 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'horizon_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_external_redirect': {'enabled': True, 'mode': 'redirect', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'acme_client': {'enabled': True, 'with_frontend': False, 'custom_member_list': []}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'horizon', 'value': {'container_name': 'horizon', 'group': 'horizon', 'enabled': True, 'image': 'registry.osism.tech/kolla/horizon:2024.2', 'environment': {'ENABLE_BLAZAR': 'no', 'ENABLE_CLOUDKITTY': 'no', 'ENABLE_DESIGNATE': 'yes', 'ENABLE_FWAAS': 'no', 'ENABLE_HEAT': 'no', 'ENABLE_IRONIC': 'no', 'ENABLE_MAGNUM': 'yes', 'ENABLE_MANILA': 'yes', 'ENABLE_MASAKARI': 'no', 'ENABLE_MISTRAL': 'no', 'ENABLE_NEUTRON_VPNAAS': 'no', 'ENABLE_OCTAVIA': 'yes', 'ENABLE_TACKER': 'no', 'ENABLE_TROVE': 'no', 'ENABLE_WATCHER': 'no', 'ENABLE_ZUN': 'no', 'FORCE_GENERATE': 'no'}, 'volumes': ['/etc/kolla/horizon/:/var/lib/kolla/config_files/:ro', '', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:80'], 'timeout': '30'}, 'haproxy': {'horizon': {'enabled': True, 'mode': 'http', 'external': False, 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_redirect': {'enabled': True, 'mode': 'redirect', 'external': False, 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'horizon_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_external_redirect': {'enabled': True, 'mode': 'redirect', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'acme_client': {'enabled': True, 'with_frontend': False, 'custom_member_list': []}}}})\u001b[0m",
                                "",
                                "TASK [horizon : Set empty custom policy] ***************************************",
                                "Tuesday 24 February 2026  00:59:37 +0000 (0:00:01.200)       0:00:02.681 ****** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [horizon : include_tasks] *************************************************",
                                "Tuesday 24 February 2026  00:59:37 +0000 (0:00:00.423)       0:00:03.105 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'name': 'cloudkitty', 'enabled': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'name': 'heat', 'enabled': 'no'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'name': 'ironic', 'enabled': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'name': 'masakari', 'enabled': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'name': 'mistral', 'enabled': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'name': 'tacker', 'enabled': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'name': 'trove', 'enabled': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'name': 'watcher', 'enabled': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'name': 'cloudkitty', 'enabled': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'name': 'heat', 'enabled': 'no'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'name': 'ironic', 'enabled': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'name': 'masakari', 'enabled': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'name': 'mistral', 'enabled': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'name': 'tacker', 'enabled': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'name': 'trove', 'enabled': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'name': 'watcher', 'enabled': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'name': 'cloudkitty', 'enabled': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'name': 'heat', 'enabled': 'no'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'name': 'ironic', 'enabled': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'name': 'masakari', 'enabled': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'name': 'mistral', 'enabled': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'name': 'tacker', 'enabled': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'name': 'trove', 'enabled': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'name': 'watcher', 'enabled': False}) \u001b[0m",
                                "\u001b[0;36mincluded: /ansible/roles/horizon/tasks/policy_item.yml for testbed-node-0, testbed-node-1, testbed-node-2 => (item={'name': 'ceilometer', 'enabled': 'yes'})\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/roles/horizon/tasks/policy_item.yml for testbed-node-0, testbed-node-1, testbed-node-2 => (item={'name': 'cinder', 'enabled': 'yes'})\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/roles/horizon/tasks/policy_item.yml for testbed-node-0, testbed-node-1, testbed-node-2 => (item={'name': 'designate', 'enabled': True})\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/roles/horizon/tasks/policy_item.yml for testbed-node-0, testbed-node-1, testbed-node-2 => (item={'name': 'glance', 'enabled': True})\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/roles/horizon/tasks/policy_item.yml for testbed-node-0, testbed-node-1, testbed-node-2 => (item={'name': 'keystone', 'enabled': True})\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/roles/horizon/tasks/policy_item.yml for testbed-node-0, testbed-node-1, testbed-node-2 => (item={'name': 'magnum', 'enabled': True})\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/roles/horizon/tasks/policy_item.yml for testbed-node-0, testbed-node-1, testbed-node-2 => (item={'name': 'manila', 'enabled': True})\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/roles/horizon/tasks/policy_item.yml for testbed-node-0, testbed-node-1, testbed-node-2 => (item={'name': 'neutron', 'enabled': True})\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/roles/horizon/tasks/policy_item.yml for testbed-node-0, testbed-node-1, testbed-node-2 => (item={'name': 'nova', 'enabled': True})\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/roles/horizon/tasks/policy_item.yml for testbed-node-0, testbed-node-1, testbed-node-2 => (item={'name': 'octavia', 'enabled': True})\u001b[0m",
                                "",
                                "TASK [horizon : Update policy file name] ***************************************",
                                "Tuesday 24 February 2026  00:59:38 +0000 (0:00:00.782)       0:00:03.887 ****** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [horizon : Check if policies shall be overwritten] ************************",
                                "Tuesday 24 February 2026  00:59:38 +0000 (0:00:00.328)       0:00:04.215 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [horizon : Update custom policy file name] ********************************",
                                "Tuesday 24 February 2026  00:59:38 +0000 (0:00:00.135)       0:00:04.351 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [horizon : Update policy file name] ***************************************",
                                "Tuesday 24 February 2026  00:59:39 +0000 (0:00:00.463)       0:00:04.814 ****** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [horizon : Check if policies shall be overwritten] ************************",
                                "Tuesday 24 February 2026  00:59:39 +0000 (0:00:00.345)       0:00:05.160 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [horizon : Update custom policy file name] ********************************",
                                "Tuesday 24 February 2026  00:59:39 +0000 (0:00:00.114)       0:00:05.275 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [horizon : Update policy file name] ***************************************",
                                "Tuesday 24 February 2026  00:59:40 +0000 (0:00:00.349)       0:00:05.624 ****** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [horizon : Check if policies shall be overwritten] ************************",
                                "Tuesday 24 February 2026  00:59:40 +0000 (0:00:00.366)       0:00:05.991 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [horizon : Update custom policy file name] ********************************",
                                "Tuesday 24 February 2026  00:59:40 +0000 (0:00:00.289)       0:00:06.281 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [horizon : Update policy file name] ***************************************",
                                "Tuesday 24 February 2026  00:59:41 +0000 (0:00:00.277)       0:00:06.558 ****** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [horizon : Check if policies shall be overwritten] ************************",
                                "Tuesday 24 February 2026  00:59:41 +0000 (0:00:00.322)       0:00:06.881 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [horizon : Update custom policy file name] ********************************",
                                "Tuesday 24 February 2026  00:59:41 +0000 (0:00:00.124)       0:00:07.005 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [horizon : Update policy file name] ***************************************",
                                "Tuesday 24 February 2026  00:59:41 +0000 (0:00:00.265)       0:00:07.271 ****** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [horizon : Check if policies shall be overwritten] ************************",
                                "Tuesday 24 February 2026  00:59:42 +0000 (0:00:00.475)       0:00:07.747 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [horizon : Update custom policy file name] ********************************",
                                "Tuesday 24 February 2026  00:59:42 +0000 (0:00:00.133)       0:00:07.881 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [horizon : Update policy file name] ***************************************",
                                "Tuesday 24 February 2026  00:59:42 +0000 (0:00:00.274)       0:00:08.155 ****** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [horizon : Check if policies shall be overwritten] ************************",
                                "Tuesday 24 February 2026  00:59:43 +0000 (0:00:00.307)       0:00:08.462 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [horizon : Update custom policy file name] ********************************",
                                "Tuesday 24 February 2026  00:59:43 +0000 (0:00:00.155)       0:00:08.617 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [horizon : Update policy file name] ***************************************",
                                "Tuesday 24 February 2026  00:59:43 +0000 (0:00:00.293)       0:00:08.911 ****** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [horizon : Check if policies shall be overwritten] ************************",
                                "Tuesday 24 February 2026  00:59:43 +0000 (0:00:00.466)       0:00:09.377 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [horizon : Update custom policy file name] ********************************",
                                "Tuesday 24 February 2026  00:59:44 +0000 (0:00:00.163)       0:00:09.541 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [horizon : Update policy file name] ***************************************",
                                "Tuesday 24 February 2026  00:59:44 +0000 (0:00:00.277)       0:00:09.818 ****** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [horizon : Check if policies shall be overwritten] ************************",
                                "Tuesday 24 February 2026  00:59:44 +0000 (0:00:00.314)       0:00:10.133 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [horizon : Update custom policy file name] ********************************",
                                "Tuesday 24 February 2026  00:59:44 +0000 (0:00:00.155)       0:00:10.289 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [horizon : Update policy file name] ***************************************",
                                "Tuesday 24 February 2026  00:59:45 +0000 (0:00:00.484)       0:00:10.773 ****** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [horizon : Check if policies shall be overwritten] ************************",
                                "Tuesday 24 February 2026  00:59:45 +0000 (0:00:00.354)       0:00:11.128 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [horizon : Update custom policy file name] ********************************",
                                "Tuesday 24 February 2026  00:59:45 +0000 (0:00:00.131)       0:00:11.259 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [horizon : Update policy file name] ***************************************",
                                "Tuesday 24 February 2026  00:59:46 +0000 (0:00:00.296)       0:00:11.556 ****** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [horizon : Check if policies shall be overwritten] ************************",
                                "Tuesday 24 February 2026  00:59:46 +0000 (0:00:00.309)       0:00:11.865 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [horizon : Update custom policy file name] ********************************",
                                "Tuesday 24 February 2026  00:59:46 +0000 (0:00:00.148)       0:00:12.014 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [horizon : Copying over config.json files for services] *******************",
                                "Tuesday 24 February 2026  00:59:47 +0000 (0:00:00.529)       0:00:12.544 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [horizon : Copying over horizon.conf] *************************************",
                                "Tuesday 24 February 2026  00:59:48 +0000 (0:00:01.616)       0:00:14.160 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=/ansible/roles/horizon/templates/horizon.conf.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=/ansible/roles/horizon/templates/horizon.conf.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=/ansible/roles/horizon/templates/horizon.conf.j2)\u001b[0m",
                                "",
                                "TASK [horizon : Copying over kolla-settings.py] ********************************",
                                "Tuesday 24 February 2026  00:59:50 +0000 (0:00:02.210)       0:00:16.371 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=/ansible/roles/horizon/templates/_9998-kolla-settings.py.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=/ansible/roles/horizon/templates/_9998-kolla-settings.py.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=/ansible/roles/horizon/templates/_9998-kolla-settings.py.j2)\u001b[0m",
                                "",
                                "TASK [horizon : Copying over custom-settings.py] *******************************",
                                "Tuesday 24 February 2026  00:59:53 +0000 (0:00:02.339)       0:00:18.710 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=/ansible/roles/horizon/templates/_9999-custom-settings.py.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=/ansible/roles/horizon/templates/_9999-custom-settings.py.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=/ansible/roles/horizon/templates/_9999-custom-settings.py.j2)\u001b[0m",
                                "",
                                "TASK [horizon : Copying over existing policy file] *****************************",
                                "Tuesday 24 February 2026  00:59:55 +0000 (0:00:02.371)       0:00:21.082 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [horizon : Copying over custom themes] ************************************",
                                "Tuesday 24 February 2026  00:59:56 +0000 (0:00:00.382)       0:00:21.464 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [horizon : include_tasks] *************************************************",
                                "Tuesday 24 February 2026  00:59:56 +0000 (0:00:00.427)       0:00:21.892 ****** ",
                                "\u001b[0;36mincluded: /ansible/roles/horizon/tasks/copy-certs.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [service-cert-copy : horizon | Copying over extra CA certificates] ********",
                                "Tuesday 24 February 2026  00:59:57 +0000 (0:00:00.918)       0:00:22.810 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'horizon', 'value': {'container_name': 'horizon', 'group': 'horizon', 'enabled': True, 'image': 'registry.osism.tech/kolla/horizon:2024.2', 'environment': {'ENABLE_BLAZAR': 'no', 'ENABLE_CLOUDKITTY': 'no', 'ENABLE_DESIGNATE': 'yes', 'ENABLE_FWAAS': 'no', 'ENABLE_HEAT': 'no', 'ENABLE_IRONIC': 'no', 'ENABLE_MAGNUM': 'yes', 'ENABLE_MANILA': 'yes', 'ENABLE_MASAKARI': 'no', 'ENABLE_MISTRAL': 'no', 'ENABLE_NEUTRON_VPNAAS': 'no', 'ENABLE_OCTAVIA': 'yes', 'ENABLE_TACKER': 'no', 'ENABLE_TROVE': 'no', 'ENABLE_WATCHER': 'no', 'ENABLE_ZUN': 'no', 'FORCE_GENERATE': 'no'}, 'volumes': ['/etc/kolla/horizon/:/var/lib/kolla/config_files/:ro', '', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:80'], 'timeout': '30'}, 'haproxy': {'horizon': {'enabled': True, 'mode': 'http', 'external': False, 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_redirect': {'enabled': True, 'mode': 'redirect', 'external': False, 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'horizon_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_external_redirect': {'enabled': True, 'mode': 'redirect', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'acme_client': {'enabled': True, 'with_frontend': False, 'custom_member_list': []}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'horizon', 'value': {'container_name': 'horizon', 'group': 'horizon', 'enabled': True, 'image': 'registry.osism.tech/kolla/horizon:2024.2', 'environment': {'ENABLE_BLAZAR': 'no', 'ENABLE_CLOUDKITTY': 'no', 'ENABLE_DESIGNATE': 'yes', 'ENABLE_FWAAS': 'no', 'ENABLE_HEAT': 'no', 'ENABLE_IRONIC': 'no', 'ENABLE_MAGNUM': 'yes', 'ENABLE_MANILA': 'yes', 'ENABLE_MASAKARI': 'no', 'ENABLE_MISTRAL': 'no', 'ENABLE_NEUTRON_VPNAAS': 'no', 'ENABLE_OCTAVIA': 'yes', 'ENABLE_TACKER': 'no', 'ENABLE_TROVE': 'no', 'ENABLE_WATCHER': 'no', 'ENABLE_ZUN': 'no', 'FORCE_GENERATE': 'no'}, 'volumes': ['/etc/kolla/horizon/:/var/lib/kolla/config_files/:ro', '', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:80'], 'timeout': '30'}, 'haproxy': {'horizon': {'enabled': True, 'mode': 'http', 'external': False, 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_redirect': {'enabled': True, 'mode': 'redirect', 'external': False, 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'horizon_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_external_redirect': {'enabled': True, 'mode': 'redirect', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'acme_client': {'enabled': True, 'with_frontend': False, 'custom_member_list': []}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'horizon', 'value': {'container_name': 'horizon', 'group': 'horizon', 'enabled': True, 'image': 'registry.osism.tech/kolla/horizon:2024.2', 'environment': {'ENABLE_BLAZAR': 'no', 'ENABLE_CLOUDKITTY': 'no', 'ENABLE_DESIGNATE': 'yes', 'ENABLE_FWAAS': 'no', 'ENABLE_HEAT': 'no', 'ENABLE_IRONIC': 'no', 'ENABLE_MAGNUM': 'yes', 'ENABLE_MANILA': 'yes', 'ENABLE_MASAKARI': 'no', 'ENABLE_MISTRAL': 'no', 'ENABLE_NEUTRON_VPNAAS': 'no', 'ENABLE_OCTAVIA': 'yes', 'ENABLE_TACKER': 'no', 'ENABLE_TROVE': 'no', 'ENABLE_WATCHER': 'no', 'ENABLE_ZUN': 'no', 'FORCE_GENERATE': 'no'}, 'volumes': ['/etc/kolla/horizon/:/var/lib/kolla/config_files/:ro', '', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:80'], 'timeout': '30'}, 'haproxy': {'horizon': {'enabled': True, 'mode': 'http', 'external': False, 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_redirect': {'enabled': True, 'mode': 'redirect', 'external': False, 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'horizon_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_external_redirect': {'enabled': True, 'mode': 'redirect', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'acme_client': {'enabled': True, 'with_frontend': False, 'custom_member_list': []}}}})\u001b[0m",
                                "",
                                "TASK [service-cert-copy : horizon | Copying over backend internal TLS certificate] ***",
                                "Tuesday 24 February 2026  00:59:59 +0000 (0:00:01.716)       0:00:24.526 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'horizon', 'value': {'container_name': 'horizon', 'group': 'horizon', 'enabled': True, 'image': 'registry.osism.tech/kolla/horizon:2024.2', 'environment': {'ENABLE_BLAZAR': 'no', 'ENABLE_CLOUDKITTY': 'no', 'ENABLE_DESIGNATE': 'yes', 'ENABLE_FWAAS': 'no', 'ENABLE_HEAT': 'no', 'ENABLE_IRONIC': 'no', 'ENABLE_MAGNUM': 'yes', 'ENABLE_MANILA': 'yes', 'ENABLE_MASAKARI': 'no', 'ENABLE_MISTRAL': 'no', 'ENABLE_NEUTRON_VPNAAS': 'no', 'ENABLE_OCTAVIA': 'yes', 'ENABLE_TACKER': 'no', 'ENABLE_TROVE': 'no', 'ENABLE_WATCHER': 'no', 'ENABLE_ZUN': 'no', 'FORCE_GENERATE': 'no'}, 'volumes': ['/etc/kolla/horizon/:/var/lib/kolla/config_files/:ro', '', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:80'], 'timeout': '30'}, 'haproxy': {'horizon': {'enabled': True, 'mode': 'http', 'external': False, 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_redirect': {'enabled': True, 'mode': 'redirect', 'external': False, 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'horizon_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_external_redirect': {'enabled': True, 'mode': 'redirect', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'acme_client': {'enabled': True, 'with_frontend': False, 'custom_member_list': []}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'horizon', 'value': {'container_name': 'horizon', 'group': 'horizon', 'enabled': True, 'image': 'registry.osism.tech/kolla/horizon:2024.2', 'environment': {'ENABLE_BLAZAR': 'no', 'ENABLE_CLOUDKITTY': 'no', 'ENABLE_DESIGNATE': 'yes', 'ENABLE_FWAAS': 'no', 'ENABLE_HEAT': 'no', 'ENABLE_IRONIC': 'no', 'ENABLE_MAGNUM': 'yes', 'ENABLE_MANILA': 'yes', 'ENABLE_MASAKARI': 'no', 'ENABLE_MISTRAL': 'no', 'ENABLE_NEUTRON_VPNAAS': 'no', 'ENABLE_OCTAVIA': 'yes', 'ENABLE_TACKER': 'no', 'ENABLE_TROVE': 'no', 'ENABLE_WATCHER': 'no', 'ENABLE_ZUN': 'no', 'FORCE_GENERATE': 'no'}, 'volumes': ['/etc/kolla/horizon/:/var/lib/kolla/config_files/:ro', '', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:80'], 'timeout': '30'}, 'haproxy': {'horizon': {'enabled': True, 'mode': 'http', 'external': False, 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_redirect': {'enabled': True, 'mode': 'redirect', 'external': False, 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'horizon_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_external_redirect': {'enabled': True, 'mode': 'redirect', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'acme_client': {'enabled': True, 'with_frontend': False, 'custom_member_list': []}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'horizon', 'value': {'container_name': 'horizon', 'group': 'horizon', 'enabled': True, 'image': 'registry.osism.tech/kolla/horizon:2024.2', 'environment': {'ENABLE_BLAZAR': 'no', 'ENABLE_CLOUDKITTY': 'no', 'ENABLE_DESIGNATE': 'yes', 'ENABLE_FWAAS': 'no', 'ENABLE_HEAT': 'no', 'ENABLE_IRONIC': 'no', 'ENABLE_MAGNUM': 'yes', 'ENABLE_MANILA': 'yes', 'ENABLE_MASAKARI': 'no', 'ENABLE_MISTRAL': 'no', 'ENABLE_NEUTRON_VPNAAS': 'no', 'ENABLE_OCTAVIA': 'yes', 'ENABLE_TACKER': 'no', 'ENABLE_TROVE': 'no', 'ENABLE_WATCHER': 'no', 'ENABLE_ZUN': 'no', 'FORCE_GENERATE': 'no'}, 'volumes': ['/etc/kolla/horizon/:/var/lib/kolla/config_files/:ro', '', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:80'], 'timeout': '30'}, 'haproxy': {'horizon': {'enabled': True, 'mode': 'http', 'external': False, 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_redirect': {'enabled': True, 'mode': 'redirect', 'external': False, 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'horizon_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_external_redirect': {'enabled': True, 'mode': 'redirect', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'acme_client': {'enabled': True, 'with_frontend': False, 'custom_member_list': []}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [service-cert-copy : horizon | Copying over backend internal TLS key] *****",
                                "Tuesday 24 February 2026  00:59:59 +0000 (0:00:00.855)       0:00:25.381 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'horizon', 'value': {'container_name': 'horizon', 'group': 'horizon', 'enabled': True, 'image': 'registry.osism.tech/kolla/horizon:2024.2', 'environment': {'ENABLE_BLAZAR': 'no', 'ENABLE_CLOUDKITTY': 'no', 'ENABLE_DESIGNATE': 'yes', 'ENABLE_FWAAS': 'no', 'ENABLE_HEAT': 'no', 'ENABLE_IRONIC': 'no', 'ENABLE_MAGNUM': 'yes', 'ENABLE_MANILA': 'yes', 'ENABLE_MASAKARI': 'no', 'ENABLE_MISTRAL': 'no', 'ENABLE_NEUTRON_VPNAAS': 'no', 'ENABLE_OCTAVIA': 'yes', 'ENABLE_TACKER': 'no', 'ENABLE_TROVE': 'no', 'ENABLE_WATCHER': 'no', 'ENABLE_ZUN': 'no', 'FORCE_GENERATE': 'no'}, 'volumes': ['/etc/kolla/horizon/:/var/lib/kolla/config_files/:ro', '', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:80'], 'timeout': '30'}, 'haproxy': {'horizon': {'enabled': True, 'mode': 'http', 'external': False, 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_redirect': {'enabled': True, 'mode': 'redirect', 'external': False, 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'horizon_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_external_redirect': {'enabled': True, 'mode': 'redirect', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'acme_client': {'enabled': True, 'with_frontend': False, 'custom_member_list': []}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'horizon', 'value': {'container_name': 'horizon', 'group': 'horizon', 'enabled': True, 'image': 'registry.osism.tech/kolla/horizon:2024.2', 'environment': {'ENABLE_BLAZAR': 'no', 'ENABLE_CLOUDKITTY': 'no', 'ENABLE_DESIGNATE': 'yes', 'ENABLE_FWAAS': 'no', 'ENABLE_HEAT': 'no', 'ENABLE_IRONIC': 'no', 'ENABLE_MAGNUM': 'yes', 'ENABLE_MANILA': 'yes', 'ENABLE_MASAKARI': 'no', 'ENABLE_MISTRAL': 'no', 'ENABLE_NEUTRON_VPNAAS': 'no', 'ENABLE_OCTAVIA': 'yes', 'ENABLE_TACKER': 'no', 'ENABLE_TROVE': 'no', 'ENABLE_WATCHER': 'no', 'ENABLE_ZUN': 'no', 'FORCE_GENERATE': 'no'}, 'volumes': ['/etc/kolla/horizon/:/var/lib/kolla/config_files/:ro', '', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:80'], 'timeout': '30'}, 'haproxy': {'horizon': {'enabled': True, 'mode': 'http', 'external': False, 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_redirect': {'enabled': True, 'mode': 'redirect', 'external': False, 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'horizon_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_external_redirect': {'enabled': True, 'mode': 'redirect', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'acme_client': {'enabled': True, 'with_frontend': False, 'custom_member_list': []}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'horizon', 'value': {'container_name': 'horizon', 'group': 'horizon', 'enabled': True, 'image': 'registry.osism.tech/kolla/horizon:2024.2', 'environment': {'ENABLE_BLAZAR': 'no', 'ENABLE_CLOUDKITTY': 'no', 'ENABLE_DESIGNATE': 'yes', 'ENABLE_FWAAS': 'no', 'ENABLE_HEAT': 'no', 'ENABLE_IRONIC': 'no', 'ENABLE_MAGNUM': 'yes', 'ENABLE_MANILA': 'yes', 'ENABLE_MASAKARI': 'no', 'ENABLE_MISTRAL': 'no', 'ENABLE_NEUTRON_VPNAAS': 'no', 'ENABLE_OCTAVIA': 'yes', 'ENABLE_TACKER': 'no', 'ENABLE_TROVE': 'no', 'ENABLE_WATCHER': 'no', 'ENABLE_ZUN': 'no', 'FORCE_GENERATE': 'no'}, 'volumes': ['/etc/kolla/horizon/:/var/lib/kolla/config_files/:ro', '', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:80'], 'timeout': '30'}, 'haproxy': {'horizon': {'enabled': True, 'mode': 'http', 'external': False, 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_redirect': {'enabled': True, 'mode': 'redirect', 'external': False, 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'horizon_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_external_redirect': {'enabled': True, 'mode': 'redirect', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'acme_client': {'enabled': True, 'with_frontend': False, 'custom_member_list': []}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [horizon : Deploy horizon container] **************************************",
                                "Tuesday 24 February 2026  01:00:01 +0000 (0:00:01.038)       0:00:26.420 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'horizon', 'value': {'container_name': 'horizon', 'group': 'horizon', 'enabled': True, 'image': 'registry.osism.tech/kolla/horizon:2024.2', 'environment': {'ENABLE_BLAZAR': 'no', 'ENABLE_CLOUDKITTY': 'no', 'ENABLE_DESIGNATE': 'yes', 'ENABLE_FWAAS': 'no', 'ENABLE_HEAT': 'no', 'ENABLE_IRONIC': 'no', 'ENABLE_MAGNUM': 'yes', 'ENABLE_MANILA': 'yes', 'ENABLE_MASAKARI': 'no', 'ENABLE_MISTRAL': 'no', 'ENABLE_NEUTRON_VPNAAS': 'no', 'ENABLE_OCTAVIA': 'yes', 'ENABLE_TACKER': 'no', 'ENABLE_TROVE': 'no', 'ENABLE_WATCHER': 'no', 'ENABLE_ZUN': 'no', 'FORCE_GENERATE': 'no'}, 'volumes': ['/etc/kolla/horizon/:/var/lib/kolla/config_files/:ro', '', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:80'], 'timeout': '30'}, 'haproxy': {'horizon': {'enabled': True, 'mode': 'http', 'external': False, 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_redirect': {'enabled': True, 'mode': 'redirect', 'external': False, 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'horizon_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_external_redirect': {'enabled': True, 'mode': 'redirect', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'acme_client': {'enabled': True, 'with_frontend': False, 'custom_member_list': []}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'horizon', 'value': {'container_name': 'horizon', 'group': 'horizon', 'enabled': True, 'image': 'registry.osism.tech/kolla/horizon:2024.2', 'environment': {'ENABLE_BLAZAR': 'no', 'ENABLE_CLOUDKITTY': 'no', 'ENABLE_DESIGNATE': 'yes', 'ENABLE_FWAAS': 'no', 'ENABLE_HEAT': 'no', 'ENABLE_IRONIC': 'no', 'ENABLE_MAGNUM': 'yes', 'ENABLE_MANILA': 'yes', 'ENABLE_MASAKARI': 'no', 'ENABLE_MISTRAL': 'no', 'ENABLE_NEUTRON_VPNAAS': 'no', 'ENABLE_OCTAVIA': 'yes', 'ENABLE_TACKER': 'no', 'ENABLE_TROVE': 'no', 'ENABLE_WATCHER': 'no', 'ENABLE_ZUN': 'no', 'FORCE_GENERATE': 'no'}, 'volumes': ['/etc/kolla/horizon/:/var/lib/kolla/config_files/:ro', '', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:80'], 'timeout': '30'}, 'haproxy': {'horizon': {'enabled': True, 'mode': 'http', 'external': False, 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_redirect': {'enabled': True, 'mode': 'redirect', 'external': False, 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'horizon_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_external_redirect': {'enabled': True, 'mode': 'redirect', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'acme_client': {'enabled': True, 'with_frontend': False, 'custom_member_list': []}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'horizon', 'value': {'container_name': 'horizon', 'group': 'horizon', 'enabled': True, 'image': 'registry.osism.tech/kolla/horizon:2024.2', 'environment': {'ENABLE_BLAZAR': 'no', 'ENABLE_CLOUDKITTY': 'no', 'ENABLE_DESIGNATE': 'yes', 'ENABLE_FWAAS': 'no', 'ENABLE_HEAT': 'no', 'ENABLE_IRONIC': 'no', 'ENABLE_MAGNUM': 'yes', 'ENABLE_MANILA': 'yes', 'ENABLE_MASAKARI': 'no', 'ENABLE_MISTRAL': 'no', 'ENABLE_NEUTRON_VPNAAS': 'no', 'ENABLE_OCTAVIA': 'yes', 'ENABLE_TACKER': 'no', 'ENABLE_TROVE': 'no', 'ENABLE_WATCHER': 'no', 'ENABLE_ZUN': 'no', 'FORCE_GENERATE': 'no'}, 'volumes': ['/etc/kolla/horizon/:/var/lib/kolla/config_files/:ro', '', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:80'], 'timeout': '30'}, 'haproxy': {'horizon': {'enabled': True, 'mode': 'http', 'external': False, 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_redirect': {'enabled': True, 'mode': 'redirect', 'external': False, 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'horizon_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '443', 'listen_port': '80', 'frontend_http_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }'], 'backend_http_extra': ['balance roundrobin'], 'tls_backend': 'no'}, 'horizon_external_redirect': {'enabled': True, 'mode': 'redirect', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '80', 'listen_port': '80', 'frontend_redirect_extra': ['use_backend acme_client_back if { path_reg ^/.well-known/acme-challenge/.+ }']}, 'acme_client': {'enabled': True, 'with_frontend': False, 'custom_member_list': []}}}})\u001b[0m",
                                "",
                                "TASK [horizon : include_tasks] *************************************************",
                                "Tuesday 24 February 2026  01:00:02 +0000 (0:00:01.756)       0:00:28.176 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [horizon : include_tasks] *************************************************",
                                "Tuesday 24 February 2026  01:00:03 +0000 (0:00:00.295)       0:00:28.471 ****** ",
                                "\u001b[0;36mincluded: /ansible/roles/horizon/tasks/bootstrap.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [horizon : Creating Horizon database] *************************************",
                                "Tuesday 24 February 2026  01:00:03 +0000 (0:00:00.586)       0:00:29.058 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [horizon : Creating Horizon database user and setting permissions] ********",
                                "Tuesday 24 February 2026  01:00:06 +0000 (0:00:02.981)       0:00:32.040 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [horizon : Running Horizon bootstrap container] ***************************",
                                "Tuesday 24 February 2026  01:00:09 +0000 (0:00:03.089)       0:00:35.130 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [horizon : Flush handlers] ************************************************",
                                "Tuesday 24 February 2026  01:00:28 +0000 (0:00:18.547)       0:00:53.677 ****** ",
                                "",
                                "TASK [horizon : Flush handlers] ************************************************",
                                "Tuesday 24 February 2026  01:00:28 +0000 (0:00:00.129)       0:00:53.807 ****** ",
                                "",
                                "TASK [horizon : Flush handlers] ************************************************",
                                "Tuesday 24 February 2026  01:00:28 +0000 (0:00:00.102)       0:00:53.910 ****** ",
                                "",
                                "RUNNING HANDLER [horizon : Restart horizon container] **************************",
                                "Tuesday 24 February 2026  01:00:28 +0000 (0:00:00.074)       0:00:53.984 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=37  \u001b[0m \u001b[0;33mchanged=11  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=25  \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=34  \u001b[0m \u001b[0;33mchanged=8   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=15  \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=34  \u001b[0m \u001b[0;33mchanged=8   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=15  \u001b[0m rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Tuesday 24 February 2026  01:01:19 +0000 (0:00:50.936)       0:01:44.921 ****** ",
                                "=============================================================================== ",
                                "horizon : Restart horizon container ------------------------------------ 50.94s",
                                "horizon : Running Horizon bootstrap container -------------------------- 18.55s",
                                "horizon : Creating Horizon database user and setting permissions -------- 3.09s",
                                "horizon : Creating Horizon database ------------------------------------- 2.98s",
                                "horizon : Copying over custom-settings.py ------------------------------- 2.37s",
                                "horizon : Copying over kolla-settings.py -------------------------------- 2.34s",
                                "horizon : Copying over horizon.conf ------------------------------------- 2.21s",
                                "horizon : Deploy horizon container -------------------------------------- 1.76s",
                                "service-cert-copy : horizon | Copying over extra CA certificates -------- 1.72s",
                                "horizon : Copying over config.json files for services ------------------- 1.62s",
                                "horizon : Ensuring config directories exist ----------------------------- 1.20s",
                                "service-cert-copy : horizon | Copying over backend internal TLS key ----- 1.04s",
                                "horizon : include_tasks ------------------------------------------------- 0.92s",
                                "service-cert-copy : horizon | Copying over backend internal TLS certificate --- 0.86s",
                                "horizon : include_tasks ------------------------------------------------- 0.78s",
                                "horizon : include_tasks ------------------------------------------------- 0.59s",
                                "horizon : Update custom policy file name -------------------------------- 0.53s",
                                "horizon : include_tasks ------------------------------------------------- 0.50s",
                                "horizon : Update custom policy file name -------------------------------- 0.48s",
                                "horizon : Update policy file name --------------------------------------- 0.48s",
                                "\u001b[32m2026-02-24 01:01:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9cb83792-8f4e-4574-a78c-9b0e654d6fd6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:01:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 01:01:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c5f0c529-69aa-4268-97af-bde1796abaa2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:01:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9cb83792-8f4e-4574-a78c-9b0e654d6fd6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:01:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 01:01:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c5f0c529-69aa-4268-97af-bde1796abaa2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:01:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9cb83792-8f4e-4574-a78c-9b0e654d6fd6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:01:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 01:01:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c5f0c529-69aa-4268-97af-bde1796abaa2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:01:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9cb83792-8f4e-4574-a78c-9b0e654d6fd6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:01:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 01:01:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c5f0c529-69aa-4268-97af-bde1796abaa2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:01:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9cb83792-8f4e-4574-a78c-9b0e654d6fd6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:01:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 01:01:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c5f0c529-69aa-4268-97af-bde1796abaa2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:01:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9cb83792-8f4e-4574-a78c-9b0e654d6fd6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:01:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 01:01:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c5f0c529-69aa-4268-97af-bde1796abaa2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:01:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9cb83792-8f4e-4574-a78c-9b0e654d6fd6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:01:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 01:01:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c5f0c529-69aa-4268-97af-bde1796abaa2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:01:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9cb83792-8f4e-4574-a78c-9b0e654d6fd6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:01:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 01:01:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c5f0c529-69aa-4268-97af-bde1796abaa2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:01:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9cb83792-8f4e-4574-a78c-9b0e654d6fd6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:01:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 01:01:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c5f0c529-69aa-4268-97af-bde1796abaa2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:01:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9cb83792-8f4e-4574-a78c-9b0e654d6fd6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:01:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 01:01:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask c5f0c529-69aa-4268-97af-bde1796abaa2 is in state SUCCESS\u001b[0m",
                                "\u001b[32m2026-02-24 01:01:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9cb83792-8f4e-4574-a78c-9b0e654d6fd6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:01:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 81556691-758e-4eda-9a1d-3b717f8d6221 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:01:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 01:01:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9cb83792-8f4e-4574-a78c-9b0e654d6fd6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:01:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 81556691-758e-4eda-9a1d-3b717f8d6221 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:01:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 01:01:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9cb83792-8f4e-4574-a78c-9b0e654d6fd6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:01:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 81556691-758e-4eda-9a1d-3b717f8d6221 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:01:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 01:02:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9cb83792-8f4e-4574-a78c-9b0e654d6fd6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:02:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 81556691-758e-4eda-9a1d-3b717f8d6221 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:02:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 01:02:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9cb83792-8f4e-4574-a78c-9b0e654d6fd6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:02:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 81556691-758e-4eda-9a1d-3b717f8d6221 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:02:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 01:02:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9cb83792-8f4e-4574-a78c-9b0e654d6fd6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:02:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 81556691-758e-4eda-9a1d-3b717f8d6221 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:02:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 01:02:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9cb83792-8f4e-4574-a78c-9b0e654d6fd6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:02:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 81556691-758e-4eda-9a1d-3b717f8d6221 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:02:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 01:02:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9cb83792-8f4e-4574-a78c-9b0e654d6fd6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:02:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 81556691-758e-4eda-9a1d-3b717f8d6221 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:02:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 01:02:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9cb83792-8f4e-4574-a78c-9b0e654d6fd6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:02:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 81556691-758e-4eda-9a1d-3b717f8d6221 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:02:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 01:02:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9cb83792-8f4e-4574-a78c-9b0e654d6fd6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:02:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 81556691-758e-4eda-9a1d-3b717f8d6221 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:02:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 01:02:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9cb83792-8f4e-4574-a78c-9b0e654d6fd6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:02:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 81556691-758e-4eda-9a1d-3b717f8d6221 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:02:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 01:02:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9cb83792-8f4e-4574-a78c-9b0e654d6fd6 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:02:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 81556691-758e-4eda-9a1d-3b717f8d6221 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:02:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 01:02:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 9cb83792-8f4e-4574-a78c-9b0e654d6fd6 is in state SUCCESS\u001b[0m",
                                "",
                                "",
                                "PLAY [Copy ceph keys to the configuration repository] **************************",
                                "",
                                "TASK [Check if ceph keys exist] ************************************************",
                                "Tuesday 24 February 2026  01:01:11 +0000 (0:00:00.182)       0:00:00.182 ****** ",
                                "\u001b[0;32mok: [testbed-manager -> testbed-node-0(192.168.16.10)] => (item=ceph.client.admin.keyring)\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager -> testbed-node-0(192.168.16.10)] => (item=ceph.client.cinder.keyring)\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager -> testbed-node-0(192.168.16.10)] => (item=ceph.client.cinder.keyring)\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager -> testbed-node-0(192.168.16.10)] => (item=ceph.client.cinder-backup.keyring)\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager -> testbed-node-0(192.168.16.10)] => (item=ceph.client.cinder.keyring)\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager -> testbed-node-0(192.168.16.10)] => (item=ceph.client.nova.keyring)\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager -> testbed-node-0(192.168.16.10)] => (item=ceph.client.glance.keyring)\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager -> testbed-node-0(192.168.16.10)] => (item=ceph.client.gnocchi.keyring)\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager -> testbed-node-0(192.168.16.10)] => (item=ceph.client.manila.keyring)\u001b[0m",
                                "",
                                "TASK [Fetch all ceph keys] *****************************************************",
                                "Tuesday 24 February 2026  01:01:17 +0000 (0:00:05.357)       0:00:05.540 ****** ",
                                "\u001b[0;32mok: [testbed-manager -> testbed-node-0(192.168.16.10)] => (item=ceph.client.admin.keyring)\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager -> testbed-node-0(192.168.16.10)] => (item=ceph.client.cinder.keyring)\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager -> testbed-node-0(192.168.16.10)] => (item=ceph.client.cinder.keyring)\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager -> testbed-node-0(192.168.16.10)] => (item=ceph.client.cinder-backup.keyring)\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager -> testbed-node-0(192.168.16.10)] => (item=ceph.client.cinder.keyring)\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager -> testbed-node-0(192.168.16.10)] => (item=ceph.client.nova.keyring)\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager -> testbed-node-0(192.168.16.10)] => (item=ceph.client.glance.keyring)\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager -> testbed-node-0(192.168.16.10)] => (item=ceph.client.gnocchi.keyring)\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager -> testbed-node-0(192.168.16.10)] => (item=ceph.client.manila.keyring)\u001b[0m",
                                "",
                                "TASK [Create share directory] **************************************************",
                                "Tuesday 24 February 2026  01:01:22 +0000 (0:00:04.972)       0:00:10.512 ****** ",
                                "\u001b[0;33mchanged: [testbed-manager -> localhost]\u001b[0m",
                                "",
                                "TASK [Write ceph keys to the share directory] **********************************",
                                "Tuesday 24 February 2026  01:01:23 +0000 (0:00:01.118)       0:00:11.631 ****** ",
                                "\u001b[0;33mchanged: [testbed-manager -> localhost] => (item=ceph.client.admin.keyring)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager -> localhost] => (item=ceph.client.cinder.keyring)\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager -> localhost] => (item=ceph.client.cinder.keyring)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager -> localhost] => (item=ceph.client.cinder-backup.keyring)\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager -> localhost] => (item=ceph.client.cinder.keyring)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager -> localhost] => (item=ceph.client.nova.keyring)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager -> localhost] => (item=ceph.client.glance.keyring)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager -> localhost] => (item=ceph.client.gnocchi.keyring)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager -> localhost] => (item=ceph.client.manila.keyring)\u001b[0m",
                                "",
                                "TASK [Check if target directories exist] ***************************************",
                                "Tuesday 24 February 2026  01:01:38 +0000 (0:00:15.235)       0:00:26.866 ****** ",
                                "\u001b[0;32mok: [testbed-manager] => (item=/opt/configuration/environments/infrastructure/files/ceph)\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager] => (item=/opt/configuration/environments/kolla/files/overlays/cinder/cinder-volume)\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager] => (item=/opt/configuration/environments/kolla/files/overlays/cinder/cinder-backup)\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager] => (item=/opt/configuration/environments/kolla/files/overlays/cinder/cinder-backup)\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager] => (item=/opt/configuration/environments/kolla/files/overlays/nova)\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager] => (item=/opt/configuration/environments/kolla/files/overlays/nova)\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager] => (item=/opt/configuration/environments/kolla/files/overlays/glance)\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager] => (item=/opt/configuration/environments/kolla/files/overlays/gnocchi)\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager] => (item=/opt/configuration/environments/kolla/files/overlays/manila)\u001b[0m",
                                "",
                                "TASK [Write ceph keys to the configuration directory] **************************",
                                "Tuesday 24 February 2026  01:01:41 +0000 (0:00:03.514)       0:00:30.380 ****** ",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=ceph.client.admin.keyring)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=ceph.client.cinder.keyring)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=ceph.client.cinder.keyring)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=ceph.client.cinder-backup.keyring)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=ceph.client.cinder.keyring)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=ceph.client.nova.keyring)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=ceph.client.glance.keyring)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=ceph.client.gnocchi.keyring)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=ceph.client.manila.keyring)\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;33mtestbed-manager\u001b[0m            : \u001b[0;32mok=6   \u001b[0m \u001b[0;33mchanged=3   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Tuesday 24 February 2026  01:01:49 +0000 (0:00:07.411)       0:00:37.792 ****** ",
                                "=============================================================================== ",
                                "Write ceph keys to the share directory --------------------------------- 15.24s",
                                "Write ceph keys to the configuration directory -------------------------- 7.41s",
                                "Check if ceph keys exist ------------------------------------------------ 5.36s",
                                "Fetch all ceph keys ----------------------------------------------------- 4.97s",
                                "Check if target directories exist --------------------------------------- 3.51s",
                                "Create share directory -------------------------------------------------- 1.12s",
                                "",
                                "",
                                "PLAY [Group hosts based on configuration] **************************************",
                                "",
                                "TASK [Group hosts based on Kolla action] ***************************************",
                                "Tuesday 24 February 2026  00:59:34 +0000 (0:00:00.251)       0:00:00.251 ****** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [Group hosts based on enabled services] ***********************************",
                                "Tuesday 24 February 2026  00:59:35 +0000 (0:00:00.346)       0:00:00.598 ****** ",
                                "\u001b[0;32mok: [testbed-node-0] => (item=enable_keystone_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item=enable_keystone_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item=enable_keystone_True)\u001b[0m",
                                "",
                                "PLAY [Apply role keystone] *****************************************************",
                                "",
                                "TASK [keystone : include_tasks] ************************************************",
                                "Tuesday 24 February 2026  00:59:35 +0000 (0:00:00.438)       0:00:01.037 ****** ",
                                "\u001b[0;36mincluded: /ansible/roles/keystone/tasks/deploy.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [keystone : Ensuring config directories exist] ****************************",
                                "Tuesday 24 February 2026  00:59:36 +0000 (0:00:00.577)       0:00:01.614 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone:2024.2', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone:2024.2', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone:2024.2', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-ssh:2024.2', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-ssh:2024.2', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-ssh:2024.2', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-fernet:2024.2', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-fernet:2024.2', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-fernet:2024.2', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}})\u001b[0m",
                                "",
                                "TASK [keystone : Check if policies shall be overwritten] ***********************",
                                "Tuesday 24 February 2026  00:59:38 +0000 (0:00:01.908)       0:00:03.523 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [keystone : Set keystone policy file] *************************************",
                                "Tuesday 24 February 2026  00:59:38 +0000 (0:00:00.123)       0:00:03.646 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [keystone : Check if Keystone domain-specific config is supplied] *********",
                                "Tuesday 24 February 2026  00:59:38 +0000 (0:00:00.466)       0:00:04.113 ****** ",
                                "\u001b[0;32mok: [testbed-node-0 -> localhost]\u001b[0m",
                                "",
                                "TASK [keystone : include_tasks] ************************************************",
                                "Tuesday 24 February 2026  00:59:39 +0000 (0:00:00.878)       0:00:04.991 ****** ",
                                "\u001b[0;36mincluded: /ansible/roles/keystone/tasks/copy-certs.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [service-cert-copy : keystone | Copying over extra CA certificates] *******",
                                "Tuesday 24 February 2026  00:59:40 +0000 (0:00:00.576)       0:00:05.568 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone:2024.2', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone:2024.2', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone:2024.2', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-ssh:2024.2', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-ssh:2024.2', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-ssh:2024.2', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-fernet:2024.2', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-fernet:2024.2', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-fernet:2024.2', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}})\u001b[0m",
                                "",
                                "TASK [service-cert-copy : keystone | Copying over backend internal TLS certificate] ***",
                                "Tuesday 24 February 2026  00:59:43 +0000 (0:00:03.540)       0:00:09.109 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone:2024.2', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-ssh:2024.2', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-fernet:2024.2', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone:2024.2', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-ssh:2024.2', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-fernet:2024.2', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone:2024.2', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-ssh:2024.2', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-fernet:2024.2', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [service-cert-copy : keystone | Copying over backend internal TLS key] ****",
                                "Tuesday 24 February 2026  00:59:44 +0000 (0:00:00.638)       0:00:09.747 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone:2024.2', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-ssh:2024.2', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-fernet:2024.2', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone:2024.2', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-ssh:2024.2', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-fernet:2024.2', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone:2024.2', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-ssh:2024.2', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-fernet:2024.2', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [keystone : Copying over config.json files for services] ******************",
                                "Tuesday 24 February 2026  00:59:45 +0000 (0:00:00.763)       0:00:10.511 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone:2024.2', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone:2024.2', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone:2024.2', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-ssh:2024.2', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-ssh:2024.2', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-ssh:2024.2', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-fernet:2024.2', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-fernet:2024.2', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-fernet:2024.2', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}})\u001b[0m",
                                "",
                                "TASK [keystone : Copying over keystone.conf] ***********************************",
                                "Tuesday 24 February 2026  00:59:48 +0000 (0:00:03.411)       0:00:13.923 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone:2024.2', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone:2024.2', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-ssh:2024.2', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-ssh:2024.2', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone:2024.2', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-ssh:2024.2', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-fernet:2024.2', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-fernet:2024.2', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-fernet:2024.2', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}})\u001b[0m",
                                "",
                                "TASK [keystone : Copying keystone-startup script for keystone] *****************",
                                "Tuesday 24 February 2026  00:59:54 +0000 (0:00:06.096)       0:00:20.019 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [keystone : Create Keystone domain-specific config directory] *************",
                                "Tuesday 24 February 2026  00:59:56 +0000 (0:00:01.571)       0:00:21.590 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [keystone : Get file list in custom domains folder] ***********************",
                                "Tuesday 24 February 2026  00:59:56 +0000 (0:00:00.655)       0:00:22.246 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [keystone : Copying Keystone Domain specific settings] ********************",
                                "Tuesday 24 February 2026  00:59:57 +0000 (0:00:00.339)       0:00:22.585 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [keystone : Copying over existing policy file] ****************************",
                                "Tuesday 24 February 2026  00:59:57 +0000 (0:00:00.662)       0:00:23.247 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone:2024.2', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-ssh:2024.2', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-fernet:2024.2', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone:2024.2', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-ssh:2024.2', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-fernet:2024.2', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone:2024.2', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-ssh:2024.2', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-fernet:2024.2', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [keystone : include_tasks] ************************************************",
                                "Tuesday 24 February 2026  00:59:58 +0000 (0:00:00.645)       0:00:23.893 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [keystone : Copying over wsgi-keystone.conf] ******************************",
                                "Tuesday 24 February 2026  00:59:58 +0000 (0:00:00.378)       0:00:24.272 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=/ansible/roles/keystone/templates/wsgi-keystone.conf.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=/ansible/roles/keystone/templates/wsgi-keystone.conf.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=/ansible/roles/keystone/templates/wsgi-keystone.conf.j2)\u001b[0m",
                                "",
                                "TASK [keystone : Checking whether keystone-paste.ini file exists] **************",
                                "Tuesday 24 February 2026  01:00:00 +0000 (0:00:01.724)       0:00:25.997 ****** ",
                                "\u001b[0;32mok: [testbed-node-0 -> localhost]\u001b[0m",
                                "",
                                "TASK [keystone : Copying over keystone-paste.ini] ******************************",
                                "Tuesday 24 February 2026  01:00:01 +0000 (0:00:01.124)       0:00:27.122 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [keystone : Generate the required cron jobs for the node] *****************",
                                "Tuesday 24 February 2026  01:00:02 +0000 (0:00:01.015)       0:00:28.137 ****** ",
                                "\u001b[0;32mok: [testbed-node-1 -> localhost]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0 -> localhost]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2 -> localhost]\u001b[0m",
                                "",
                                "TASK [keystone : Set fact with the generated cron jobs for building the crontab later] ***",
                                "Tuesday 24 February 2026  01:00:03 +0000 (0:00:01.051)       0:00:29.188 ****** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [keystone : Copying files for keystone-fernet] ****************************",
                                "Tuesday 24 February 2026  01:00:04 +0000 (0:00:00.296)       0:00:29.485 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'src': 'crontab.j2', 'dest': 'crontab'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'src': 'crontab.j2', 'dest': 'crontab'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'src': 'crontab.j2', 'dest': 'crontab'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'src': 'fernet-rotate.sh.j2', 'dest': 'fernet-rotate.sh'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'src': 'fernet-rotate.sh.j2', 'dest': 'fernet-rotate.sh'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'src': 'fernet-rotate.sh.j2', 'dest': 'fernet-rotate.sh'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'src': 'fernet-node-sync.sh.j2', 'dest': 'fernet-node-sync.sh'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'src': 'fernet-node-sync.sh.j2', 'dest': 'fernet-node-sync.sh'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'src': 'fernet-node-sync.sh.j2', 'dest': 'fernet-node-sync.sh'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'src': 'fernet-push.sh.j2', 'dest': 'fernet-push.sh'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'src': 'fernet-push.sh.j2', 'dest': 'fernet-push.sh'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'src': 'fernet-push.sh.j2', 'dest': 'fernet-push.sh'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'src': 'fernet-healthcheck.sh.j2', 'dest': 'fernet-healthcheck.sh'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'src': 'fernet-healthcheck.sh.j2', 'dest': 'fernet-healthcheck.sh'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'src': 'fernet-healthcheck.sh.j2', 'dest': 'fernet-healthcheck.sh'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'src': 'id_rsa', 'dest': 'id_rsa'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'src': 'id_rsa', 'dest': 'id_rsa'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'src': 'id_rsa', 'dest': 'id_rsa'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'src': 'ssh_config.j2', 'dest': 'ssh_config'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'src': 'ssh_config.j2', 'dest': 'ssh_config'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'src': 'ssh_config.j2', 'dest': 'ssh_config'})\u001b[0m",
                                "",
                                "TASK [keystone : Copying files for keystone-ssh] *******************************",
                                "Tuesday 24 February 2026  01:00:13 +0000 (0:00:09.426)       0:00:38.912 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'src': 'sshd_config.j2', 'dest': 'sshd_config'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'src': 'sshd_config.j2', 'dest': 'sshd_config'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'src': 'sshd_config.j2', 'dest': 'sshd_config'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'src': 'id_rsa.pub', 'dest': 'id_rsa.pub'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'src': 'id_rsa.pub', 'dest': 'id_rsa.pub'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'src': 'id_rsa.pub', 'dest': 'id_rsa.pub'})\u001b[0m",
                                "",
                                "TASK [keystone : Check keystone containers] ************************************",
                                "Tuesday 24 February 2026  01:00:16 +0000 (0:00:02.888)       0:00:41.800 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone:2024.2', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone:2024.2', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'keystone', 'value': {'container_name': 'keystone', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone:2024.2', 'volumes': ['/etc/kolla/keystone/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', '', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:5000'], 'timeout': '30'}, 'haproxy': {'keystone_internal': {'enabled': True, 'mode': 'http', 'external': False, 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}, 'keystone_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'tls_backend': 'no', 'port': '5000', 'listen_port': '5000', 'backend_http_extra': ['balance roundrobin']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-ssh:2024.2', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-ssh:2024.2', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'keystone-ssh', 'value': {'container_name': 'keystone_ssh', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-ssh:2024.2', 'volumes': ['/etc/kolla/keystone-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8023'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-fernet:2024.2', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-fernet:2024.2', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'keystone-fernet', 'value': {'container_name': 'keystone_fernet', 'group': 'keystone', 'enabled': True, 'image': 'registry.osism.tech/kolla/keystone-fernet:2024.2', 'volumes': ['/etc/kolla/keystone-fernet/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'keystone_fernet_tokens:/etc/keystone/fernet-keys'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', '/usr/bin/fernet-healthcheck.sh'], 'timeout': '30'}}})\u001b[0m",
                                "",
                                "TASK [keystone : include_tasks] ************************************************",
                                "Tuesday 24 February 2026  01:00:19 +0000 (0:00:02.779)       0:00:44.579 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [keystone : Creating keystone database] ***********************************",
                                "Tuesday 24 February 2026  01:00:19 +0000 (0:00:00.387)       0:00:44.967 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [keystone : Creating Keystone database user and setting permissions] ******",
                                "Tuesday 24 February 2026  01:00:22 +0000 (0:00:02.942)       0:00:47.910 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [keystone : Checking for any running keystone_fernet containers] **********",
                                "Tuesday 24 February 2026  01:00:25 +0000 (0:00:02.689)       0:00:50.600 ****** ",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [keystone : Group nodes where keystone_fernet is running] *****************",
                                "Tuesday 24 February 2026  01:00:26 +0000 (0:00:01.140)       0:00:51.741 ****** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [keystone : Fail if any hosts need bootstrapping and not all hosts targeted] ***",
                                "Tuesday 24 February 2026  01:00:26 +0000 (0:00:00.334)       0:00:52.075 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [keystone : Running Keystone bootstrap container] *************************",
                                "Tuesday 24 February 2026  01:00:27 +0000 (0:00:00.344)       0:00:52.419 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [keystone : Running Keystone fernet bootstrap container] ******************",
                                "Tuesday 24 February 2026  01:00:44 +0000 (0:00:17.780)       0:01:10.199 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [keystone : Flush handlers] ***********************************************",
                                "Tuesday 24 February 2026  01:00:57 +0000 (0:00:12.957)       0:01:23.156 ****** ",
                                "",
                                "TASK [keystone : Flush handlers] ***********************************************",
                                "Tuesday 24 February 2026  01:00:57 +0000 (0:00:00.070)       0:01:23.227 ****** ",
                                "",
                                "TASK [keystone : Flush handlers] ***********************************************",
                                "Tuesday 24 February 2026  01:00:57 +0000 (0:00:00.068)       0:01:23.296 ****** ",
                                "",
                                "RUNNING HANDLER [keystone : Restart keystone-ssh container] ********************",
                                "Tuesday 24 February 2026  01:00:57 +0000 (0:00:00.071)       0:01:23.367 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [keystone : Restart keystone-fernet container] *****************",
                                "Tuesday 24 February 2026  01:01:10 +0000 (0:00:12.578)       0:01:35.947 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [keystone : Restart keystone container] ************************",
                                "Tuesday 24 February 2026  01:01:15 +0000 (0:00:05.419)       0:01:41.366 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [keystone : include_tasks] ************************************************",
                                "Tuesday 24 February 2026  01:01:27 +0000 (0:00:11.600)       0:01:52.967 ****** ",
                                "\u001b[0;36mincluded: /ansible/roles/keystone/tasks/distribute_fernet.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [keystone : Waiting for Keystone SSH port to be UP] ***********************",
                                "Tuesday 24 February 2026  01:01:28 +0000 (0:00:00.824)       0:01:53.792 ****** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [keystone : Run key distribution] *****************************************",
                                "Tuesday 24 February 2026  01:01:29 +0000 (0:00:00.893)       0:01:54.686 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [keystone : Creating admin project, user, role, service, and endpoint] ****",
                                "Tuesday 24 February 2026  01:01:31 +0000 (0:00:01.872)       0:01:56.559 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=RegionOne)\u001b[0m",
                                "",
                                "TASK [service-ks-register : keystone | Creating services] **********************",
                                "Tuesday 24 February 2026  01:01:45 +0000 (0:00:14.151)       0:02:10.710 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=keystone (identity))\u001b[0m",
                                "",
                                "TASK [service-ks-register : keystone | Creating endpoints] *********************",
                                "Tuesday 24 February 2026  01:02:13 +0000 (0:00:28.261)       0:02:38.972 ****** ",
                                "\u001b[0;32mok: [testbed-node-0] => (item=keystone -> https://api-int.testbed.osism.xyz:5000 -> internal)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0] => (item=keystone -> https://api.testbed.osism.xyz:5000 -> public)\u001b[0m",
                                "",
                                "TASK [service-ks-register : keystone | Creating projects] **********************",
                                "Tuesday 24 February 2026  01:02:21 +0000 (0:00:08.079)       0:02:47.051 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [service-ks-register : keystone | Creating users] *************************",
                                "Tuesday 24 February 2026  01:02:21 +0000 (0:00:00.144)       0:02:47.196 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [service-ks-register : keystone | Creating roles] *************************",
                                "Tuesday 24 February 2026  01:02:21 +0000 (0:00:00.124)       0:02:47.320 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [service-ks-register : keystone | Granting user roles] ********************",
                                "Tuesday 24 February 2026  01:02:22 +0000 (0:00:00.131)       0:02:47.452 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [keystone : Creating default user role] ***********************************",
                                "Tuesday 24 February 2026  01:02:22 +0000 (0:00:00.620)       0:02:48.072 ****** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [keystone : include_tasks] ************************************************",
                                "Tuesday 24 February 2026  01:02:26 +0000 (0:00:03.925)       0:02:51.997 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=33  \u001b[0m \u001b[0;33mchanged=19  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=17  \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=22  \u001b[0m \u001b[0;33mchanged=12  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=12  \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=22  \u001b[0m \u001b[0;33mchanged=12  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=12  \u001b[0m rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Tuesday 24 February 2026  01:02:27 +0000 (0:00:00.461)       0:02:52.459 ****** ",
                                "=============================================================================== ",
                                "service-ks-register : keystone | Creating services --------------------- 28.26s",
                                "keystone : Running Keystone bootstrap container ------------------------ 17.78s",
                                "keystone : Creating admin project, user, role, service, and endpoint --- 14.15s",
                                "keystone : Running Keystone fernet bootstrap container ----------------- 12.96s",
                                "keystone : Restart keystone-ssh container ------------------------------ 12.58s",
                                "keystone : Restart keystone container ---------------------------------- 11.60s",
                                "keystone : Copying files for keystone-fernet ---------------------------- 9.43s",
                                "service-ks-register : keystone | Creating endpoints --------------------- 8.08s",
                                "keystone : Copying over keystone.conf ----------------------------------- 6.10s",
                                "keystone : Restart keystone-fernet container ---------------------------- 5.42s",
                                "keystone : Creating default user role ----------------------------------- 3.93s",
                                "service-cert-copy : keystone | Copying over extra CA certificates ------- 3.54s",
                                "keystone : Copying over config.json files for services ------------------ 3.41s",
                                "keystone : Creating keystone database ----------------------------------- 2.94s",
                                "keystone : Copying files for keystone-ssh ------------------------------- 2.89s",
                                "keystone : Check keystone containers ------------------------------------ 2.78s",
                                "keystone : Creating Keystone database user and setting permissions ------ 2.69s",
                                "keystone : Ensuring config directories exist ---------------------------- 1.91s",
                                "keystone : Run key distribution ----------------------------------------- 1.87s",
                                "keystone : Copying over wsgi-keystone.conf ------------------------------ 1.73s",
                                "\u001b[32m2026-02-24 01:02:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 81556691-758e-4eda-9a1d-3b717f8d6221 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:02:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 01:02:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f714ea82-7678-4e20-85bd-ee1502bb5463 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:02:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask beffba80-d510-4944-9e2b-b4ff3c4fc669 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:02:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ac05741a-1939-415f-aeff-a7eed21049c9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:02:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 81556691-758e-4eda-9a1d-3b717f8d6221 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:02:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44aa611b-f898-4352-b174-ada638d8f758 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:02:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 01:02:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f714ea82-7678-4e20-85bd-ee1502bb5463 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:02:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask beffba80-d510-4944-9e2b-b4ff3c4fc669 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:02:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ac05741a-1939-415f-aeff-a7eed21049c9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:02:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 81556691-758e-4eda-9a1d-3b717f8d6221 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:02:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44aa611b-f898-4352-b174-ada638d8f758 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:02:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 01:02:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f714ea82-7678-4e20-85bd-ee1502bb5463 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:02:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask beffba80-d510-4944-9e2b-b4ff3c4fc669 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:02:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ac05741a-1939-415f-aeff-a7eed21049c9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:02:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 81556691-758e-4eda-9a1d-3b717f8d6221 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:02:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44aa611b-f898-4352-b174-ada638d8f758 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:02:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 01:02:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f714ea82-7678-4e20-85bd-ee1502bb5463 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:02:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask beffba80-d510-4944-9e2b-b4ff3c4fc669 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:02:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ac05741a-1939-415f-aeff-a7eed21049c9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:02:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 81556691-758e-4eda-9a1d-3b717f8d6221 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:02:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44aa611b-f898-4352-b174-ada638d8f758 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:02:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 01:02:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f714ea82-7678-4e20-85bd-ee1502bb5463 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:02:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask beffba80-d510-4944-9e2b-b4ff3c4fc669 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:02:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ac05741a-1939-415f-aeff-a7eed21049c9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:02:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 81556691-758e-4eda-9a1d-3b717f8d6221 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:02:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44aa611b-f898-4352-b174-ada638d8f758 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:02:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 01:02:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f714ea82-7678-4e20-85bd-ee1502bb5463 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:02:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask beffba80-d510-4944-9e2b-b4ff3c4fc669 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:02:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ac05741a-1939-415f-aeff-a7eed21049c9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:02:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 81556691-758e-4eda-9a1d-3b717f8d6221 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:02:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44aa611b-f898-4352-b174-ada638d8f758 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:02:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 01:02:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f714ea82-7678-4e20-85bd-ee1502bb5463 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:02:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask beffba80-d510-4944-9e2b-b4ff3c4fc669 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:02:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ac05741a-1939-415f-aeff-a7eed21049c9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:02:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 81556691-758e-4eda-9a1d-3b717f8d6221 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:02:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44aa611b-f898-4352-b174-ada638d8f758 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:02:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 01:02:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f714ea82-7678-4e20-85bd-ee1502bb5463 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:02:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask beffba80-d510-4944-9e2b-b4ff3c4fc669 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:02:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ac05741a-1939-415f-aeff-a7eed21049c9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:02:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 81556691-758e-4eda-9a1d-3b717f8d6221 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:02:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44aa611b-f898-4352-b174-ada638d8f758 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:02:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 01:02:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f714ea82-7678-4e20-85bd-ee1502bb5463 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:02:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask beffba80-d510-4944-9e2b-b4ff3c4fc669 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:02:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ac05741a-1939-415f-aeff-a7eed21049c9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:02:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 81556691-758e-4eda-9a1d-3b717f8d6221 is in state SUCCESS\u001b[0m",
                                "\u001b[32m2026-02-24 01:02:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6fd9e63e-df8b-4c9f-86ca-d2316078b744 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:02:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44aa611b-f898-4352-b174-ada638d8f758 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:02:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 01:02:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f714ea82-7678-4e20-85bd-ee1502bb5463 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:02:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask beffba80-d510-4944-9e2b-b4ff3c4fc669 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:02:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ac05741a-1939-415f-aeff-a7eed21049c9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:02:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6fd9e63e-df8b-4c9f-86ca-d2316078b744 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:02:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44aa611b-f898-4352-b174-ada638d8f758 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:02:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 01:03:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f714ea82-7678-4e20-85bd-ee1502bb5463 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:03:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask beffba80-d510-4944-9e2b-b4ff3c4fc669 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:03:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ac05741a-1939-415f-aeff-a7eed21049c9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:03:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6fd9e63e-df8b-4c9f-86ca-d2316078b744 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:03:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44aa611b-f898-4352-b174-ada638d8f758 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:03:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 01:03:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f714ea82-7678-4e20-85bd-ee1502bb5463 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:03:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask beffba80-d510-4944-9e2b-b4ff3c4fc669 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:03:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ac05741a-1939-415f-aeff-a7eed21049c9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:03:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6fd9e63e-df8b-4c9f-86ca-d2316078b744 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:03:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44aa611b-f898-4352-b174-ada638d8f758 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:03:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 01:03:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f714ea82-7678-4e20-85bd-ee1502bb5463 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:03:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask beffba80-d510-4944-9e2b-b4ff3c4fc669 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:03:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ac05741a-1939-415f-aeff-a7eed21049c9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:03:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6fd9e63e-df8b-4c9f-86ca-d2316078b744 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:03:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44aa611b-f898-4352-b174-ada638d8f758 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:03:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 01:03:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f714ea82-7678-4e20-85bd-ee1502bb5463 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:03:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask beffba80-d510-4944-9e2b-b4ff3c4fc669 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:03:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ac05741a-1939-415f-aeff-a7eed21049c9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:03:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6fd9e63e-df8b-4c9f-86ca-d2316078b744 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:03:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44aa611b-f898-4352-b174-ada638d8f758 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:03:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 01:03:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f714ea82-7678-4e20-85bd-ee1502bb5463 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:03:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask beffba80-d510-4944-9e2b-b4ff3c4fc669 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:03:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ac05741a-1939-415f-aeff-a7eed21049c9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:03:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6fd9e63e-df8b-4c9f-86ca-d2316078b744 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:03:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44aa611b-f898-4352-b174-ada638d8f758 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:03:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 01:03:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f714ea82-7678-4e20-85bd-ee1502bb5463 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:03:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask beffba80-d510-4944-9e2b-b4ff3c4fc669 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:03:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ac05741a-1939-415f-aeff-a7eed21049c9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:03:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6fd9e63e-df8b-4c9f-86ca-d2316078b744 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:03:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44aa611b-f898-4352-b174-ada638d8f758 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:03:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 01:03:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f714ea82-7678-4e20-85bd-ee1502bb5463 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:03:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask beffba80-d510-4944-9e2b-b4ff3c4fc669 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:03:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ac05741a-1939-415f-aeff-a7eed21049c9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:03:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6fd9e63e-df8b-4c9f-86ca-d2316078b744 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:03:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44aa611b-f898-4352-b174-ada638d8f758 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:03:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 01:03:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f714ea82-7678-4e20-85bd-ee1502bb5463 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:03:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask beffba80-d510-4944-9e2b-b4ff3c4fc669 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:03:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ac05741a-1939-415f-aeff-a7eed21049c9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:03:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6fd9e63e-df8b-4c9f-86ca-d2316078b744 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:03:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44aa611b-f898-4352-b174-ada638d8f758 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:03:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 01:03:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f714ea82-7678-4e20-85bd-ee1502bb5463 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:03:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask beffba80-d510-4944-9e2b-b4ff3c4fc669 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:03:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ac05741a-1939-415f-aeff-a7eed21049c9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:03:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6fd9e63e-df8b-4c9f-86ca-d2316078b744 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:03:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44aa611b-f898-4352-b174-ada638d8f758 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:03:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 01:03:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f714ea82-7678-4e20-85bd-ee1502bb5463 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:03:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask beffba80-d510-4944-9e2b-b4ff3c4fc669 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:03:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ac05741a-1939-415f-aeff-a7eed21049c9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:03:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6fd9e63e-df8b-4c9f-86ca-d2316078b744 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:03:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44aa611b-f898-4352-b174-ada638d8f758 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:03:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 01:03:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f714ea82-7678-4e20-85bd-ee1502bb5463 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:03:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask beffba80-d510-4944-9e2b-b4ff3c4fc669 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:03:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ac05741a-1939-415f-aeff-a7eed21049c9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:03:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6fd9e63e-df8b-4c9f-86ca-d2316078b744 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:03:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44aa611b-f898-4352-b174-ada638d8f758 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:03:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 01:03:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f714ea82-7678-4e20-85bd-ee1502bb5463 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:03:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask beffba80-d510-4944-9e2b-b4ff3c4fc669 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:03:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ac05741a-1939-415f-aeff-a7eed21049c9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:03:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6fd9e63e-df8b-4c9f-86ca-d2316078b744 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:03:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44aa611b-f898-4352-b174-ada638d8f758 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:03:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 01:03:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f714ea82-7678-4e20-85bd-ee1502bb5463 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:03:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask beffba80-d510-4944-9e2b-b4ff3c4fc669 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:03:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ac05741a-1939-415f-aeff-a7eed21049c9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:03:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6fd9e63e-df8b-4c9f-86ca-d2316078b744 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:03:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44aa611b-f898-4352-b174-ada638d8f758 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:03:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 01:03:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f714ea82-7678-4e20-85bd-ee1502bb5463 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:03:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask beffba80-d510-4944-9e2b-b4ff3c4fc669 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:03:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ac05741a-1939-415f-aeff-a7eed21049c9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:03:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6fd9e63e-df8b-4c9f-86ca-d2316078b744 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:03:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44aa611b-f898-4352-b174-ada638d8f758 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:03:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 01:03:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f714ea82-7678-4e20-85bd-ee1502bb5463 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:03:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask beffba80-d510-4944-9e2b-b4ff3c4fc669 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:03:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ac05741a-1939-415f-aeff-a7eed21049c9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:03:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6fd9e63e-df8b-4c9f-86ca-d2316078b744 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:03:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44aa611b-f898-4352-b174-ada638d8f758 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:03:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 01:03:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f714ea82-7678-4e20-85bd-ee1502bb5463 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:03:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask beffba80-d510-4944-9e2b-b4ff3c4fc669 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:03:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ac05741a-1939-415f-aeff-a7eed21049c9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:03:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6fd9e63e-df8b-4c9f-86ca-d2316078b744 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:03:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44aa611b-f898-4352-b174-ada638d8f758 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:03:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 01:03:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f714ea82-7678-4e20-85bd-ee1502bb5463 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:03:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask beffba80-d510-4944-9e2b-b4ff3c4fc669 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:03:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ac05741a-1939-415f-aeff-a7eed21049c9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:03:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6fd9e63e-df8b-4c9f-86ca-d2316078b744 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:03:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44aa611b-f898-4352-b174-ada638d8f758 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:03:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 01:03:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f714ea82-7678-4e20-85bd-ee1502bb5463 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:03:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask beffba80-d510-4944-9e2b-b4ff3c4fc669 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:03:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ac05741a-1939-415f-aeff-a7eed21049c9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:03:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6fd9e63e-df8b-4c9f-86ca-d2316078b744 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:03:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44aa611b-f898-4352-b174-ada638d8f758 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:03:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 01:03:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f714ea82-7678-4e20-85bd-ee1502bb5463 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:03:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask beffba80-d510-4944-9e2b-b4ff3c4fc669 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:03:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ac05741a-1939-415f-aeff-a7eed21049c9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:03:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6fd9e63e-df8b-4c9f-86ca-d2316078b744 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:03:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44aa611b-f898-4352-b174-ada638d8f758 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:03:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 01:03:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f714ea82-7678-4e20-85bd-ee1502bb5463 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:03:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask beffba80-d510-4944-9e2b-b4ff3c4fc669 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:03:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ac05741a-1939-415f-aeff-a7eed21049c9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:03:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6fd9e63e-df8b-4c9f-86ca-d2316078b744 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:03:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44aa611b-f898-4352-b174-ada638d8f758 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:03:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 01:04:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f714ea82-7678-4e20-85bd-ee1502bb5463 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:04:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask beffba80-d510-4944-9e2b-b4ff3c4fc669 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:04:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ac05741a-1939-415f-aeff-a7eed21049c9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:04:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6fd9e63e-df8b-4c9f-86ca-d2316078b744 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:04:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44aa611b-f898-4352-b174-ada638d8f758 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:04:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 01:04:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f714ea82-7678-4e20-85bd-ee1502bb5463 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:04:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask beffba80-d510-4944-9e2b-b4ff3c4fc669 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:04:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ac05741a-1939-415f-aeff-a7eed21049c9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:04:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6fd9e63e-df8b-4c9f-86ca-d2316078b744 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:04:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44aa611b-f898-4352-b174-ada638d8f758 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:04:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 01:04:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f714ea82-7678-4e20-85bd-ee1502bb5463 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:04:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask beffba80-d510-4944-9e2b-b4ff3c4fc669 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:04:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask ac05741a-1939-415f-aeff-a7eed21049c9 is in state SUCCESS\u001b[0m",
                                "\u001b[32m2026-02-24 01:04:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6fd9e63e-df8b-4c9f-86ca-d2316078b744 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:04:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44aa611b-f898-4352-b174-ada638d8f758 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:04:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3046662f-74e4-408a-bb14-4c7f7919dd4b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:04:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 01:04:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f714ea82-7678-4e20-85bd-ee1502bb5463 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:04:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask beffba80-d510-4944-9e2b-b4ff3c4fc669 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:04:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6fd9e63e-df8b-4c9f-86ca-d2316078b744 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:04:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44aa611b-f898-4352-b174-ada638d8f758 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:04:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3046662f-74e4-408a-bb14-4c7f7919dd4b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:04:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 01:04:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f714ea82-7678-4e20-85bd-ee1502bb5463 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:04:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask beffba80-d510-4944-9e2b-b4ff3c4fc669 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:04:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6fd9e63e-df8b-4c9f-86ca-d2316078b744 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:04:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44aa611b-f898-4352-b174-ada638d8f758 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:04:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3046662f-74e4-408a-bb14-4c7f7919dd4b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:04:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 01:04:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f714ea82-7678-4e20-85bd-ee1502bb5463 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:04:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask beffba80-d510-4944-9e2b-b4ff3c4fc669 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:04:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6fd9e63e-df8b-4c9f-86ca-d2316078b744 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:04:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44aa611b-f898-4352-b174-ada638d8f758 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:04:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3046662f-74e4-408a-bb14-4c7f7919dd4b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:04:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 01:04:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f714ea82-7678-4e20-85bd-ee1502bb5463 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:04:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask beffba80-d510-4944-9e2b-b4ff3c4fc669 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:04:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6fd9e63e-df8b-4c9f-86ca-d2316078b744 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:04:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44aa611b-f898-4352-b174-ada638d8f758 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:04:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3046662f-74e4-408a-bb14-4c7f7919dd4b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:04:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 01:04:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f714ea82-7678-4e20-85bd-ee1502bb5463 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:04:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask beffba80-d510-4944-9e2b-b4ff3c4fc669 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:04:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6fd9e63e-df8b-4c9f-86ca-d2316078b744 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:04:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44aa611b-f898-4352-b174-ada638d8f758 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:04:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3046662f-74e4-408a-bb14-4c7f7919dd4b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:04:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 01:04:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f714ea82-7678-4e20-85bd-ee1502bb5463 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:04:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask beffba80-d510-4944-9e2b-b4ff3c4fc669 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:04:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6fd9e63e-df8b-4c9f-86ca-d2316078b744 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:04:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44aa611b-f898-4352-b174-ada638d8f758 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:04:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3046662f-74e4-408a-bb14-4c7f7919dd4b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:04:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 01:04:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f714ea82-7678-4e20-85bd-ee1502bb5463 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:04:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask beffba80-d510-4944-9e2b-b4ff3c4fc669 is in state STARTED\u001b[0m",
                                "",
                                "",
                                "PLAY [Apply role cephclient] ***************************************************",
                                "",
                                "TASK [osism.services.cephclient : Include container tasks] *********************",
                                "Tuesday 24 February 2026  01:01:54 +0000 (0:00:00.271)       0:00:00.271 ****** ",
                                "\u001b[0;36mincluded: /usr/share/ansible/collections/ansible_collections/osism/services/roles/cephclient/tasks/container.yml for testbed-manager\u001b[0m",
                                "",
                                "TASK [osism.services.cephclient : Create required directories] *****************",
                                "Tuesday 24 February 2026  01:01:54 +0000 (0:00:00.238)       0:00:00.510 ****** ",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=/opt/cephclient/configuration)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=/opt/cephclient/data)\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager] => (item=/opt/cephclient)\u001b[0m",
                                "",
                                "TASK [osism.services.cephclient : Copy configuration files] ********************",
                                "Tuesday 24 February 2026  01:01:56 +0000 (0:00:01.377)       0:00:01.887 ****** ",
                                "\u001b[0;33mchanged: [testbed-manager] => (item={'src': 'ceph.conf.j2', 'dest': '/opt/cephclient/configuration/ceph.conf'})\u001b[0m",
                                "",
                                "TASK [osism.services.cephclient : Copy keyring file] ***************************",
                                "Tuesday 24 February 2026  01:01:57 +0000 (0:00:01.635)       0:00:03.523 ****** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.services.cephclient : Copy docker-compose.yml file] ****************",
                                "Tuesday 24 February 2026  01:01:58 +0000 (0:00:00.993)       0:00:04.516 ****** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.services.cephclient : Manage cephclient service] *******************",
                                "Tuesday 24 February 2026  01:01:59 +0000 (0:00:01.028)       0:00:05.545 ****** ",
                                "\u001b[1;30mFAILED - RETRYING: [testbed-manager]: Manage cephclient service (10 retries left).\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.services.cephclient : Copy wrapper scripts] ************************",
                                "Tuesday 24 February 2026  01:02:42 +0000 (0:00:42.879)       0:00:48.424 ****** ",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=ceph)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=ceph-authtool)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=rados)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=radosgw-admin)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=rbd)\u001b[0m",
                                "",
                                "TASK [osism.services.cephclient : Remove old wrapper scripts] ******************",
                                "Tuesday 24 February 2026  01:02:47 +0000 (0:00:04.646)       0:00:53.071 ****** ",
                                "\u001b[0;32mok: [testbed-manager] => (item=crushtool)\u001b[0m",
                                "",
                                "TASK [osism.services.cephclient : Include package tasks] ***********************",
                                "Tuesday 24 February 2026  01:02:47 +0000 (0:00:00.511)       0:00:53.583 ****** ",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.services.cephclient : Include rook task] ***************************",
                                "Tuesday 24 February 2026  01:02:47 +0000 (0:00:00.144)       0:00:53.727 ****** ",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "",
                                "RUNNING HANDLER [osism.services.cephclient : Restart cephclient service] *******",
                                "Tuesday 24 February 2026  01:02:48 +0000 (0:00:00.596)       0:00:54.324 ****** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "RUNNING HANDLER [osism.services.cephclient : Ensure that all containers are up] ***",
                                "Tuesday 24 February 2026  01:02:50 +0000 (0:00:01.671)       0:00:55.996 ****** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "RUNNING HANDLER [osism.services.cephclient : Wait for an healthy service] ******",
                                "Tuesday 24 February 2026  01:02:51 +0000 (0:00:00.842)       0:00:56.838 ****** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "RUNNING HANDLER [osism.services.cephclient : Copy bash completion scripts] *****",
                                "Tuesday 24 February 2026  01:02:51 +0000 (0:00:00.712)       0:00:57.551 ****** ",
                                "\u001b[0;32mok: [testbed-manager] => (item=ceph)\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager] => (item=rados)\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager] => (item=radosgw-admin)\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager] => (item=rbd)\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;33mtestbed-manager\u001b[0m            : \u001b[0;32mok=12  \u001b[0m \u001b[0;33mchanged=8   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Tuesday 24 February 2026  01:02:53 +0000 (0:00:01.707)       0:00:59.258 ****** ",
                                "=============================================================================== ",
                                "osism.services.cephclient : Manage cephclient service ------------------ 42.88s",
                                "osism.services.cephclient : Copy wrapper scripts ------------------------ 4.65s",
                                "osism.services.cephclient : Copy bash completion scripts ---------------- 1.71s",
                                "osism.services.cephclient : Restart cephclient service ------------------ 1.67s",
                                "osism.services.cephclient : Copy configuration files -------------------- 1.64s",
                                "osism.services.cephclient : Create required directories ----------------- 1.38s",
                                "osism.services.cephclient : Copy docker-compose.yml file ---------------- 1.03s",
                                "osism.services.cephclient : Copy keyring file --------------------------- 0.99s",
                                "osism.services.cephclient : Ensure that all containers are up ----------- 0.84s",
                                "osism.services.cephclient : Wait for an healthy service ----------------- 0.71s",
                                "osism.services.cephclient : Include rook task --------------------------- 0.60s",
                                "osism.services.cephclient : Remove old wrapper scripts ------------------ 0.51s",
                                "osism.services.cephclient : Include container tasks --------------------- 0.24s",
                                "osism.services.cephclient : Include package tasks ----------------------- 0.14s",
                                "",
                                "",
                                "PLAY [Download ironic ipa images] **********************************************",
                                "",
                                "TASK [Ensure the destination directory exists] *********************************",
                                "Tuesday 24 February 2026  01:02:32 +0000 (0:00:00.086)       0:00:00.086 ****** ",
                                "\u001b[0;33mchanged: [localhost]\u001b[0m",
                                "",
                                "TASK [Download ironic-agent initramfs] *****************************************",
                                "Tuesday 24 February 2026  01:02:32 +0000 (0:00:00.851)       0:00:00.938 ****** ",
                                "\u001b[1;30mFAILED - RETRYING: [localhost]: Download ironic-agent initramfs (3 retries left).\u001b[0m",
                                "\u001b[1;30mFAILED - RETRYING: [localhost]: Download ironic-agent initramfs (2 retries left).\u001b[0m",
                                "\u001b[1;30mFAILED - RETRYING: [localhost]: Download ironic-agent initramfs (1 retries left).\u001b[0m",
                                "\u001b[0;31mfatal: [localhost]: FAILED! => {\"attempts\": 3, \"changed\": false, \"dest\": \"/share/ironic/ironic/ironic-agent.initramfs\", \"elapsed\": 10, \"msg\": \"Request failed: <urlopen error [Errno 101] Network is unreachable>\", \"url\": \"https://tarballs.opendev.org/openstack/ironic-python-agent/dib/files/ipa-centos9-stable-2024.2.initramfs.sha256\"}\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;31mlocalhost\u001b[0m                  : \u001b[0;32mok=1   \u001b[0m \u001b[0;33mchanged=1   \u001b[0m unreachable=0    \u001b[0;31mfailed=1   \u001b[0m skipped=0    rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Tuesday 24 February 2026  01:04:07 +0000 (0:01:34.415)       0:01:35.354 ****** ",
                                "=============================================================================== ",
                                "Download ironic-agent initramfs ---------------------------------------- 94.42s",
                                "Ensure the destination directory exists --------------------------------- 0.85s",
                                "",
                                "\u001b[1;35m[WARNING]: Collection community.general does not support Ansible version\u001b[0m",
                                "\u001b[1;35m2.16.14\u001b[0m",
                                "",
                                "PLAY [Bootstraph ceph dashboard] ***********************************************",
                                "",
                                "TASK [Disable the ceph dashboard] **********************************************",
                                "Tuesday 24 February 2026  01:02:58 +0000 (0:00:00.279)       0:00:00.279 ****** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [Set mgr/dashboard/ssl to false] ******************************************",
                                "Tuesday 24 February 2026  01:03:00 +0000 (0:00:01.654)       0:00:01.934 ****** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [Set mgr/dashboard/server_port to 7000] ***********************************",
                                "Tuesday 24 February 2026  01:03:01 +0000 (0:00:01.214)       0:00:03.149 ****** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [Set mgr/dashboard/server_addr to 0.0.0.0] ********************************",
                                "Tuesday 24 February 2026  01:03:02 +0000 (0:00:01.129)       0:00:04.279 ****** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [Set mgr/dashboard/standby_behaviour to error] ****************************",
                                "Tuesday 24 February 2026  01:03:04 +0000 (0:00:01.368)       0:00:05.647 ****** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [Set mgr/dashboard/standby_error_status_code to 404] **********************",
                                "Tuesday 24 February 2026  01:03:05 +0000 (0:00:01.191)       0:00:06.839 ****** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [Enable the ceph dashboard] ***********************************************",
                                "Tuesday 24 February 2026  01:03:06 +0000 (0:00:01.198)       0:00:08.038 ****** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [Write ceph_dashboard_password to temporary file] *************************",
                                "Tuesday 24 February 2026  01:03:08 +0000 (0:00:02.110)       0:00:10.148 ****** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [Create admin user] *******************************************************",
                                "Tuesday 24 February 2026  01:03:10 +0000 (0:00:01.431)       0:00:11.580 ****** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [Remove temporary file for ceph_dashboard_password] ***********************",
                                "Tuesday 24 February 2026  01:04:05 +0000 (0:00:54.942)       0:01:06.522 ****** ",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "",
                                "PLAY [Restart ceph manager services] *******************************************",
                                "",
                                "TASK [Restart ceph manager service] ********************************************",
                                "Tuesday 24 February 2026  01:04:05 +0000 (0:00:00.157)       0:01:06.680 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "PLAY [Restart ceph manager services] *******************************************",
                                "",
                                "TASK [Restart ceph manager service] ********************************************",
                                "Tuesday 24 February 2026  01:04:17 +0000 (0:00:11.863)       0:01:18.544 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "",
                                "PLAY [Restart ceph manager services] *******************************************",
                                "",
                                "TASK [Restart ceph manager service] ********************************************",
                                "Tuesday 24 February 2026  01:04:28 +0000 (0:00:11.288)       0:01:29.832 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;33mtestbed-manager\u001b[0m            : \u001b[0;32mok=9   \u001b[0m \u001b[0;33mchanged=9   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=1   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=1   \u001b[0m \u001b[0;33mchanged=1   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=1   \u001b[0m \u001b[0;33mchanged=1   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=1   \u001b[0m \u001b[0;33mchanged=1   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Tuesday 24 February 2026  01:04:29 +0000 (0:00:01.125)       0:01:30.957 ****** ",
                                "=============================================================================== ",
                                "Create admin user ------------------------------------------------------ 54.94s",
                                "Restart ceph manager service ------------------------------------------- 24.28s",
                                "Enable the ceph dashboard ----------------------------------------------- 2.11s",
                                "Disable the ceph dashboard ---------------------------------------------- 1.65s",
                                "Write ceph_dashboard_password to temporary file ------------------------- 1.43s",
                                "Set mgr/dashboard/server_addr to 0.0.0.0 -------------------------------- 1.37s",
                                "Set mgr/dashboard/ssl to false ------------------------------------------ 1.21s",
                                "Set mgr/dashboard/standby_error_status_code to 404 ---------------------- 1.20s",
                                "Set mgr/dashboard/standby_behaviour to error ---------------------------- 1.19s",
                                "Set mgr/dashboard/server_port to 7000 ----------------------------------- 1.13s",
                                "Remove temporary file for ceph_dashboard_password ----------------------- 0.16s",
                                "\u001b[32m2026-02-24 01:04:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 6fd9e63e-df8b-4c9f-86ca-d2316078b744 is in state SUCCESS\u001b[0m",
                                "\u001b[32m2026-02-24 01:04:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44aa611b-f898-4352-b174-ada638d8f758 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:04:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3046662f-74e4-408a-bb14-4c7f7919dd4b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:04:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 01:04:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f714ea82-7678-4e20-85bd-ee1502bb5463 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:04:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask beffba80-d510-4944-9e2b-b4ff3c4fc669 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:04:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44aa611b-f898-4352-b174-ada638d8f758 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:04:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3046662f-74e4-408a-bb14-4c7f7919dd4b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:04:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 01:04:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f714ea82-7678-4e20-85bd-ee1502bb5463 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:04:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask beffba80-d510-4944-9e2b-b4ff3c4fc669 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:04:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44aa611b-f898-4352-b174-ada638d8f758 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:04:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3046662f-74e4-408a-bb14-4c7f7919dd4b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:04:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 01:04:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f714ea82-7678-4e20-85bd-ee1502bb5463 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:04:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask beffba80-d510-4944-9e2b-b4ff3c4fc669 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:04:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44aa611b-f898-4352-b174-ada638d8f758 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:04:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3046662f-74e4-408a-bb14-4c7f7919dd4b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:04:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 01:04:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f714ea82-7678-4e20-85bd-ee1502bb5463 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:04:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask beffba80-d510-4944-9e2b-b4ff3c4fc669 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:04:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44aa611b-f898-4352-b174-ada638d8f758 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:04:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3046662f-74e4-408a-bb14-4c7f7919dd4b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:04:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 01:04:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f714ea82-7678-4e20-85bd-ee1502bb5463 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:04:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask beffba80-d510-4944-9e2b-b4ff3c4fc669 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:04:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44aa611b-f898-4352-b174-ada638d8f758 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:04:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3046662f-74e4-408a-bb14-4c7f7919dd4b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:04:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 01:04:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f714ea82-7678-4e20-85bd-ee1502bb5463 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:04:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask beffba80-d510-4944-9e2b-b4ff3c4fc669 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:04:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44aa611b-f898-4352-b174-ada638d8f758 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:04:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3046662f-74e4-408a-bb14-4c7f7919dd4b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:04:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 01:04:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f714ea82-7678-4e20-85bd-ee1502bb5463 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:04:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask beffba80-d510-4944-9e2b-b4ff3c4fc669 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:04:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44aa611b-f898-4352-b174-ada638d8f758 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:04:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3046662f-74e4-408a-bb14-4c7f7919dd4b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:04:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 01:04:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f714ea82-7678-4e20-85bd-ee1502bb5463 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:04:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask beffba80-d510-4944-9e2b-b4ff3c4fc669 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:04:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44aa611b-f898-4352-b174-ada638d8f758 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:04:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3046662f-74e4-408a-bb14-4c7f7919dd4b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:04:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 01:04:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f714ea82-7678-4e20-85bd-ee1502bb5463 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:04:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask beffba80-d510-4944-9e2b-b4ff3c4fc669 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:04:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44aa611b-f898-4352-b174-ada638d8f758 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:04:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3046662f-74e4-408a-bb14-4c7f7919dd4b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:04:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 01:05:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f714ea82-7678-4e20-85bd-ee1502bb5463 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:05:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask beffba80-d510-4944-9e2b-b4ff3c4fc669 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:05:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44aa611b-f898-4352-b174-ada638d8f758 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:05:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3046662f-74e4-408a-bb14-4c7f7919dd4b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:05:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 01:05:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f714ea82-7678-4e20-85bd-ee1502bb5463 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:05:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask beffba80-d510-4944-9e2b-b4ff3c4fc669 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:05:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44aa611b-f898-4352-b174-ada638d8f758 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:05:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3046662f-74e4-408a-bb14-4c7f7919dd4b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:05:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 01:05:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f714ea82-7678-4e20-85bd-ee1502bb5463 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:05:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask beffba80-d510-4944-9e2b-b4ff3c4fc669 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:05:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44aa611b-f898-4352-b174-ada638d8f758 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:05:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3046662f-74e4-408a-bb14-4c7f7919dd4b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:05:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 01:05:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f714ea82-7678-4e20-85bd-ee1502bb5463 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:05:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask beffba80-d510-4944-9e2b-b4ff3c4fc669 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:05:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44aa611b-f898-4352-b174-ada638d8f758 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:05:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3046662f-74e4-408a-bb14-4c7f7919dd4b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:05:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 01:05:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f714ea82-7678-4e20-85bd-ee1502bb5463 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:05:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask beffba80-d510-4944-9e2b-b4ff3c4fc669 is in state SUCCESS\u001b[0m",
                                "",
                                "",
                                "PLAY [Group hosts based on configuration] **************************************",
                                "",
                                "TASK [Group hosts based on Kolla action] ***************************************",
                                "Tuesday 24 February 2026  01:02:32 +0000 (0:00:00.378)       0:00:00.378 ****** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [Group hosts based on enabled services] ***********************************",
                                "Tuesday 24 February 2026  01:02:32 +0000 (0:00:00.320)       0:00:00.698 ****** ",
                                "\u001b[0;32mok: [testbed-node-0] => (item=enable_barbican_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item=enable_barbican_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item=enable_barbican_True)\u001b[0m",
                                "",
                                "PLAY [Apply role barbican] *****************************************************",
                                "",
                                "TASK [barbican : include_tasks] ************************************************",
                                "Tuesday 24 February 2026  01:02:33 +0000 (0:00:00.503)       0:00:01.202 ****** ",
                                "\u001b[0;36mincluded: /ansible/roles/barbican/tasks/deploy.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [service-ks-register : barbican | Creating services] **********************",
                                "Tuesday 24 February 2026  01:02:33 +0000 (0:00:00.522)       0:00:01.725 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=barbican (key-manager))\u001b[0m",
                                "",
                                "TASK [service-ks-register : barbican | Creating endpoints] *********************",
                                "Tuesday 24 February 2026  01:02:37 +0000 (0:00:03.782)       0:00:05.507 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=barbican -> https://api-int.testbed.osism.xyz:9311 -> internal)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=barbican -> https://api.testbed.osism.xyz:9311 -> public)\u001b[0m",
                                "",
                                "TASK [service-ks-register : barbican | Creating projects] **********************",
                                "Tuesday 24 February 2026  01:02:46 +0000 (0:00:08.751)       0:00:14.258 ****** ",
                                "\u001b[1;30mFAILED - RETRYING: [testbed-node-0]: barbican | Creating projects (5 retries left).\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0] => (item=service)\u001b[0m",
                                "",
                                "TASK [service-ks-register : barbican | Creating users] *************************",
                                "Tuesday 24 February 2026  01:03:04 +0000 (0:00:18.156)       0:00:32.415 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=barbican -> service)\u001b[0m",
                                "\u001b[1;35m[WARNING]: Module did not set no_log for update_password\u001b[0m",
                                "",
                                "TASK [service-ks-register : barbican | Creating roles] *************************",
                                "Tuesday 24 February 2026  01:03:09 +0000 (0:00:04.862)       0:00:37.278 ****** ",
                                "\u001b[0;32mok: [testbed-node-0] => (item=admin)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=key-manager:service-admin)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=creator)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=observer)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=audit)\u001b[0m",
                                "",
                                "TASK [service-ks-register : barbican | Granting user roles] ********************",
                                "Tuesday 24 February 2026  01:03:26 +0000 (0:00:16.594)       0:00:53.873 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=barbican -> service -> admin)\u001b[0m",
                                "",
                                "TASK [barbican : Ensuring config directories exist] ****************************",
                                "Tuesday 24 February 2026  01:03:31 +0000 (0:00:05.327)       0:00:59.200 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-api:2024.2', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-api:2024.2', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-api:2024.2', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-keystone-listener:2024.2', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-keystone-listener:2024.2', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-keystone-listener:2024.2', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-worker:2024.2', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-worker:2024.2', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-worker:2024.2', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}})\u001b[0m",
                                "",
                                "TASK [barbican : Ensuring vassals config directories exist] ********************",
                                "Tuesday 24 February 2026  01:03:34 +0000 (0:00:03.309)       0:01:02.510 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=barbican-api/vassals)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=barbican-api/vassals)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=barbican-api/vassals)\u001b[0m",
                                "",
                                "TASK [barbican : Check if policies shall be overwritten] ***********************",
                                "Tuesday 24 February 2026  01:03:36 +0000 (0:00:01.719)       0:01:04.229 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [barbican : Set barbican policy file] *************************************",
                                "Tuesday 24 February 2026  01:03:36 +0000 (0:00:00.472)       0:01:04.702 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [barbican : include_tasks] ************************************************",
                                "Tuesday 24 February 2026  01:03:38 +0000 (0:00:01.433)       0:01:06.135 ****** ",
                                "\u001b[0;36mincluded: /ansible/roles/barbican/tasks/copy-certs.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [service-cert-copy : barbican | Copying over extra CA certificates] *******",
                                "Tuesday 24 February 2026  01:03:38 +0000 (0:00:00.646)       0:01:06.782 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-api:2024.2', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-api:2024.2', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-api:2024.2', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-keystone-listener:2024.2', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-keystone-listener:2024.2', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-keystone-listener:2024.2', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-worker:2024.2', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-worker:2024.2', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-worker:2024.2', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}})\u001b[0m",
                                "",
                                "TASK [service-cert-copy : barbican | Copying over backend internal TLS certificate] ***",
                                "Tuesday 24 February 2026  01:03:43 +0000 (0:00:04.227)       0:01:11.009 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-api:2024.2', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-keystone-listener:2024.2', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-worker:2024.2', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-api:2024.2', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-keystone-listener:2024.2', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-worker:2024.2', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-api:2024.2', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-keystone-listener:2024.2', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-worker:2024.2', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [service-cert-copy : barbican | Copying over backend internal TLS key] ****",
                                "Tuesday 24 February 2026  01:03:44 +0000 (0:00:01.494)       0:01:12.503 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-api:2024.2', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-keystone-listener:2024.2', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-worker:2024.2', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-api:2024.2', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-keystone-listener:2024.2', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-worker:2024.2', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-api:2024.2', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-keystone-listener:2024.2', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-worker:2024.2', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [barbican : Copying over config.json files for services] ******************",
                                "Tuesday 24 February 2026  01:03:46 +0000 (0:00:02.048)       0:01:14.551 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-api:2024.2', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-api:2024.2', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-api:2024.2', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-keystone-listener:2024.2', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-keystone-listener:2024.2', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-keystone-listener:2024.2', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-worker:2024.2', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-worker:2024.2', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-worker:2024.2', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}})\u001b[0m",
                                "",
                                "TASK [barbican : Copying over barbican-api.ini] ********************************",
                                "Tuesday 24 February 2026  01:03:50 +0000 (0:00:03.899)       0:01:18.451 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [barbican : Checking whether barbican-api-paste.ini file exists] **********",
                                "Tuesday 24 February 2026  01:03:54 +0000 (0:00:03.470)       0:01:21.922 ****** ",
                                "\u001b[0;32mok: [testbed-node-0 -> localhost]\u001b[0m",
                                "",
                                "TASK [barbican : Copying over barbican-api-paste.ini] **************************",
                                "Tuesday 24 February 2026  01:03:56 +0000 (0:00:02.912)       0:01:24.835 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [barbican : Copying over barbican.conf] ***********************************",
                                "Tuesday 24 February 2026  01:03:58 +0000 (0:00:01.175)       0:01:26.010 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-api:2024.2', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-api:2024.2', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-api:2024.2', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-keystone-listener:2024.2', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-keystone-listener:2024.2', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-keystone-listener:2024.2', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-worker:2024.2', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-worker:2024.2', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-worker:2024.2', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}})\u001b[0m",
                                "",
                                "TASK [barbican : Copying over existing policy file] ****************************",
                                "Tuesday 24 February 2026  01:04:07 +0000 (0:00:09.742)       0:01:35.752 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-api:2024.2', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-keystone-listener:2024.2', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-worker:2024.2', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-api:2024.2', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-keystone-listener:2024.2', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-worker:2024.2', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-api:2024.2', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-keystone-listener:2024.2', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-worker:2024.2', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [barbican : Check barbican containers] ************************************",
                                "Tuesday 24 February 2026  01:04:10 +0000 (0:00:02.225)       0:01:37.978 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-api:2024.2', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-api:2024.2', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'barbican-api', 'value': {'container_name': 'barbican_api', 'group': 'barbican-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-api:2024.2', 'volumes': ['/etc/kolla/barbican-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'barbican:/var/lib/barbican/', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9311'], 'timeout': '30'}, 'haproxy': {'barbican_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}, 'barbican_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9311', 'listen_port': '9311', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-keystone-listener:2024.2', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-keystone-listener:2024.2', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'barbican-keystone-listener', 'value': {'container_name': 'barbican_keystone_listener', 'group': 'barbican-keystone-listener', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-keystone-listener:2024.2', 'volumes': ['/etc/kolla/barbican-keystone-listener/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-keystone-listener 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-worker:2024.2', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-worker:2024.2', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'barbican-worker', 'value': {'container_name': 'barbican_worker', 'group': 'barbican-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/barbican-worker:2024.2', 'volumes': ['/etc/kolla/barbican-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port barbican-worker 5672'], 'timeout': '30'}}})\u001b[0m",
                                "",
                                "TASK [barbican : include_tasks] ************************************************",
                                "Tuesday 24 February 2026  01:04:16 +0000 (0:00:06.884)       0:01:44.863 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [barbican : Creating barbican database] ***********************************",
                                "Tuesday 24 February 2026  01:04:18 +0000 (0:00:01.027)       0:01:45.890 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [barbican : Creating barbican database user and setting permissions] ******",
                                "Tuesday 24 February 2026  01:04:20 +0000 (0:00:02.964)       0:01:48.855 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [barbican : Running barbican bootstrap container] *************************",
                                "Tuesday 24 February 2026  01:04:24 +0000 (0:00:03.481)       0:01:52.337 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [barbican : Flush handlers] ***********************************************",
                                "Tuesday 24 February 2026  01:04:37 +0000 (0:00:12.888)       0:02:05.225 ****** ",
                                "",
                                "TASK [barbican : Flush handlers] ***********************************************",
                                "Tuesday 24 February 2026  01:04:37 +0000 (0:00:00.152)       0:02:05.378 ****** ",
                                "",
                                "TASK [barbican : Flush handlers] ***********************************************",
                                "Tuesday 24 February 2026  01:04:37 +0000 (0:00:00.125)       0:02:05.503 ****** ",
                                "",
                                "RUNNING HANDLER [barbican : Restart barbican-api container] ********************",
                                "Tuesday 24 February 2026  01:04:37 +0000 (0:00:00.067)       0:02:05.570 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [barbican : Restart barbican-keystone-listener container] ******",
                                "Tuesday 24 February 2026  01:04:51 +0000 (0:00:13.707)       0:02:19.278 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "",
                                "RUNNING HANDLER [barbican : Restart barbican-worker container] *****************",
                                "Tuesday 24 February 2026  01:05:01 +0000 (0:00:10.325)       0:02:29.604 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=24  \u001b[0m \u001b[0;33mchanged=18  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=7   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=14  \u001b[0m \u001b[0;33mchanged=10  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=6   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=14  \u001b[0m \u001b[0;33mchanged=10  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=6   \u001b[0m rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Tuesday 24 February 2026  01:05:10 +0000 (0:00:08.933)       0:02:38.537 ****** ",
                                "=============================================================================== ",
                                "service-ks-register : barbican | Creating projects --------------------- 18.16s",
                                "service-ks-register : barbican | Creating roles ------------------------ 16.59s",
                                "barbican : Restart barbican-api container ------------------------------ 13.71s",
                                "barbican : Running barbican bootstrap container ------------------------ 12.89s",
                                "barbican : Restart barbican-keystone-listener container ---------------- 10.33s",
                                "barbican : Copying over barbican.conf ----------------------------------- 9.74s",
                                "barbican : Restart barbican-worker container ---------------------------- 8.93s",
                                "service-ks-register : barbican | Creating endpoints --------------------- 8.75s",
                                "barbican : Check barbican containers ------------------------------------ 6.88s",
                                "service-ks-register : barbican | Granting user roles -------------------- 5.33s",
                                "service-ks-register : barbican | Creating users ------------------------- 4.86s",
                                "service-cert-copy : barbican | Copying over extra CA certificates ------- 4.23s",
                                "barbican : Copying over config.json files for services ------------------ 3.90s",
                                "service-ks-register : barbican | Creating services ---------------------- 3.78s",
                                "barbican : Creating barbican database user and setting permissions ------ 3.48s",
                                "barbican : Copying over barbican-api.ini -------------------------------- 3.47s",
                                "barbican : Ensuring config directories exist ---------------------------- 3.31s",
                                "barbican : Creating barbican database ----------------------------------- 2.96s",
                                "barbican : Checking whether barbican-api-paste.ini file exists ---------- 2.91s",
                                "barbican : Copying over existing policy file ---------------------------- 2.23s",
                                "\u001b[32m2026-02-24 01:05:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a8ebcda5-cdc0-4611-9f70-ca910ead0481 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:05:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44aa611b-f898-4352-b174-ada638d8f758 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:05:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3046662f-74e4-408a-bb14-4c7f7919dd4b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:05:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 01:05:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f714ea82-7678-4e20-85bd-ee1502bb5463 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:05:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a8ebcda5-cdc0-4611-9f70-ca910ead0481 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:05:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44aa611b-f898-4352-b174-ada638d8f758 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:05:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3046662f-74e4-408a-bb14-4c7f7919dd4b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:05:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 01:05:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f714ea82-7678-4e20-85bd-ee1502bb5463 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:05:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a8ebcda5-cdc0-4611-9f70-ca910ead0481 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:05:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44aa611b-f898-4352-b174-ada638d8f758 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:05:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3046662f-74e4-408a-bb14-4c7f7919dd4b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:05:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 01:05:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f714ea82-7678-4e20-85bd-ee1502bb5463 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:05:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a8ebcda5-cdc0-4611-9f70-ca910ead0481 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:05:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44aa611b-f898-4352-b174-ada638d8f758 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:05:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3046662f-74e4-408a-bb14-4c7f7919dd4b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:05:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 01:05:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f714ea82-7678-4e20-85bd-ee1502bb5463 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:05:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a8ebcda5-cdc0-4611-9f70-ca910ead0481 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:05:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44aa611b-f898-4352-b174-ada638d8f758 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:05:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3046662f-74e4-408a-bb14-4c7f7919dd4b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:05:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 01:05:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f714ea82-7678-4e20-85bd-ee1502bb5463 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:05:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a8ebcda5-cdc0-4611-9f70-ca910ead0481 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:05:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44aa611b-f898-4352-b174-ada638d8f758 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:05:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3046662f-74e4-408a-bb14-4c7f7919dd4b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:05:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 01:05:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f714ea82-7678-4e20-85bd-ee1502bb5463 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:05:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a8ebcda5-cdc0-4611-9f70-ca910ead0481 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:05:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44aa611b-f898-4352-b174-ada638d8f758 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:05:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3046662f-74e4-408a-bb14-4c7f7919dd4b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:05:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 01:05:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f714ea82-7678-4e20-85bd-ee1502bb5463 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:05:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a8ebcda5-cdc0-4611-9f70-ca910ead0481 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:05:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44aa611b-f898-4352-b174-ada638d8f758 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:05:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3046662f-74e4-408a-bb14-4c7f7919dd4b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:05:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 01:05:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f714ea82-7678-4e20-85bd-ee1502bb5463 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:05:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a8ebcda5-cdc0-4611-9f70-ca910ead0481 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:05:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44aa611b-f898-4352-b174-ada638d8f758 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:05:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3046662f-74e4-408a-bb14-4c7f7919dd4b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:05:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 01:05:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f714ea82-7678-4e20-85bd-ee1502bb5463 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:05:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a8ebcda5-cdc0-4611-9f70-ca910ead0481 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:05:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44aa611b-f898-4352-b174-ada638d8f758 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:05:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3046662f-74e4-408a-bb14-4c7f7919dd4b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:05:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 01:05:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f714ea82-7678-4e20-85bd-ee1502bb5463 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:05:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a8ebcda5-cdc0-4611-9f70-ca910ead0481 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:05:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44aa611b-f898-4352-b174-ada638d8f758 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:05:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3046662f-74e4-408a-bb14-4c7f7919dd4b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:05:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 01:05:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f714ea82-7678-4e20-85bd-ee1502bb5463 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:05:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a8ebcda5-cdc0-4611-9f70-ca910ead0481 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:05:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44aa611b-f898-4352-b174-ada638d8f758 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:05:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3046662f-74e4-408a-bb14-4c7f7919dd4b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:05:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 01:05:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f714ea82-7678-4e20-85bd-ee1502bb5463 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:05:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a8ebcda5-cdc0-4611-9f70-ca910ead0481 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:05:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44aa611b-f898-4352-b174-ada638d8f758 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:05:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3046662f-74e4-408a-bb14-4c7f7919dd4b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:05:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 01:05:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f714ea82-7678-4e20-85bd-ee1502bb5463 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:05:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a8ebcda5-cdc0-4611-9f70-ca910ead0481 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:05:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44aa611b-f898-4352-b174-ada638d8f758 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:05:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3046662f-74e4-408a-bb14-4c7f7919dd4b is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:05:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 01:05:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f714ea82-7678-4e20-85bd-ee1502bb5463 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:05:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a8ebcda5-cdc0-4611-9f70-ca910ead0481 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:05:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 47bc2a2f-3681-4d97-beda-759966d298ca is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:05:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44aa611b-f898-4352-b174-ada638d8f758 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:05:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 3046662f-74e4-408a-bb14-4c7f7919dd4b is in state SUCCESS\u001b[0m",
                                "\u001b[32m2026-02-24 01:05:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "",
                                "",
                                "PLAY [Group hosts based on configuration] **************************************",
                                "",
                                "TASK [Group hosts based on Kolla action] ***************************************",
                                "Tuesday 24 February 2026  01:04:19 +0000 (0:00:00.339)       0:00:00.339 ****** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [Group hosts based on enabled services] ***********************************",
                                "Tuesday 24 February 2026  01:04:19 +0000 (0:00:00.566)       0:00:00.906 ****** ",
                                "\u001b[0;32mok: [testbed-node-0] => (item=enable_placement_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item=enable_placement_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item=enable_placement_True)\u001b[0m",
                                "",
                                "PLAY [Apply role placement] ****************************************************",
                                "",
                                "TASK [placement : include_tasks] ***********************************************",
                                "Tuesday 24 February 2026  01:04:21 +0000 (0:00:01.650)       0:00:02.556 ****** ",
                                "\u001b[0;36mincluded: /ansible/roles/placement/tasks/deploy.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [service-ks-register : placement | Creating services] *********************",
                                "Tuesday 24 February 2026  01:04:22 +0000 (0:00:00.951)       0:00:03.508 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=placement (placement))\u001b[0m",
                                "",
                                "TASK [service-ks-register : placement | Creating endpoints] ********************",
                                "Tuesday 24 February 2026  01:04:26 +0000 (0:00:04.198)       0:00:07.707 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=placement -> https://api-int.testbed.osism.xyz:8780 -> internal)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=placement -> https://api.testbed.osism.xyz:8780 -> public)\u001b[0m",
                                "",
                                "TASK [service-ks-register : placement | Creating projects] *********************",
                                "Tuesday 24 February 2026  01:04:34 +0000 (0:00:07.633)       0:00:15.340 ****** ",
                                "\u001b[0;32mok: [testbed-node-0] => (item=service)\u001b[0m",
                                "",
                                "TASK [service-ks-register : placement | Creating users] ************************",
                                "Tuesday 24 February 2026  01:04:38 +0000 (0:00:03.948)       0:00:19.288 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=placement -> service)\u001b[0m",
                                "\u001b[1;35m[WARNING]: Module did not set no_log for update_password\u001b[0m",
                                "",
                                "TASK [service-ks-register : placement | Creating roles] ************************",
                                "Tuesday 24 February 2026  01:04:42 +0000 (0:00:04.837)       0:00:24.126 ****** ",
                                "\u001b[0;32mok: [testbed-node-0] => (item=admin)\u001b[0m",
                                "",
                                "TASK [service-ks-register : placement | Granting user roles] *******************",
                                "Tuesday 24 February 2026  01:04:47 +0000 (0:00:04.433)       0:00:28.559 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=placement -> service -> admin)\u001b[0m",
                                "",
                                "TASK [placement : include_tasks] ***********************************************",
                                "Tuesday 24 February 2026  01:04:52 +0000 (0:00:04.969)       0:00:33.529 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [placement : Ensuring config directories exist] ***************************",
                                "Tuesday 24 February 2026  01:04:52 +0000 (0:00:00.429)       0:00:33.958 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/placement-api:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8780'], 'timeout': '30'}, 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/placement-api:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8780'], 'timeout': '30'}, 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/placement-api:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8780'], 'timeout': '30'}, 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}}})\u001b[0m",
                                "",
                                "TASK [placement : Check if policies shall be overwritten] **********************",
                                "Tuesday 24 February 2026  01:04:54 +0000 (0:00:01.711)       0:00:35.670 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [placement : Set placement policy file] ***********************************",
                                "Tuesday 24 February 2026  01:04:54 +0000 (0:00:00.289)       0:00:35.960 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [placement : include_tasks] ***********************************************",
                                "Tuesday 24 February 2026  01:04:55 +0000 (0:00:00.872)       0:00:36.832 ****** ",
                                "\u001b[0;36mincluded: /ansible/roles/placement/tasks/copy-certs.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [service-cert-copy : placement | Copying over extra CA certificates] ******",
                                "Tuesday 24 February 2026  01:04:56 +0000 (0:00:00.627)       0:00:37.459 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/placement-api:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8780'], 'timeout': '30'}, 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/placement-api:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8780'], 'timeout': '30'}, 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/placement-api:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8780'], 'timeout': '30'}, 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}}})\u001b[0m",
                                "",
                                "TASK [service-cert-copy : placement | Copying over backend internal TLS certificate] ***",
                                "Tuesday 24 February 2026  01:04:58 +0000 (0:00:02.017)       0:00:39.477 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/placement-api:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8780'], 'timeout': '30'}, 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/placement-api:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8780'], 'timeout': '30'}, 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/placement-api:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8780'], 'timeout': '30'}, 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [service-cert-copy : placement | Copying over backend internal TLS key] ***",
                                "Tuesday 24 February 2026  01:04:59 +0000 (0:00:00.918)       0:00:40.395 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/placement-api:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8780'], 'timeout': '30'}, 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/placement-api:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8780'], 'timeout': '30'}, 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/placement-api:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8780'], 'timeout': '30'}, 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [placement : Copying over config.json files for services] *****************",
                                "Tuesday 24 February 2026  01:05:00 +0000 (0:00:01.788)       0:00:42.184 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/placement-api:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8780'], 'timeout': '30'}, 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/placement-api:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8780'], 'timeout': '30'}, 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/placement-api:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8780'], 'timeout': '30'}, 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}}})\u001b[0m",
                                "",
                                "TASK [placement : Copying over placement.conf] *********************************",
                                "Tuesday 24 February 2026  01:05:03 +0000 (0:00:02.110)       0:00:44.294 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/placement-api:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8780'], 'timeout': '30'}, 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/placement-api:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8780'], 'timeout': '30'}, 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/placement-api:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8780'], 'timeout': '30'}, 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}}})\u001b[0m",
                                "",
                                "TASK [placement : Copying over placement-api wsgi configuration] ***************",
                                "Tuesday 24 February 2026  01:05:09 +0000 (0:00:06.241)       0:00:50.536 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=/ansible/roles/placement/templates/placement-api-wsgi.conf.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=/ansible/roles/placement/templates/placement-api-wsgi.conf.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=/ansible/roles/placement/templates/placement-api-wsgi.conf.j2)\u001b[0m",
                                "",
                                "TASK [placement : Copying over migrate-db.rc.j2 configuration] *****************",
                                "Tuesday 24 February 2026  01:05:11 +0000 (0:00:02.441)       0:00:52.978 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [placement : Copying over existing policy file] ***************************",
                                "Tuesday 24 February 2026  01:05:13 +0000 (0:00:02.040)       0:00:55.018 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/placement-api:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8780'], 'timeout': '30'}, 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/placement-api:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8780'], 'timeout': '30'}, 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/placement-api:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8780'], 'timeout': '30'}, 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [placement : Check placement containers] **********************************",
                                "Tuesday 24 February 2026  01:05:14 +0000 (0:00:00.676)       0:00:55.695 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/placement-api:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8780'], 'timeout': '30'}, 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/placement-api:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8780'], 'timeout': '30'}, 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'placement-api', 'value': {'container_name': 'placement_api', 'group': 'placement-api', 'image': 'registry.osism.tech/kolla/placement-api:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/placement-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8780'], 'timeout': '30'}, 'haproxy': {'placement_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}, 'placement_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8780', 'listen_port': '8780', 'tls_backend': 'no'}}}})\u001b[0m",
                                "",
                                "TASK [placement : Creating placement databases] ********************************",
                                "Tuesday 24 February 2026  01:05:16 +0000 (0:00:01.584)       0:00:57.280 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [placement : Creating placement databases user and setting permissions] ***",
                                "Tuesday 24 February 2026  01:05:21 +0000 (0:00:05.082)       0:01:02.363 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [placement : Running placement bootstrap container] ***********************",
                                "Tuesday 24 February 2026  01:05:24 +0000 (0:00:03.089)       0:01:05.452 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [placement : Flush handlers] **********************************************",
                                "Tuesday 24 February 2026  01:05:40 +0000 (0:00:16.111)       0:01:21.563 ****** ",
                                "",
                                "TASK [placement : Flush handlers] **********************************************",
                                "Tuesday 24 February 2026  01:05:40 +0000 (0:00:00.073)       0:01:21.637 ****** ",
                                "",
                                "TASK [placement : Flush handlers] **********************************************",
                                "Tuesday 24 February 2026  01:05:40 +0000 (0:00:00.118)       0:01:21.756 ****** ",
                                "",
                                "RUNNING HANDLER [placement : Restart placement-api container] ******************",
                                "Tuesday 24 February 2026  01:05:40 +0000 (0:00:00.079)       0:01:21.835 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=21  \u001b[0m \u001b[0;33mchanged=15  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=6   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=12  \u001b[0m \u001b[0;33mchanged=8   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=5   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=12  \u001b[0m \u001b[0;33mchanged=8   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=5   \u001b[0m rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Tuesday 24 February 2026  01:05:52 +0000 (0:00:12.133)       0:01:33.968 ****** ",
                                "=============================================================================== ",
                                "placement : Running placement bootstrap container ---------------------- 16.11s",
                                "placement : Restart placement-api container ---------------------------- 12.13s",
                                "service-ks-register : placement | Creating endpoints -------------------- 7.63s",
                                "placement : Copying over placement.conf --------------------------------- 6.24s",
                                "placement : Creating placement databases -------------------------------- 5.08s",
                                "service-ks-register : placement | Granting user roles ------------------- 4.97s",
                                "service-ks-register : placement | Creating users ------------------------ 4.84s",
                                "service-ks-register : placement | Creating roles ------------------------ 4.43s",
                                "service-ks-register : placement | Creating services --------------------- 4.20s",
                                "service-ks-register : placement | Creating projects --------------------- 3.95s",
                                "placement : Creating placement databases user and setting permissions --- 3.09s",
                                "placement : Copying over placement-api wsgi configuration --------------- 2.44s",
                                "placement : Copying over config.json files for services ----------------- 2.11s",
                                "placement : Copying over migrate-db.rc.j2 configuration ----------------- 2.04s",
                                "service-cert-copy : placement | Copying over extra CA certificates ------ 2.02s",
                                "service-cert-copy : placement | Copying over backend internal TLS key --- 1.79s",
                                "placement : Ensuring config directories exist --------------------------- 1.71s",
                                "Group hosts based on enabled services ----------------------------------- 1.65s",
                                "placement : Check placement containers ---------------------------------- 1.58s",
                                "placement : include_tasks ----------------------------------------------- 0.95s",
                                "\u001b[32m2026-02-24 01:05:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f714ea82-7678-4e20-85bd-ee1502bb5463 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:05:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a8ebcda5-cdc0-4611-9f70-ca910ead0481 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:05:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 47bc2a2f-3681-4d97-beda-759966d298ca is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:05:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44aa611b-f898-4352-b174-ada638d8f758 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:05:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 01:06:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f714ea82-7678-4e20-85bd-ee1502bb5463 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:06:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a8ebcda5-cdc0-4611-9f70-ca910ead0481 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:06:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a29489db-95ad-418c-bc67-f87a0737c252 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:06:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 47bc2a2f-3681-4d97-beda-759966d298ca is in state SUCCESS\u001b[0m",
                                "\u001b[32m2026-02-24 01:06:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44aa611b-f898-4352-b174-ada638d8f758 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:06:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 01:06:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f714ea82-7678-4e20-85bd-ee1502bb5463 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:06:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a8ebcda5-cdc0-4611-9f70-ca910ead0481 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:06:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a29489db-95ad-418c-bc67-f87a0737c252 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:06:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44aa611b-f898-4352-b174-ada638d8f758 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:06:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 01:06:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f714ea82-7678-4e20-85bd-ee1502bb5463 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:06:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a8ebcda5-cdc0-4611-9f70-ca910ead0481 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:06:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a29489db-95ad-418c-bc67-f87a0737c252 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:06:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44aa611b-f898-4352-b174-ada638d8f758 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:06:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 01:06:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f714ea82-7678-4e20-85bd-ee1502bb5463 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:06:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a8ebcda5-cdc0-4611-9f70-ca910ead0481 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:06:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a29489db-95ad-418c-bc67-f87a0737c252 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:06:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44aa611b-f898-4352-b174-ada638d8f758 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:06:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 01:06:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask f714ea82-7678-4e20-85bd-ee1502bb5463 is in state SUCCESS\u001b[0m",
                                "",
                                "",
                                "PLAY [Group hosts based on configuration] **************************************",
                                "",
                                "TASK [Group hosts based on Kolla action] ***************************************",
                                "Tuesday 24 February 2026  01:05:58 +0000 (0:00:00.208)       0:00:00.208 ****** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [Group hosts based on enabled services] ***********************************",
                                "Tuesday 24 February 2026  01:05:58 +0000 (0:00:00.311)       0:00:00.520 ****** ",
                                "\u001b[0;32mok: [testbed-node-0] => (item=enable_keystone_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item=enable_keystone_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item=enable_keystone_True)\u001b[0m",
                                "",
                                "PLAY [Wait for the Keystone service] *******************************************",
                                "",
                                "TASK [Waiting for Keystone public port to be UP] *******************************",
                                "Tuesday 24 February 2026  01:05:59 +0000 (0:00:00.748)       0:00:01.268 ****** ",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;32mtestbed-node-0\u001b[0m             : \u001b[0;32mok=3   \u001b[0m changed=0    unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "\u001b[0;32mtestbed-node-1\u001b[0m             : \u001b[0;32mok=3   \u001b[0m changed=0    unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "\u001b[0;32mtestbed-node-2\u001b[0m             : \u001b[0;32mok=3   \u001b[0m changed=0    unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Tuesday 24 February 2026  01:06:00 +0000 (0:00:00.662)       0:00:01.931 ****** ",
                                "=============================================================================== ",
                                "Group hosts based on enabled services ----------------------------------- 0.75s",
                                "Waiting for Keystone public port to be UP ------------------------------- 0.66s",
                                "Group hosts based on Kolla action --------------------------------------- 0.31s",
                                "",
                                "",
                                "PLAY [Group hosts based on configuration] **************************************",
                                "",
                                "TASK [Group hosts based on Kolla action] ***************************************",
                                "Tuesday 24 February 2026  01:02:32 +0000 (0:00:00.288)       0:00:00.288 ****** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [Group hosts based on enabled services] ***********************************",
                                "Tuesday 24 February 2026  01:02:32 +0000 (0:00:00.301)       0:00:00.590 ****** ",
                                "\u001b[0;32mok: [testbed-node-0] => (item=enable_designate_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item=enable_designate_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item=enable_designate_True)\u001b[0m",
                                "",
                                "PLAY [Apply role designate] ****************************************************",
                                "",
                                "TASK [designate : include_tasks] ***********************************************",
                                "Tuesday 24 February 2026  01:02:33 +0000 (0:00:00.448)       0:00:01.038 ****** ",
                                "\u001b[0;36mincluded: /ansible/roles/designate/tasks/deploy.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [service-ks-register : designate | Creating services] *********************",
                                "Tuesday 24 February 2026  01:02:33 +0000 (0:00:00.510)       0:00:01.549 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=designate (dns))\u001b[0m",
                                "",
                                "TASK [service-ks-register : designate | Creating endpoints] ********************",
                                "Tuesday 24 February 2026  01:02:38 +0000 (0:00:04.482)       0:00:06.032 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=designate -> https://api-int.testbed.osism.xyz:9001 -> internal)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=designate -> https://api.testbed.osism.xyz:9001 -> public)\u001b[0m",
                                "",
                                "TASK [service-ks-register : designate | Creating projects] *********************",
                                "Tuesday 24 February 2026  01:02:46 +0000 (0:00:08.061)       0:00:14.093 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=service)\u001b[0m",
                                "",
                                "TASK [service-ks-register : designate | Creating users] ************************",
                                "Tuesday 24 February 2026  01:02:50 +0000 (0:00:04.045)       0:00:18.139 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=designate -> service)\u001b[0m",
                                "\u001b[1;35m[WARNING]: Module did not set no_log for update_password\u001b[0m",
                                "",
                                "TASK [service-ks-register : designate | Creating roles] ************************",
                                "Tuesday 24 February 2026  01:02:54 +0000 (0:00:04.576)       0:00:22.715 ****** ",
                                "\u001b[0;32mok: [testbed-node-0] => (item=admin)\u001b[0m",
                                "",
                                "TASK [service-ks-register : designate | Granting user roles] *******************",
                                "Tuesday 24 February 2026  01:02:59 +0000 (0:00:04.160)       0:00:26.876 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=designate -> service -> admin)\u001b[0m",
                                "",
                                "TASK [designate : Ensuring config directories exist] ***************************",
                                "Tuesday 24 February 2026  01:03:03 +0000 (0:00:04.804)       0:00:31.680 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-api:2024.2', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-api:2024.2', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-api:2024.2', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-backend-bind9:2024.2', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-backend-bind9:2024.2', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-backend-bind9:2024.2', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-central:2024.2', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-central:2024.2', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-central:2024.2', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-mdns:2024.2', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-mdns:2024.2', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-mdns:2024.2', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-producer:2024.2', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-producer:2024.2', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-producer:2024.2', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-worker:2024.2', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-worker:2024.2', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-worker:2024.2', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}})\u001b[0m",
                                "",
                                "TASK [designate : Check if policies shall be overwritten] **********************",
                                "Tuesday 24 February 2026  01:03:07 +0000 (0:00:03.279)       0:00:34.959 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [designate : Set designate policy file] ***********************************",
                                "Tuesday 24 February 2026  01:03:07 +0000 (0:00:00.151)       0:00:35.111 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [designate : include_tasks] ***********************************************",
                                "Tuesday 24 February 2026  01:03:07 +0000 (0:00:00.287)       0:00:35.398 ****** ",
                                "\u001b[0;36mincluded: /ansible/roles/designate/tasks/copy-certs.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [service-cert-copy : designate | Copying over extra CA certificates] ******",
                                "Tuesday 24 February 2026  01:03:08 +0000 (0:00:00.860)       0:00:36.258 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-api:2024.2', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-api:2024.2', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-api:2024.2', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-backend-bind9:2024.2', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-backend-bind9:2024.2', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-backend-bind9:2024.2', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-central:2024.2', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-central:2024.2', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-central:2024.2', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-mdns:2024.2', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-mdns:2024.2', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-mdns:2024.2', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-producer:2024.2', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-producer:2024.2', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-producer:2024.2', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-worker:2024.2', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-worker:2024.2', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-worker:2024.2', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}})\u001b[0m",
                                "",
                                "TASK [service-cert-copy : designate | Copying over backend internal TLS certificate] ***",
                                "Tuesday 24 February 2026  01:03:14 +0000 (0:00:06.356)       0:00:42.615 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-api:2024.2', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-backend-bind9:2024.2', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-central:2024.2', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-mdns:2024.2', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-producer:2024.2', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-worker:2024.2', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-api:2024.2', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-backend-bind9:2024.2', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-central:2024.2', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-mdns:2024.2', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-producer:2024.2', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-worker:2024.2', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-api:2024.2', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-backend-bind9:2024.2', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-central:2024.2', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-mdns:2024.2', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-producer:2024.2', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-worker:2024.2', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [service-cert-copy : designate | Copying over backend internal TLS key] ***",
                                "Tuesday 24 February 2026  01:03:16 +0000 (0:00:01.398)       0:00:44.014 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-api:2024.2', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-backend-bind9:2024.2', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-central:2024.2', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-mdns:2024.2', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-producer:2024.2', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-api:2024.2', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-worker:2024.2', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-backend-bind9:2024.2', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-central:2024.2', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-mdns:2024.2', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-producer:2024.2', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-worker:2024.2', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-api:2024.2', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-backend-bind9:2024.2', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-central:2024.2', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-mdns:2024.2', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-producer:2024.2', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-worker:2024.2', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [designate : Copying over config.json files for services] *****************",
                                "Tuesday 24 February 2026  01:03:18 +0000 (0:00:02.139)       0:00:46.154 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-api:2024.2', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-api:2024.2', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-api:2024.2', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-backend-bind9:2024.2', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-backend-bind9:2024.2', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-backend-bind9:2024.2', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-central:2024.2', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-central:2024.2', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-central:2024.2', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-mdns:2024.2', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-mdns:2024.2', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-mdns:2024.2', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-producer:2024.2', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-producer:2024.2', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-producer:2024.2', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-worker:2024.2', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-worker:2024.2', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-worker:2024.2', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}})\u001b[0m",
                                "",
                                "TASK [designate : Copying over designate.conf] *********************************",
                                "Tuesday 24 February 2026  01:03:25 +0000 (0:00:07.464)       0:00:53.619 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-api:2024.2', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-api:2024.2', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-api:2024.2', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-backend-bind9:2024.2', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-backend-bind9:2024.2', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-backend-bind9:2024.2', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-central:2024.2', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-central:2024.2', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-central:2024.2', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-mdns:2024.2', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-mdns:2024.2', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-mdns:2024.2', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-producer:2024.2', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-producer:2024.2', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-producer:2024.2', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-worker:2024.2', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-worker:2024.2', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-worker:2024.2', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}})\u001b[0m",
                                "",
                                "TASK [designate : Copying over pools.yaml] *************************************",
                                "Tuesday 24 February 2026  01:03:50 +0000 (0:00:24.326)       0:01:17.945 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=/ansible/roles/designate/templates/pools.yaml.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=/ansible/roles/designate/templates/pools.yaml.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=/ansible/roles/designate/templates/pools.yaml.j2)\u001b[0m",
                                "",
                                "TASK [designate : Copying over named.conf] *************************************",
                                "Tuesday 24 February 2026  01:03:58 +0000 (0:00:08.679)       0:01:26.625 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=/ansible/roles/designate/templates/named.conf.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=/ansible/roles/designate/templates/named.conf.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=/ansible/roles/designate/templates/named.conf.j2)\u001b[0m",
                                "",
                                "TASK [designate : Copying over rndc.conf] **************************************",
                                "Tuesday 24 February 2026  01:04:03 +0000 (0:00:04.651)       0:01:31.276 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-api:2024.2', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-api:2024.2', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-api:2024.2', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}}) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-backend-bind9:2024.2', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-central:2024.2', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-mdns:2024.2', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-producer:2024.2', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-backend-bind9:2024.2', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-central:2024.2', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-mdns:2024.2', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-producer:2024.2', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-backend-bind9:2024.2', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-central:2024.2', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-mdns:2024.2', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-producer:2024.2', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-worker:2024.2', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-worker:2024.2', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-worker:2024.2', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}})\u001b[0m",
                                "",
                                "TASK [designate : Copying over rndc.key] ***************************************",
                                "Tuesday 24 February 2026  01:04:07 +0000 (0:00:03.940)       0:01:35.216 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-api:2024.2', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-api:2024.2', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-api:2024.2', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}}) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-backend-bind9:2024.2', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-central:2024.2', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-backend-bind9:2024.2', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-mdns:2024.2', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-central:2024.2', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-mdns:2024.2', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-producer:2024.2', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-producer:2024.2', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-backend-bind9:2024.2', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-central:2024.2', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-mdns:2024.2', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-producer:2024.2', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-worker:2024.2', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-worker:2024.2', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-worker:2024.2', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}})\u001b[0m",
                                "",
                                "TASK [designate : include_tasks] ***********************************************",
                                "Tuesday 24 February 2026  01:04:12 +0000 (0:00:05.065)       0:01:40.281 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [designate : Copying over existing policy file] ***************************",
                                "Tuesday 24 February 2026  01:04:13 +0000 (0:00:01.104)       0:01:41.385 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-api:2024.2', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-backend-bind9:2024.2', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-central:2024.2', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-mdns:2024.2', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-producer:2024.2', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-worker:2024.2', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-api:2024.2', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-backend-bind9:2024.2', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-central:2024.2', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-mdns:2024.2', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-producer:2024.2', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-worker:2024.2', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-api:2024.2', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-backend-bind9:2024.2', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-central:2024.2', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-mdns:2024.2', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-producer:2024.2', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-worker:2024.2', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [designate : Check designate containers] **********************************",
                                "Tuesday 24 February 2026  01:04:14 +0000 (0:00:01.014)       0:01:42.400 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-api:2024.2', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-backend-bind9:2024.2', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-api:2024.2', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-api', 'value': {'container_name': 'designate_api', 'group': 'designate-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-api:2024.2', 'volumes': ['/etc/kolla/designate-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9001'], 'timeout': '30'}, 'haproxy': {'designate_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9001', 'listen_port': '9001'}, 'designate_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9001', 'listen_port': '9001'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-central:2024.2', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-backend-bind9:2024.2', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-backend-bind9', 'value': {'container_name': 'designate_backend_bind9', 'group': 'designate-backend-bind9', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-backend-bind9:2024.2', 'volumes': ['/etc/kolla/designate-backend-bind9/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', 'designate_backend_bind9:/var/lib/named/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen named 53'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-mdns:2024.2', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-central:2024.2', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-central', 'value': {'container_name': 'designate_central', 'group': 'designate-central', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-central:2024.2', 'volumes': ['/etc/kolla/designate-central/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-central 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-producer:2024.2', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-mdns:2024.2', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-mdns', 'value': {'container_name': 'designate_mdns', 'group': 'designate-mdns', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-mdns:2024.2', 'volumes': ['/etc/kolla/designate-mdns/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-mdns 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-producer:2024.2', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-worker:2024.2', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-producer', 'value': {'container_name': 'designate_producer', 'group': 'designate-producer', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-producer:2024.2', 'volumes': ['/etc/kolla/designate-producer/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-producer 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-worker:2024.2', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'designate-worker', 'value': {'container_name': 'designate_worker', 'group': 'designate-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/designate-worker:2024.2', 'volumes': ['/etc/kolla/designate-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port designate-worker 5672'], 'timeout': '30'}}})\u001b[0m",
                                "",
                                "TASK [designate : include_tasks] ***********************************************",
                                "Tuesday 24 February 2026  01:04:21 +0000 (0:00:07.070)       0:01:49.470 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [designate : Creating Designate databases] ********************************",
                                "Tuesday 24 February 2026  01:04:22 +0000 (0:00:00.488)       0:01:49.958 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=designate)\u001b[0m",
                                "",
                                "TASK [designate : Creating Designate databases user and setting permissions] ***",
                                "Tuesday 24 February 2026  01:04:24 +0000 (0:00:02.750)       0:01:52.709 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=None)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0 -> {{ groups['designate-central'][0] }}]\u001b[0m",
                                "",
                                "TASK [designate : Running Designate bootstrap container] ***********************",
                                "Tuesday 24 February 2026  01:04:27 +0000 (0:00:02.892)       0:01:55.602 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [designate : Flush handlers] **********************************************",
                                "Tuesday 24 February 2026  01:04:47 +0000 (0:00:19.447)       0:02:15.050 ****** ",
                                "",
                                "TASK [designate : Flush handlers] **********************************************",
                                "Tuesday 24 February 2026  01:04:47 +0000 (0:00:00.070)       0:02:15.120 ****** ",
                                "",
                                "TASK [designate : Flush handlers] **********************************************",
                                "Tuesday 24 February 2026  01:04:47 +0000 (0:00:00.067)       0:02:15.187 ****** ",
                                "",
                                "RUNNING HANDLER [designate : Restart designate-backend-bind9 container] ********",
                                "Tuesday 24 February 2026  01:04:47 +0000 (0:00:00.065)       0:02:15.253 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [designate : Restart designate-api container] ******************",
                                "Tuesday 24 February 2026  01:05:01 +0000 (0:00:13.945)       0:02:29.199 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [designate : Restart designate-central container] **************",
                                "Tuesday 24 February 2026  01:05:15 +0000 (0:00:14.513)       0:02:43.712 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "",
                                "RUNNING HANDLER [designate : Restart designate-producer container] *************",
                                "Tuesday 24 February 2026  01:05:31 +0000 (0:00:15.292)       0:02:59.004 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "RUNNING HANDLER [designate : Restart designate-mdns container] *****************",
                                "Tuesday 24 February 2026  01:05:40 +0000 (0:00:09.370)       0:03:08.375 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [designate : Restart designate-worker container] ***************",
                                "Tuesday 24 February 2026  01:05:52 +0000 (0:00:11.935)       0:03:20.310 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [designate : Non-destructive DNS pools update] ****************************",
                                "Tuesday 24 February 2026  01:06:04 +0000 (0:00:11.520)       0:03:31.831 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=29  \u001b[0m \u001b[0;33mchanged=24  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=7   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=19  \u001b[0m \u001b[0;33mchanged=15  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=6   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=19  \u001b[0m \u001b[0;33mchanged=15  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=6   \u001b[0m rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Tuesday 24 February 2026  01:06:12 +0000 (0:00:08.881)       0:03:40.713 ****** ",
                                "=============================================================================== ",
                                "designate : Copying over designate.conf -------------------------------- 24.33s",
                                "designate : Running Designate bootstrap container ---------------------- 19.45s",
                                "designate : Restart designate-central container ------------------------ 15.29s",
                                "designate : Restart designate-api container ---------------------------- 14.51s",
                                "designate : Restart designate-backend-bind9 container ------------------ 13.95s",
                                "designate : Restart designate-mdns container --------------------------- 11.94s",
                                "designate : Restart designate-worker container ------------------------- 11.52s",
                                "designate : Restart designate-producer container ------------------------ 9.37s",
                                "designate : Non-destructive DNS pools update ---------------------------- 8.88s",
                                "designate : Copying over pools.yaml ------------------------------------- 8.68s",
                                "service-ks-register : designate | Creating endpoints -------------------- 8.06s",
                                "designate : Copying over config.json files for services ----------------- 7.46s",
                                "designate : Check designate containers ---------------------------------- 7.07s",
                                "service-cert-copy : designate | Copying over extra CA certificates ------ 6.36s",
                                "designate : Copying over rndc.key --------------------------------------- 5.07s",
                                "service-ks-register : designate | Granting user roles ------------------- 4.80s",
                                "designate : Copying over named.conf ------------------------------------- 4.65s",
                                "service-ks-register : designate | Creating users ------------------------ 4.58s",
                                "service-ks-register : designate | Creating services --------------------- 4.48s",
                                "service-ks-register : designate | Creating roles ------------------------ 4.16s",
                                "\u001b[32m2026-02-24 01:06:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b44e23a9-4820-4202-81f0-47da60f63246 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:06:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a8ebcda5-cdc0-4611-9f70-ca910ead0481 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:06:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a29489db-95ad-418c-bc67-f87a0737c252 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:06:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44aa611b-f898-4352-b174-ada638d8f758 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:06:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 01:06:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b44e23a9-4820-4202-81f0-47da60f63246 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:06:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a8ebcda5-cdc0-4611-9f70-ca910ead0481 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:06:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a29489db-95ad-418c-bc67-f87a0737c252 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:06:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44aa611b-f898-4352-b174-ada638d8f758 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:06:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 01:06:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b44e23a9-4820-4202-81f0-47da60f63246 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:06:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a8ebcda5-cdc0-4611-9f70-ca910ead0481 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:06:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a29489db-95ad-418c-bc67-f87a0737c252 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:06:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44aa611b-f898-4352-b174-ada638d8f758 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:06:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 01:06:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b44e23a9-4820-4202-81f0-47da60f63246 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:06:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a8ebcda5-cdc0-4611-9f70-ca910ead0481 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:06:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a29489db-95ad-418c-bc67-f87a0737c252 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:06:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44aa611b-f898-4352-b174-ada638d8f758 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:06:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 01:06:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b44e23a9-4820-4202-81f0-47da60f63246 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:06:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a8ebcda5-cdc0-4611-9f70-ca910ead0481 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:06:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a29489db-95ad-418c-bc67-f87a0737c252 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:06:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44aa611b-f898-4352-b174-ada638d8f758 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:06:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 01:06:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b44e23a9-4820-4202-81f0-47da60f63246 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:06:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a8ebcda5-cdc0-4611-9f70-ca910ead0481 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:06:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a29489db-95ad-418c-bc67-f87a0737c252 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:06:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44aa611b-f898-4352-b174-ada638d8f758 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:06:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 01:06:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b44e23a9-4820-4202-81f0-47da60f63246 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:06:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a8ebcda5-cdc0-4611-9f70-ca910ead0481 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:06:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a29489db-95ad-418c-bc67-f87a0737c252 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:06:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44aa611b-f898-4352-b174-ada638d8f758 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:06:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 01:06:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b44e23a9-4820-4202-81f0-47da60f63246 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:06:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a8ebcda5-cdc0-4611-9f70-ca910ead0481 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:06:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a29489db-95ad-418c-bc67-f87a0737c252 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:06:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44aa611b-f898-4352-b174-ada638d8f758 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:06:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 01:06:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b44e23a9-4820-4202-81f0-47da60f63246 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:06:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a8ebcda5-cdc0-4611-9f70-ca910ead0481 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:06:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a29489db-95ad-418c-bc67-f87a0737c252 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:06:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44aa611b-f898-4352-b174-ada638d8f758 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:06:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 01:06:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b44e23a9-4820-4202-81f0-47da60f63246 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:06:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a8ebcda5-cdc0-4611-9f70-ca910ead0481 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:06:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a29489db-95ad-418c-bc67-f87a0737c252 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:06:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44aa611b-f898-4352-b174-ada638d8f758 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:06:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 01:06:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b44e23a9-4820-4202-81f0-47da60f63246 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:06:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a8ebcda5-cdc0-4611-9f70-ca910ead0481 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:06:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a29489db-95ad-418c-bc67-f87a0737c252 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:06:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44aa611b-f898-4352-b174-ada638d8f758 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:06:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 01:06:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b44e23a9-4820-4202-81f0-47da60f63246 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:06:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a8ebcda5-cdc0-4611-9f70-ca910ead0481 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:06:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a29489db-95ad-418c-bc67-f87a0737c252 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:06:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44aa611b-f898-4352-b174-ada638d8f758 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:06:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 01:06:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b44e23a9-4820-4202-81f0-47da60f63246 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:06:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a8ebcda5-cdc0-4611-9f70-ca910ead0481 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:06:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a29489db-95ad-418c-bc67-f87a0737c252 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:06:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44aa611b-f898-4352-b174-ada638d8f758 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:06:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 01:06:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b44e23a9-4820-4202-81f0-47da60f63246 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:06:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a8ebcda5-cdc0-4611-9f70-ca910ead0481 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:06:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a29489db-95ad-418c-bc67-f87a0737c252 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:06:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44aa611b-f898-4352-b174-ada638d8f758 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:06:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 01:06:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b44e23a9-4820-4202-81f0-47da60f63246 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:06:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a8ebcda5-cdc0-4611-9f70-ca910ead0481 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:06:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a29489db-95ad-418c-bc67-f87a0737c252 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:06:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44aa611b-f898-4352-b174-ada638d8f758 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:06:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 01:07:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask b44e23a9-4820-4202-81f0-47da60f63246 is in state SUCCESS\u001b[0m",
                                "\u001b[32m2026-02-24 01:07:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a8ebcda5-cdc0-4611-9f70-ca910ead0481 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:07:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a29489db-95ad-418c-bc67-f87a0737c252 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:07:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44aa611b-f898-4352-b174-ada638d8f758 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:07:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 01:07:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fba6bf07-65a1-432e-9a3d-72209557a21d is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:07:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a8ebcda5-cdc0-4611-9f70-ca910ead0481 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:07:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a29489db-95ad-418c-bc67-f87a0737c252 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:07:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44aa611b-f898-4352-b174-ada638d8f758 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:07:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 01:07:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fba6bf07-65a1-432e-9a3d-72209557a21d is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:07:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a8ebcda5-cdc0-4611-9f70-ca910ead0481 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:07:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a29489db-95ad-418c-bc67-f87a0737c252 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:07:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44aa611b-f898-4352-b174-ada638d8f758 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:07:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 01:07:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fba6bf07-65a1-432e-9a3d-72209557a21d is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:07:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a8ebcda5-cdc0-4611-9f70-ca910ead0481 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:07:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a29489db-95ad-418c-bc67-f87a0737c252 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:07:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44aa611b-f898-4352-b174-ada638d8f758 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:07:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 01:07:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fba6bf07-65a1-432e-9a3d-72209557a21d is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:07:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a8ebcda5-cdc0-4611-9f70-ca910ead0481 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:07:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a29489db-95ad-418c-bc67-f87a0737c252 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:07:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44aa611b-f898-4352-b174-ada638d8f758 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:07:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 01:07:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fba6bf07-65a1-432e-9a3d-72209557a21d is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:07:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a8ebcda5-cdc0-4611-9f70-ca910ead0481 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:07:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a29489db-95ad-418c-bc67-f87a0737c252 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:07:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44aa611b-f898-4352-b174-ada638d8f758 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:07:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 01:07:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fba6bf07-65a1-432e-9a3d-72209557a21d is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:07:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a8ebcda5-cdc0-4611-9f70-ca910ead0481 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:07:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a29489db-95ad-418c-bc67-f87a0737c252 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:07:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44aa611b-f898-4352-b174-ada638d8f758 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:07:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 01:07:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fba6bf07-65a1-432e-9a3d-72209557a21d is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:07:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a8ebcda5-cdc0-4611-9f70-ca910ead0481 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:07:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a29489db-95ad-418c-bc67-f87a0737c252 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:07:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44aa611b-f898-4352-b174-ada638d8f758 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:07:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 01:07:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fba6bf07-65a1-432e-9a3d-72209557a21d is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:07:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a8ebcda5-cdc0-4611-9f70-ca910ead0481 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:07:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a29489db-95ad-418c-bc67-f87a0737c252 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:07:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44aa611b-f898-4352-b174-ada638d8f758 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:07:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 01:07:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fba6bf07-65a1-432e-9a3d-72209557a21d is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:07:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a8ebcda5-cdc0-4611-9f70-ca910ead0481 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:07:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a29489db-95ad-418c-bc67-f87a0737c252 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:07:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44aa611b-f898-4352-b174-ada638d8f758 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:07:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 01:07:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fba6bf07-65a1-432e-9a3d-72209557a21d is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:07:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a8ebcda5-cdc0-4611-9f70-ca910ead0481 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:07:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a29489db-95ad-418c-bc67-f87a0737c252 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:07:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44aa611b-f898-4352-b174-ada638d8f758 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:07:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 01:07:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fba6bf07-65a1-432e-9a3d-72209557a21d is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:07:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a8ebcda5-cdc0-4611-9f70-ca910ead0481 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:07:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a29489db-95ad-418c-bc67-f87a0737c252 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:07:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44aa611b-f898-4352-b174-ada638d8f758 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:07:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 01:07:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fba6bf07-65a1-432e-9a3d-72209557a21d is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:07:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask eddb1a5f-448f-4517-b064-c30da006d691 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:07:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a8ebcda5-cdc0-4611-9f70-ca910ead0481 is in state SUCCESS\u001b[0m",
                                "",
                                "",
                                "PLAY [Group hosts based on configuration] **************************************",
                                "",
                                "TASK [Group hosts based on Kolla action] ***************************************",
                                "Tuesday 24 February 2026  01:06:19 +0000 (0:00:00.431)       0:00:00.431 ****** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Group hosts based on enabled services] ***********************************",
                                "Tuesday 24 February 2026  01:06:20 +0000 (0:00:01.058)       0:00:01.489 ****** ",
                                "\u001b[0;32mok: [testbed-manager] => (item=enable_ceph_rgw_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0] => (item=enable_ceph_rgw_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item=enable_ceph_rgw_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item=enable_ceph_rgw_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item=enable_ceph_rgw_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item=enable_ceph_rgw_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item=enable_ceph_rgw_True)\u001b[0m",
                                "",
                                "PLAY [Apply role ceph-rgw] *****************************************************",
                                "",
                                "TASK [ceph-rgw : include_tasks] ************************************************",
                                "Tuesday 24 February 2026  01:06:22 +0000 (0:00:01.650)       0:00:03.139 ****** ",
                                "\u001b[0;36mincluded: /ansible/roles/ceph-rgw/tasks/deploy.yml for testbed-manager, testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m",
                                "",
                                "TASK [service-ks-register : ceph-rgw | Creating services] **********************",
                                "Tuesday 24 February 2026  01:06:26 +0000 (0:00:03.865)       0:00:07.005 ****** ",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=swift (object-store))\u001b[0m",
                                "",
                                "TASK [service-ks-register : ceph-rgw | Creating endpoints] *********************",
                                "Tuesday 24 February 2026  01:06:30 +0000 (0:00:03.828)       0:00:10.833 ****** ",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=swift -> https://api-int.testbed.osism.xyz:6780/swift/v1/AUTH_%(project_id)s -> internal)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=swift -> https://api.testbed.osism.xyz:6780/swift/v1/AUTH_%(project_id)s -> public)\u001b[0m",
                                "",
                                "TASK [service-ks-register : ceph-rgw | Creating projects] **********************",
                                "Tuesday 24 February 2026  01:06:37 +0000 (0:00:07.352)       0:00:18.185 ****** ",
                                "\u001b[0;32mok: [testbed-manager] => (item=service)\u001b[0m",
                                "",
                                "TASK [service-ks-register : ceph-rgw | Creating users] *************************",
                                "Tuesday 24 February 2026  01:06:40 +0000 (0:00:03.458)       0:00:21.644 ****** ",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=ceph_rgw -> service)\u001b[0m",
                                "\u001b[1;35m[WARNING]: Module did not set no_log for update_password\u001b[0m",
                                "",
                                "TASK [service-ks-register : ceph-rgw | Creating roles] *************************",
                                "Tuesday 24 February 2026  01:06:46 +0000 (0:00:05.445)       0:00:27.090 ****** ",
                                "\u001b[0;32mok: [testbed-manager] => (item=admin)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=ResellerAdmin)\u001b[0m",
                                "",
                                "TASK [service-ks-register : ceph-rgw | Granting user roles] ********************",
                                "Tuesday 24 February 2026  01:06:53 +0000 (0:00:07.412)       0:00:34.502 ****** ",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=ceph_rgw -> service -> admin)\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;33mtestbed-manager\u001b[0m            : \u001b[0;32mok=9   \u001b[0m \u001b[0;33mchanged=5   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "\u001b[0;32mtestbed-node-0\u001b[0m             : \u001b[0;32mok=3   \u001b[0m changed=0    unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "\u001b[0;32mtestbed-node-1\u001b[0m             : \u001b[0;32mok=3   \u001b[0m changed=0    unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "\u001b[0;32mtestbed-node-2\u001b[0m             : \u001b[0;32mok=3   \u001b[0m changed=0    unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "\u001b[0;32mtestbed-node-3\u001b[0m             : \u001b[0;32mok=3   \u001b[0m changed=0    unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "\u001b[0;32mtestbed-node-4\u001b[0m             : \u001b[0;32mok=3   \u001b[0m changed=0    unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "\u001b[0;32mtestbed-node-5\u001b[0m             : \u001b[0;32mok=3   \u001b[0m changed=0    unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Tuesday 24 February 2026  01:06:58 +0000 (0:00:05.225)       0:00:39.727 ****** ",
                                "=============================================================================== ",
                                "service-ks-register : ceph-rgw | Creating roles ------------------------- 7.41s",
                                "service-ks-register : ceph-rgw | Creating endpoints --------------------- 7.35s",
                                "service-ks-register : ceph-rgw | Creating users ------------------------- 5.45s",
                                "service-ks-register : ceph-rgw | Granting user roles -------------------- 5.23s",
                                "ceph-rgw : include_tasks ------------------------------------------------ 3.86s",
                                "service-ks-register : ceph-rgw | Creating services ---------------------- 3.83s",
                                "service-ks-register : ceph-rgw | Creating projects ---------------------- 3.46s",
                                "Group hosts based on enabled services ----------------------------------- 1.65s",
                                "Group hosts based on Kolla action --------------------------------------- 1.06s",
                                "",
                                "",
                                "PLAY [Group hosts based on configuration] **************************************",
                                "",
                                "TASK [Group hosts based on Kolla action] ***************************************",
                                "Tuesday 24 February 2026  01:05:22 +0000 (0:00:01.108)       0:00:01.108 ****** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [Group hosts based on enabled services] ***********************************",
                                "Tuesday 24 February 2026  01:05:22 +0000 (0:00:00.446)       0:00:01.555 ****** ",
                                "\u001b[0;32mok: [testbed-node-0] => (item=enable_magnum_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item=enable_magnum_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item=enable_magnum_True)\u001b[0m",
                                "",
                                "PLAY [Apply role magnum] *******************************************************",
                                "",
                                "TASK [magnum : include_tasks] **************************************************",
                                "Tuesday 24 February 2026  01:05:23 +0000 (0:00:00.633)       0:00:02.189 ****** ",
                                "\u001b[0;36mincluded: /ansible/roles/magnum/tasks/deploy.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [service-ks-register : magnum | Creating services] ************************",
                                "Tuesday 24 February 2026  01:05:23 +0000 (0:00:00.641)       0:00:02.830 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=magnum (container-infra))\u001b[0m",
                                "",
                                "TASK [service-ks-register : magnum | Creating endpoints] ***********************",
                                "Tuesday 24 February 2026  01:05:28 +0000 (0:00:04.321)       0:00:07.151 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=magnum -> https://api-int.testbed.osism.xyz:9511/v1 -> internal)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=magnum -> https://api.testbed.osism.xyz:9511/v1 -> public)\u001b[0m",
                                "",
                                "TASK [service-ks-register : magnum | Creating projects] ************************",
                                "Tuesday 24 February 2026  01:05:36 +0000 (0:00:08.209)       0:00:15.361 ****** ",
                                "\u001b[0;32mok: [testbed-node-0] => (item=service)\u001b[0m",
                                "",
                                "TASK [service-ks-register : magnum | Creating users] ***************************",
                                "Tuesday 24 February 2026  01:05:40 +0000 (0:00:04.070)       0:00:19.432 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=magnum -> service)\u001b[0m",
                                "\u001b[1;35m[WARNING]: Module did not set no_log for update_password\u001b[0m",
                                "",
                                "TASK [service-ks-register : magnum | Creating roles] ***************************",
                                "Tuesday 24 February 2026  01:05:45 +0000 (0:00:04.896)       0:00:24.328 ****** ",
                                "\u001b[0;32mok: [testbed-node-0] => (item=admin)\u001b[0m",
                                "",
                                "TASK [service-ks-register : magnum | Granting user roles] **********************",
                                "Tuesday 24 February 2026  01:05:49 +0000 (0:00:04.366)       0:00:28.695 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=magnum -> service -> admin)\u001b[0m",
                                "",
                                "TASK [magnum : Creating Magnum trustee domain] *********************************",
                                "Tuesday 24 February 2026  01:05:54 +0000 (0:00:04.721)       0:00:33.416 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [magnum : Creating Magnum trustee user] ***********************************",
                                "Tuesday 24 February 2026  01:05:58 +0000 (0:00:03.987)       0:00:37.404 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [magnum : Creating Magnum trustee user role] ******************************",
                                "Tuesday 24 February 2026  01:06:02 +0000 (0:00:04.211)       0:00:41.615 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [magnum : Ensuring config directories exist] ******************************",
                                "Tuesday 24 February 2026  01:06:07 +0000 (0:00:04.403)       0:00:46.019 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-api:2024.2', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-api:2024.2', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-api:2024.2', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-conductor:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.11,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-conductor:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.10,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-conductor:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.12,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}})\u001b[0m",
                                "",
                                "TASK [magnum : Check if policies shall be overwritten] *************************",
                                "Tuesday 24 February 2026  01:06:08 +0000 (0:00:01.379)       0:00:47.398 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [magnum : Set magnum policy file] *****************************************",
                                "Tuesday 24 February 2026  01:06:08 +0000 (0:00:00.144)       0:00:47.543 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [magnum : Check if kubeconfig file is supplied] ***************************",
                                "Tuesday 24 February 2026  01:06:09 +0000 (0:00:00.744)       0:00:48.288 ****** ",
                                "\u001b[0;32mok: [testbed-node-0 -> localhost]\u001b[0m",
                                "",
                                "TASK [magnum : Copying over kubeconfig file] ***********************************",
                                "Tuesday 24 February 2026  01:06:10 +0000 (0:00:01.697)       0:00:49.985 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-api:2024.2', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-api:2024.2', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-api:2024.2', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-conductor:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.12,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-conductor:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.11,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-conductor:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.10,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}})\u001b[0m",
                                "",
                                "TASK [magnum : Set magnum kubeconfig file's path] ******************************",
                                "Tuesday 24 February 2026  01:06:13 +0000 (0:00:02.739)       0:00:52.724 ****** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [magnum : include_tasks] **************************************************",
                                "Tuesday 24 February 2026  01:06:14 +0000 (0:00:00.356)       0:00:53.081 ****** ",
                                "\u001b[0;36mincluded: /ansible/roles/magnum/tasks/copy-certs.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [service-cert-copy : magnum | Copying over extra CA certificates] *********",
                                "Tuesday 24 February 2026  01:06:15 +0000 (0:00:01.477)       0:00:54.559 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-api:2024.2', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-api:2024.2', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-api:2024.2', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-conductor:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.10,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-conductor:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.11,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-conductor:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.12,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}})\u001b[0m",
                                "",
                                "TASK [service-cert-copy : magnum | Copying over backend internal TLS certificate] ***",
                                "Tuesday 24 February 2026  01:06:18 +0000 (0:00:03.228)       0:00:57.787 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-api:2024.2', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-conductor:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.10,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-api:2024.2', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-conductor:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.11,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-api:2024.2', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-conductor:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.12,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [service-cert-copy : magnum | Copying over backend internal TLS key] ******",
                                "Tuesday 24 February 2026  01:06:19 +0000 (0:00:01.016)       0:00:58.804 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-api:2024.2', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-api:2024.2', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-conductor:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.10,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-conductor:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.11,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-api:2024.2', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-conductor:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.12,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [magnum : Copying over config.json files for services] ********************",
                                "Tuesday 24 February 2026  01:06:22 +0000 (0:00:02.593)       0:01:01.397 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-api:2024.2', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-api:2024.2', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-api:2024.2', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-conductor:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.10,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-conductor:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.12,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-conductor:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.11,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}})\u001b[0m",
                                "",
                                "TASK [magnum : Copying over magnum.conf] ***************************************",
                                "Tuesday 24 February 2026  01:06:26 +0000 (0:00:03.834)       0:01:05.232 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-api:2024.2', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-api:2024.2', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-api:2024.2', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-conductor:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.12,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-conductor:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.10,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-conductor:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.11,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}})\u001b[0m",
                                "",
                                "TASK [magnum : Copying over existing policy file] ******************************",
                                "Tuesday 24 February 2026  01:06:36 +0000 (0:00:10.617)       0:01:15.850 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-api:2024.2', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-conductor:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.10,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-api:2024.2', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-conductor:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.11,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-api:2024.2', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-conductor:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.12,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [magnum : Check magnum containers] ****************************************",
                                "Tuesday 24 February 2026  01:06:37 +0000 (0:00:00.662)       0:01:16.512 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-api:2024.2', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-api:2024.2', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'magnum-api', 'value': {'container_name': 'magnum_api', 'group': 'magnum-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-api:2024.2', 'environment': {'DUMMY_ENVIRONMENT': 'kolla_useless_env'}, 'volumes': ['/etc/kolla/magnum-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9511'], 'timeout': '30'}, 'haproxy': {'magnum_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9511', 'listen_port': '9511'}, 'magnum_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9511', 'listen_port': '9511'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-conductor:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.11,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-conductor:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.10,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'magnum-conductor', 'value': {'container_name': 'magnum_conductor', 'group': 'magnum-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/magnum-conductor:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.12,192.168.16.9'}, 'volumes': ['/etc/kolla/magnum-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'magnum:/var/lib/magnum/', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port magnum-conductor 5672'], 'timeout': '30'}}})\u001b[0m",
                                "",
                                "TASK [magnum : include_tasks] **************************************************",
                                "Tuesday 24 February 2026  01:06:39 +0000 (0:00:02.405)       0:01:18.917 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [magnum : Creating Magnum database] ***************************************",
                                "Tuesday 24 February 2026  01:06:40 +0000 (0:00:00.530)       0:01:19.447 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [magnum : Creating Magnum database user and setting permissions] **********",
                                "Tuesday 24 February 2026  01:06:43 +0000 (0:00:02.683)       0:01:22.130 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [magnum : Running Magnum bootstrap container] *****************************",
                                "Tuesday 24 February 2026  01:06:46 +0000 (0:00:03.531)       0:01:25.662 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [magnum : Flush handlers] *************************************************",
                                "Tuesday 24 February 2026  01:07:05 +0000 (0:00:18.495)       0:01:44.157 ****** ",
                                "",
                                "TASK [magnum : Flush handlers] *************************************************",
                                "Tuesday 24 February 2026  01:07:05 +0000 (0:00:00.140)       0:01:44.298 ****** ",
                                "",
                                "TASK [magnum : Flush handlers] *************************************************",
                                "Tuesday 24 February 2026  01:07:05 +0000 (0:00:00.068)       0:01:44.367 ****** ",
                                "",
                                "RUNNING HANDLER [magnum : Restart magnum-api container] ************************",
                                "Tuesday 24 February 2026  01:07:05 +0000 (0:00:00.106)       0:01:44.473 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [magnum : Restart magnum-conductor container] ******************",
                                "Tuesday 24 February 2026  01:07:21 +0000 (0:00:15.914)       0:02:00.388 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=26  \u001b[0m \u001b[0;33mchanged=18  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=6   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=13  \u001b[0m \u001b[0;33mchanged=8   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=5   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=13  \u001b[0m \u001b[0;33mchanged=8   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=5   \u001b[0m rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Tuesday 24 February 2026  01:07:35 +0000 (0:00:13.930)       0:02:14.319 ****** ",
                                "=============================================================================== ",
                                "magnum : Running Magnum bootstrap container ---------------------------- 18.50s",
                                "magnum : Restart magnum-api container ---------------------------------- 15.91s",
                                "magnum : Restart magnum-conductor container ---------------------------- 13.93s",
                                "magnum : Copying over magnum.conf -------------------------------------- 10.62s",
                                "service-ks-register : magnum | Creating endpoints ----------------------- 8.21s",
                                "service-ks-register : magnum | Creating users --------------------------- 4.90s",
                                "service-ks-register : magnum | Granting user roles ---------------------- 4.72s",
                                "magnum : Creating Magnum trustee user role ------------------------------ 4.40s",
                                "service-ks-register : magnum | Creating roles --------------------------- 4.37s",
                                "service-ks-register : magnum | Creating services ------------------------ 4.32s",
                                "magnum : Creating Magnum trustee user ----------------------------------- 4.21s",
                                "service-ks-register : magnum | Creating projects ------------------------ 4.07s",
                                "magnum : Creating Magnum trustee domain --------------------------------- 3.99s",
                                "magnum : Copying over config.json files for services -------------------- 3.83s",
                                "magnum : Creating Magnum database user and setting permissions ---------- 3.53s",
                                "service-cert-copy : magnum | Copying over extra CA certificates --------- 3.23s",
                                "magnum : Copying over kubeconfig file ----------------------------------- 2.74s",
                                "magnum : Creating Magnum database --------------------------------------- 2.68s",
                                "service-cert-copy : magnum | Copying over backend internal TLS key ------ 2.59s",
                                "magnum : Check magnum containers ---------------------------------------- 2.41s",
                                "\u001b[32m2026-02-24 01:07:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a29489db-95ad-418c-bc67-f87a0737c252 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:07:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44aa611b-f898-4352-b174-ada638d8f758 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:07:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 01:07:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fba6bf07-65a1-432e-9a3d-72209557a21d is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:07:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask eddb1a5f-448f-4517-b064-c30da006d691 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:07:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a29489db-95ad-418c-bc67-f87a0737c252 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:07:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44aa611b-f898-4352-b174-ada638d8f758 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:07:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 01:07:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fba6bf07-65a1-432e-9a3d-72209557a21d is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:07:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask eddb1a5f-448f-4517-b064-c30da006d691 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:07:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a29489db-95ad-418c-bc67-f87a0737c252 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:07:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44aa611b-f898-4352-b174-ada638d8f758 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:07:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 01:07:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fba6bf07-65a1-432e-9a3d-72209557a21d is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:07:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask eddb1a5f-448f-4517-b064-c30da006d691 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:07:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a29489db-95ad-418c-bc67-f87a0737c252 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:07:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44aa611b-f898-4352-b174-ada638d8f758 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:07:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 01:07:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fba6bf07-65a1-432e-9a3d-72209557a21d is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:07:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask eddb1a5f-448f-4517-b064-c30da006d691 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:07:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a29489db-95ad-418c-bc67-f87a0737c252 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:07:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44aa611b-f898-4352-b174-ada638d8f758 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:07:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 01:07:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fba6bf07-65a1-432e-9a3d-72209557a21d is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:07:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask eddb1a5f-448f-4517-b064-c30da006d691 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:07:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a29489db-95ad-418c-bc67-f87a0737c252 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:07:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4baf3a0c-387b-4215-b490-5edd972623e9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:07:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 44aa611b-f898-4352-b174-ada638d8f758 is in state SUCCESS\u001b[0m",
                                "\u001b[32m2026-02-24 01:07:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "",
                                "",
                                "PLAY [Group hosts based on configuration] **************************************",
                                "",
                                "TASK [Group hosts based on Kolla action] ***************************************",
                                "Tuesday 24 February 2026  01:02:32 +0000 (0:00:00.205)       0:00:00.205 ****** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Group hosts based on enabled services] ***********************************",
                                "Tuesday 24 February 2026  01:02:32 +0000 (0:00:00.706)       0:00:00.912 ****** ",
                                "\u001b[0;32mok: [testbed-node-0] => (item=enable_neutron_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item=enable_neutron_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item=enable_neutron_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item=enable_neutron_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item=enable_neutron_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item=enable_neutron_True)\u001b[0m",
                                "",
                                "PLAY [Apply role neutron] ******************************************************",
                                "",
                                "TASK [neutron : include_tasks] *************************************************",
                                "Tuesday 24 February 2026  01:02:33 +0000 (0:00:00.636)       0:00:01.549 ****** ",
                                "\u001b[0;36mincluded: /ansible/roles/neutron/tasks/deploy.yml for testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m",
                                "",
                                "TASK [neutron : Get container facts] *******************************************",
                                "Tuesday 24 February 2026  01:02:34 +0000 (0:00:01.217)       0:00:02.766 ****** ",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [neutron : Get container volume facts] ************************************",
                                "Tuesday 24 February 2026  01:02:35 +0000 (0:00:01.244)       0:00:04.010 ****** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [neutron : Check for ML2/OVN presence] ************************************",
                                "Tuesday 24 February 2026  01:02:37 +0000 (0:00:01.199)       0:00:05.210 ****** ",
                                "\u001b[0;32mok: [testbed-node-0] => {\u001b[0m",
                                "\u001b[0;32m    \"changed\": false,\u001b[0m",
                                "\u001b[0;32m    \"msg\": \"All assertions passed\"\u001b[0m",
                                "\u001b[0;32m}\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => {\u001b[0m",
                                "\u001b[0;32m    \"changed\": false,\u001b[0m",
                                "\u001b[0;32m    \"msg\": \"All assertions passed\"\u001b[0m",
                                "\u001b[0;32m}\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => {\u001b[0m",
                                "\u001b[0;32m    \"changed\": false,\u001b[0m",
                                "\u001b[0;32m    \"msg\": \"All assertions passed\"\u001b[0m",
                                "\u001b[0;32m}\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => {\u001b[0m",
                                "\u001b[0;32m    \"changed\": false,\u001b[0m",
                                "\u001b[0;32m    \"msg\": \"All assertions passed\"\u001b[0m",
                                "\u001b[0;32m}\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => {\u001b[0m",
                                "\u001b[0;32m    \"changed\": false,\u001b[0m",
                                "\u001b[0;32m    \"msg\": \"All assertions passed\"\u001b[0m",
                                "\u001b[0;32m}\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => {\u001b[0m",
                                "\u001b[0;32m    \"changed\": false,\u001b[0m",
                                "\u001b[0;32m    \"msg\": \"All assertions passed\"\u001b[0m",
                                "\u001b[0;32m}\u001b[0m",
                                "",
                                "TASK [neutron : Check for ML2/OVS presence] ************************************",
                                "Tuesday 24 February 2026  01:02:37 +0000 (0:00:00.827)       0:00:06.037 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [service-ks-register : neutron | Creating services] ***********************",
                                "Tuesday 24 February 2026  01:02:38 +0000 (0:00:00.694)       0:00:06.731 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=neutron (network))\u001b[0m",
                                "",
                                "TASK [service-ks-register : neutron | Creating endpoints] **********************",
                                "Tuesday 24 February 2026  01:02:42 +0000 (0:00:03.804)       0:00:10.536 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=neutron -> https://api-int.testbed.osism.xyz:9696 -> internal)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=neutron -> https://api.testbed.osism.xyz:9696 -> public)\u001b[0m",
                                "",
                                "TASK [service-ks-register : neutron | Creating projects] ***********************",
                                "Tuesday 24 February 2026  01:02:50 +0000 (0:00:07.910)       0:00:18.446 ****** ",
                                "\u001b[0;32mok: [testbed-node-0] => (item=service)\u001b[0m",
                                "",
                                "TASK [service-ks-register : neutron | Creating users] **************************",
                                "Tuesday 24 February 2026  01:02:54 +0000 (0:00:03.900)       0:00:22.346 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=neutron -> service)\u001b[0m",
                                "\u001b[1;35m[WARNING]: Module did not set no_log for update_password\u001b[0m",
                                "",
                                "TASK [service-ks-register : neutron | Creating roles] **************************",
                                "Tuesday 24 February 2026  01:02:58 +0000 (0:00:04.487)       0:00:26.834 ****** ",
                                "\u001b[0;32mok: [testbed-node-0] => (item=admin)\u001b[0m",
                                "",
                                "TASK [service-ks-register : neutron | Granting user roles] *********************",
                                "Tuesday 24 February 2026  01:03:02 +0000 (0:00:04.108)       0:00:30.943 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=neutron -> service -> admin)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=neutron -> service -> service)\u001b[0m",
                                "",
                                "TASK [neutron : include_tasks] *************************************************",
                                "Tuesday 24 February 2026  01:03:11 +0000 (0:00:08.840)       0:00:39.783 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Load and persist kernel modules] *****************************************",
                                "Tuesday 24 February 2026  01:03:12 +0000 (0:00:00.924)       0:00:40.707 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [neutron : Check IPv6 support] ********************************************",
                                "Tuesday 24 February 2026  01:03:15 +0000 (0:00:02.463)       0:00:43.171 ****** ",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Setting sysctl values] ***************************************************",
                                "Tuesday 24 February 2026  01:03:16 +0000 (0:00:01.462)       0:00:44.633 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [neutron : Ensuring config directories exist] *****************************",
                                "Tuesday 24 February 2026  01:03:20 +0000 (0:00:03.540)       0:00:48.174 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2024.2', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2024.2', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2024.2', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}})\u001b[0m",
                                "",
                                "TASK [neutron : Check if extra ml2 plugins exists] *****************************",
                                "Tuesday 24 February 2026  01:03:23 +0000 (0:00:03.577)       0:00:51.752 ****** ",
                                "\u001b[1;35m[WARNING]: Skipped\u001b[0m",
                                "\u001b[1;35m'/opt/configuration/environments/kolla/files/overlays/neutron/plugins/' path\u001b[0m",
                                "\u001b[1;35mdue to this access issue:\u001b[0m",
                                "\u001b[1;35m'/opt/configuration/environments/kolla/files/overlays/neutron/plugins/' is not\u001b[0m",
                                "\u001b[1;35ma directory\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0 -> localhost]\u001b[0m",
                                "",
                                "TASK [neutron : include_tasks] *************************************************",
                                "Tuesday 24 February 2026  01:03:24 +0000 (0:00:01.044)       0:00:52.796 ****** ",
                                "\u001b[0;36mincluded: /ansible/roles/neutron/tasks/copy-certs.yml for testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m",
                                "",
                                "TASK [service-cert-copy : neutron | Copying over extra CA certificates] ********",
                                "Tuesday 24 February 2026  01:03:26 +0000 (0:00:01.417)       0:00:54.214 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2024.2', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2024.2', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2024.2', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}})\u001b[0m",
                                "",
                                "TASK [service-cert-copy : neutron | Copying over backend internal TLS certificate] ***",
                                "Tuesday 24 February 2026  01:03:30 +0000 (0:00:04.087)       0:00:58.302 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2024.2', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2024.2', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2024.2', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [service-cert-copy : neutron | Copying over backend internal TLS key] *****",
                                "Tuesday 24 February 2026  01:03:34 +0000 (0:00:04.521)       0:01:02.823 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2024.2', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2024.2', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2024.2', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [neutron : Creating TLS backend PEM File] *********************************",
                                "Tuesday 24 February 2026  01:03:39 +0000 (0:00:04.500)       0:01:07.324 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [neutron : Check if policies shall be overwritten] ************************",
                                "Tuesday 24 February 2026  01:03:42 +0000 (0:00:03.343)       0:01:10.668 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [neutron : Set neutron policy file] ***************************************",
                                "Tuesday 24 February 2026  01:03:42 +0000 (0:00:00.137)       0:01:10.805 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [neutron : Copying over existing policy file] *****************************",
                                "Tuesday 24 February 2026  01:03:43 +0000 (0:00:00.802)       0:01:11.608 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2024.2', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2024.2', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2024.2', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [neutron : Copying over config.json files for services] *******************",
                                "Tuesday 24 February 2026  01:03:46 +0000 (0:00:03.366)       0:01:14.975 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2024.2', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2024.2', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2024.2', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}})\u001b[0m",
                                "",
                                "TASK [neutron : Copying over neutron.conf] *************************************",
                                "Tuesday 24 February 2026  01:03:51 +0000 (0:00:04.364)       0:01:19.340 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2024.2', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2024.2', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2024.2', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}})\u001b[0m",
                                "",
                                "TASK [neutron : Copying over neutron_vpnaas.conf] ******************************",
                                "Tuesday 24 February 2026  01:03:59 +0000 (0:00:08.285)       0:01:27.626 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2024.2', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2024.2', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2024.2', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [neutron : Copying over ssh key] ******************************************",
                                "Tuesday 24 February 2026  01:04:03 +0000 (0:00:04.009)       0:01:31.636 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [neutron : Copying over ml2_conf.ini] *************************************",
                                "Tuesday 24 February 2026  01:04:07 +0000 (0:00:03.835)       0:01:35.471 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2024.2', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2024.2', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2024.2', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}})\u001b[0m",
                                "",
                                "TASK [neutron : Copying over linuxbridge_agent.ini] ****************************",
                                "Tuesday 24 February 2026  01:04:14 +0000 (0:00:06.979)       0:01:42.451 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [neutron : Copying over openvswitch_agent.ini] ****************************",
                                "Tuesday 24 February 2026  01:04:18 +0000 (0:00:04.459)       0:01:46.910 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [neutron : Copying over sriov_agent.ini] **********************************",
                                "Tuesday 24 February 2026  01:04:22 +0000 (0:00:03.744)       0:01:50.654 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [neutron : Copying over mlnx_agent.ini] ***********************************",
                                "Tuesday 24 February 2026  01:04:25 +0000 (0:00:02.611)       0:01:53.266 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [neutron : Copying over eswitchd.conf] ************************************",
                                "Tuesday 24 February 2026  01:04:27 +0000 (0:00:02.467)       0:01:55.734 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [neutron : Copying over dhcp_agent.ini] ***********************************",
                                "Tuesday 24 February 2026  01:04:30 +0000 (0:00:03.238)       0:01:58.973 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [neutron : Copying over dnsmasq.conf] *************************************",
                                "Tuesday 24 February 2026  01:04:33 +0000 (0:00:02.532)       0:02:01.506 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=/ansible/roles/neutron/templates/dnsmasq.conf.j2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=/ansible/roles/neutron/templates/dnsmasq.conf.j2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=/ansible/roles/neutron/templates/dnsmasq.conf.j2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=/ansible/roles/neutron/templates/dnsmasq.conf.j2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=/ansible/roles/neutron/templates/dnsmasq.conf.j2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=/ansible/roles/neutron/templates/dnsmasq.conf.j2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [neutron : Copying over l3_agent.ini] *************************************",
                                "Tuesday 24 February 2026  01:04:35 +0000 (0:00:02.547)       0:02:04.053 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2024.2', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2024.2', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2024.2', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [neutron : Copying over fwaas_driver.ini] *********************************",
                                "Tuesday 24 February 2026  01:04:38 +0000 (0:00:02.698)       0:02:06.752 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2024.2', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2024.2', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2024.2', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [neutron : Copying over metadata_agent.ini] *******************************",
                                "Tuesday 24 February 2026  01:04:41 +0000 (0:00:03.254)       0:02:10.006 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [neutron : Copying over neutron_ovn_metadata_agent.ini] *******************",
                                "Tuesday 24 February 2026  01:04:44 +0000 (0:00:02.323)       0:02:12.330 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [neutron : Copying over metering_agent.ini] *******************************",
                                "Tuesday 24 February 2026  01:04:48 +0000 (0:00:04.486)       0:02:16.816 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [neutron : Copying over ironic_neutron_agent.ini] *************************",
                                "Tuesday 24 February 2026  01:04:52 +0000 (0:00:03.925)       0:02:20.742 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [neutron : Copying over bgp_dragent.ini] **********************************",
                                "Tuesday 24 February 2026  01:04:55 +0000 (0:00:03.198)       0:02:23.940 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [neutron : Copying over ovn_agent.ini] ************************************",
                                "Tuesday 24 February 2026  01:04:58 +0000 (0:00:02.808)       0:02:26.749 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [neutron : Copying over nsx.ini] ******************************************",
                                "Tuesday 24 February 2026  01:05:02 +0000 (0:00:03.648)       0:02:30.398 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [neutron : Copy neutron-l3-agent-wrapper script] **************************",
                                "Tuesday 24 February 2026  01:05:07 +0000 (0:00:05.364)       0:02:35.762 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [neutron : Copying over extra ml2 plugins] ********************************",
                                "Tuesday 24 February 2026  01:05:11 +0000 (0:00:03.693)       0:02:39.456 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [neutron : Copying over neutron-tls-proxy.cfg] ****************************",
                                "Tuesday 24 February 2026  01:05:14 +0000 (0:00:03.127)       0:02:42.584 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=/ansible/roles/neutron/templates/neutron-tls-proxy.cfg.j2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=/ansible/roles/neutron/templates/neutron-tls-proxy.cfg.j2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=/ansible/roles/neutron/templates/neutron-tls-proxy.cfg.j2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=/ansible/roles/neutron/templates/neutron-tls-proxy.cfg.j2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=/ansible/roles/neutron/templates/neutron-tls-proxy.cfg.j2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=/ansible/roles/neutron/templates/neutron-tls-proxy.cfg.j2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [neutron : Copying over neutron_taas.conf] ********************************",
                                "Tuesday 24 February 2026  01:05:19 +0000 (0:00:04.731)       0:02:47.315 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2024.2', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2024.2', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2024.2', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [neutron : Check neutron containers] **************************************",
                                "Tuesday 24 February 2026  01:05:23 +0000 (0:00:04.114)       0:02:51.429 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2024.2', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2024.2', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'neutron-server', 'value': {'container_name': 'neutron_server', 'image': 'registry.osism.tech/kolla/neutron-server:2024.2', 'enabled': True, 'group': 'neutron-server', 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9696'], 'timeout': '30'}, 'haproxy': {'neutron_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9696', 'listen_port': '9696'}, 'neutron_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9696', 'listen_port': '9696'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'neutron-ovn-metadata-agent', 'value': {'container_name': 'neutron_ovn_metadata_agent', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'privileged': True, 'enabled': True, 'host_in_groups': True, 'volumes': ['/etc/kolla/neutron-ovn-metadata-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', 'neutron_metadata_socket:/var/lib/neutron/kolla/', '/run/openvswitch:/run/openvswitch:shared', '/run/netns:/run/netns:shared', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port neutron-ovn-metadata-agent 6640'], 'timeout': '30'}}})\u001b[0m",
                                "",
                                "TASK [neutron : include_tasks] *************************************************",
                                "Tuesday 24 February 2026  01:05:26 +0000 (0:00:03.510)       0:02:54.939 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [neutron : Creating Neutron database] *************************************",
                                "Tuesday 24 February 2026  01:05:27 +0000 (0:00:00.608)       0:02:55.548 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [neutron : Creating Neutron database user and setting permissions] ********",
                                "Tuesday 24 February 2026  01:05:29 +0000 (0:00:02.512)       0:02:58.061 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [neutron : Running Neutron bootstrap container] ***************************",
                                "Tuesday 24 February 2026  01:05:32 +0000 (0:00:02.915)       0:03:00.976 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [neutron : Flush Handlers] ************************************************",
                                "Tuesday 24 February 2026  01:06:20 +0000 (0:00:47.724)       0:03:48.700 ****** ",
                                "",
                                "TASK [neutron : Flush Handlers] ************************************************",
                                "Tuesday 24 February 2026  01:06:20 +0000 (0:00:00.069)       0:03:48.770 ****** ",
                                "",
                                "TASK [neutron : Flush Handlers] ************************************************",
                                "Tuesday 24 February 2026  01:06:21 +0000 (0:00:00.397)       0:03:49.167 ****** ",
                                "",
                                "TASK [neutron : Flush Handlers] ************************************************",
                                "Tuesday 24 February 2026  01:06:21 +0000 (0:00:00.074)       0:03:49.241 ****** ",
                                "",
                                "TASK [neutron : Flush Handlers] ************************************************",
                                "Tuesday 24 February 2026  01:06:21 +0000 (0:00:00.120)       0:03:49.361 ****** ",
                                "",
                                "TASK [neutron : Flush Handlers] ************************************************",
                                "Tuesday 24 February 2026  01:06:21 +0000 (0:00:00.135)       0:03:49.496 ****** ",
                                "",
                                "RUNNING HANDLER [neutron : Restart neutron-server container] *******************",
                                "Tuesday 24 February 2026  01:06:21 +0000 (0:00:00.146)       0:03:49.643 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "",
                                "RUNNING HANDLER [neutron : Restart neutron-ovn-metadata-agent container] *******",
                                "Tuesday 24 February 2026  01:06:56 +0000 (0:00:34.425)       0:04:24.068 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=26  \u001b[0m \u001b[0;33mchanged=15  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=32  \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=16  \u001b[0m \u001b[0;33mchanged=8   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=31  \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=16  \u001b[0m \u001b[0;33mchanged=8   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=31  \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-3\u001b[0m             : \u001b[0;32mok=15  \u001b[0m \u001b[0;33mchanged=7   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=32  \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-4\u001b[0m             : \u001b[0;32mok=15  \u001b[0m \u001b[0;33mchanged=7   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=32  \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-5\u001b[0m             : \u001b[0;32mok=15  \u001b[0m \u001b[0;33mchanged=7   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=32  \u001b[0m rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Tuesday 24 February 2026  01:07:50 +0000 (0:00:54.110)       0:05:18.179 ****** ",
                                "=============================================================================== ",
                                "neutron : Restart neutron-ovn-metadata-agent container ----------------- 54.11s",
                                "neutron : Running Neutron bootstrap container -------------------------- 47.72s",
                                "neutron : Restart neutron-server container ----------------------------- 34.43s",
                                "service-ks-register : neutron | Granting user roles --------------------- 8.84s",
                                "neutron : Copying over neutron.conf ------------------------------------- 8.29s",
                                "service-ks-register : neutron | Creating endpoints ---------------------- 7.91s",
                                "neutron : Copying over ml2_conf.ini ------------------------------------- 6.98s",
                                "neutron : Copying over nsx.ini ------------------------------------------ 5.36s",
                                "neutron : Copying over neutron-tls-proxy.cfg ---------------------------- 4.73s",
                                "service-cert-copy : neutron | Copying over backend internal TLS certificate --- 4.52s",
                                "service-cert-copy : neutron | Copying over backend internal TLS key ----- 4.50s",
                                "service-ks-register : neutron | Creating users -------------------------- 4.49s",
                                "neutron : Copying over neutron_ovn_metadata_agent.ini ------------------- 4.49s",
                                "neutron : Copying over linuxbridge_agent.ini ---------------------------- 4.46s",
                                "neutron : Copying over config.json files for services ------------------- 4.36s",
                                "neutron : Copying over neutron_taas.conf -------------------------------- 4.11s",
                                "service-ks-register : neutron | Creating roles -------------------------- 4.11s",
                                "service-cert-copy : neutron | Copying over extra CA certificates -------- 4.09s",
                                "neutron : Copying over neutron_vpnaas.conf ------------------------------ 4.01s",
                                "neutron : Copying over metering_agent.ini ------------------------------- 3.93s",
                                "\u001b[32m2026-02-24 01:07:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fba6bf07-65a1-432e-9a3d-72209557a21d is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:07:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask eddb1a5f-448f-4517-b064-c30da006d691 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:07:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a29489db-95ad-418c-bc67-f87a0737c252 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:07:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4baf3a0c-387b-4215-b490-5edd972623e9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:07:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 01:07:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fba6bf07-65a1-432e-9a3d-72209557a21d is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:07:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask eddb1a5f-448f-4517-b064-c30da006d691 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:07:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a29489db-95ad-418c-bc67-f87a0737c252 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:07:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4baf3a0c-387b-4215-b490-5edd972623e9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:07:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 01:08:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fba6bf07-65a1-432e-9a3d-72209557a21d is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:08:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask eddb1a5f-448f-4517-b064-c30da006d691 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:08:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a29489db-95ad-418c-bc67-f87a0737c252 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:08:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4baf3a0c-387b-4215-b490-5edd972623e9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:08:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 01:08:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fba6bf07-65a1-432e-9a3d-72209557a21d is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:08:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask eddb1a5f-448f-4517-b064-c30da006d691 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:08:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a29489db-95ad-418c-bc67-f87a0737c252 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:08:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4baf3a0c-387b-4215-b490-5edd972623e9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:08:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 01:08:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fba6bf07-65a1-432e-9a3d-72209557a21d is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:08:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask eddb1a5f-448f-4517-b064-c30da006d691 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:08:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a29489db-95ad-418c-bc67-f87a0737c252 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:08:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4baf3a0c-387b-4215-b490-5edd972623e9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:08:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 01:08:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fba6bf07-65a1-432e-9a3d-72209557a21d is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:08:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask eddb1a5f-448f-4517-b064-c30da006d691 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:08:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a29489db-95ad-418c-bc67-f87a0737c252 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:08:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4baf3a0c-387b-4215-b490-5edd972623e9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:08:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 01:08:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fba6bf07-65a1-432e-9a3d-72209557a21d is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:08:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask eddb1a5f-448f-4517-b064-c30da006d691 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:08:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a29489db-95ad-418c-bc67-f87a0737c252 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:08:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4baf3a0c-387b-4215-b490-5edd972623e9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:08:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 01:08:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fba6bf07-65a1-432e-9a3d-72209557a21d is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:08:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask eddb1a5f-448f-4517-b064-c30da006d691 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:08:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a29489db-95ad-418c-bc67-f87a0737c252 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:08:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4baf3a0c-387b-4215-b490-5edd972623e9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:08:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 01:08:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fba6bf07-65a1-432e-9a3d-72209557a21d is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:08:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask eddb1a5f-448f-4517-b064-c30da006d691 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:08:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a29489db-95ad-418c-bc67-f87a0737c252 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:08:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4baf3a0c-387b-4215-b490-5edd972623e9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:08:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 01:08:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fba6bf07-65a1-432e-9a3d-72209557a21d is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:08:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask eddb1a5f-448f-4517-b064-c30da006d691 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:08:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a29489db-95ad-418c-bc67-f87a0737c252 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:08:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4baf3a0c-387b-4215-b490-5edd972623e9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:08:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 01:08:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fba6bf07-65a1-432e-9a3d-72209557a21d is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:08:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask eddb1a5f-448f-4517-b064-c30da006d691 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:08:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a29489db-95ad-418c-bc67-f87a0737c252 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:08:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4baf3a0c-387b-4215-b490-5edd972623e9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:08:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 01:08:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fba6bf07-65a1-432e-9a3d-72209557a21d is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:08:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask eddb1a5f-448f-4517-b064-c30da006d691 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:08:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a29489db-95ad-418c-bc67-f87a0737c252 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:08:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4baf3a0c-387b-4215-b490-5edd972623e9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:08:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 01:08:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fba6bf07-65a1-432e-9a3d-72209557a21d is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:08:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask eddb1a5f-448f-4517-b064-c30da006d691 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:08:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a29489db-95ad-418c-bc67-f87a0737c252 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:08:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4baf3a0c-387b-4215-b490-5edd972623e9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:08:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 01:08:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fba6bf07-65a1-432e-9a3d-72209557a21d is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:08:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask eddb1a5f-448f-4517-b064-c30da006d691 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:08:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a29489db-95ad-418c-bc67-f87a0737c252 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:08:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4baf3a0c-387b-4215-b490-5edd972623e9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:08:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 01:08:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fba6bf07-65a1-432e-9a3d-72209557a21d is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:08:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask eddb1a5f-448f-4517-b064-c30da006d691 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:08:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a29489db-95ad-418c-bc67-f87a0737c252 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:08:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4baf3a0c-387b-4215-b490-5edd972623e9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:08:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 01:08:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fba6bf07-65a1-432e-9a3d-72209557a21d is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:08:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask eddb1a5f-448f-4517-b064-c30da006d691 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:08:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a29489db-95ad-418c-bc67-f87a0737c252 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:08:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4baf3a0c-387b-4215-b490-5edd972623e9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:08:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 01:08:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fba6bf07-65a1-432e-9a3d-72209557a21d is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:08:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask eddb1a5f-448f-4517-b064-c30da006d691 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:08:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a29489db-95ad-418c-bc67-f87a0737c252 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:08:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4baf3a0c-387b-4215-b490-5edd972623e9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:08:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 01:08:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fba6bf07-65a1-432e-9a3d-72209557a21d is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:08:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask eddb1a5f-448f-4517-b064-c30da006d691 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:08:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a29489db-95ad-418c-bc67-f87a0737c252 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:08:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4baf3a0c-387b-4215-b490-5edd972623e9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:08:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 01:08:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fba6bf07-65a1-432e-9a3d-72209557a21d is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:08:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask eddb1a5f-448f-4517-b064-c30da006d691 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:08:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a29489db-95ad-418c-bc67-f87a0737c252 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:08:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4baf3a0c-387b-4215-b490-5edd972623e9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:08:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 01:08:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fba6bf07-65a1-432e-9a3d-72209557a21d is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:08:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask eddb1a5f-448f-4517-b064-c30da006d691 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:08:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a29489db-95ad-418c-bc67-f87a0737c252 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:08:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4baf3a0c-387b-4215-b490-5edd972623e9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:08:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 01:08:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fba6bf07-65a1-432e-9a3d-72209557a21d is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:08:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask eddb1a5f-448f-4517-b064-c30da006d691 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:08:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a29489db-95ad-418c-bc67-f87a0737c252 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:08:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4baf3a0c-387b-4215-b490-5edd972623e9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:08:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 01:08:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fba6bf07-65a1-432e-9a3d-72209557a21d is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:08:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask eddb1a5f-448f-4517-b064-c30da006d691 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:08:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a29489db-95ad-418c-bc67-f87a0737c252 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:08:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4baf3a0c-387b-4215-b490-5edd972623e9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:08:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 01:09:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fba6bf07-65a1-432e-9a3d-72209557a21d is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:09:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask eddb1a5f-448f-4517-b064-c30da006d691 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:09:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a29489db-95ad-418c-bc67-f87a0737c252 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:09:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4baf3a0c-387b-4215-b490-5edd972623e9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:09:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 01:09:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fba6bf07-65a1-432e-9a3d-72209557a21d is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:09:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask eddb1a5f-448f-4517-b064-c30da006d691 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:09:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a29489db-95ad-418c-bc67-f87a0737c252 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:09:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4baf3a0c-387b-4215-b490-5edd972623e9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:09:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 01:09:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fba6bf07-65a1-432e-9a3d-72209557a21d is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:09:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask eddb1a5f-448f-4517-b064-c30da006d691 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:09:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a29489db-95ad-418c-bc67-f87a0737c252 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:09:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4baf3a0c-387b-4215-b490-5edd972623e9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:09:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 01:09:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fba6bf07-65a1-432e-9a3d-72209557a21d is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:09:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask eddb1a5f-448f-4517-b064-c30da006d691 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:09:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a29489db-95ad-418c-bc67-f87a0737c252 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:09:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4baf3a0c-387b-4215-b490-5edd972623e9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:09:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 01:09:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fba6bf07-65a1-432e-9a3d-72209557a21d is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:09:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask eddb1a5f-448f-4517-b064-c30da006d691 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:09:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a29489db-95ad-418c-bc67-f87a0737c252 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:09:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4baf3a0c-387b-4215-b490-5edd972623e9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:09:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 01:09:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fba6bf07-65a1-432e-9a3d-72209557a21d is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:09:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask eddb1a5f-448f-4517-b064-c30da006d691 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:09:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a29489db-95ad-418c-bc67-f87a0737c252 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:09:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4baf3a0c-387b-4215-b490-5edd972623e9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:09:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 01:09:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fba6bf07-65a1-432e-9a3d-72209557a21d is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:09:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask eddb1a5f-448f-4517-b064-c30da006d691 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:09:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a29489db-95ad-418c-bc67-f87a0737c252 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:09:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4baf3a0c-387b-4215-b490-5edd972623e9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:09:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 01:09:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fba6bf07-65a1-432e-9a3d-72209557a21d is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:09:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask eddb1a5f-448f-4517-b064-c30da006d691 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:09:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a29489db-95ad-418c-bc67-f87a0737c252 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:09:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4baf3a0c-387b-4215-b490-5edd972623e9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:09:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 01:09:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fba6bf07-65a1-432e-9a3d-72209557a21d is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:09:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask eddb1a5f-448f-4517-b064-c30da006d691 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:09:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a29489db-95ad-418c-bc67-f87a0737c252 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:09:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4baf3a0c-387b-4215-b490-5edd972623e9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:09:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 01:09:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fba6bf07-65a1-432e-9a3d-72209557a21d is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:09:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask eddb1a5f-448f-4517-b064-c30da006d691 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:09:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a29489db-95ad-418c-bc67-f87a0737c252 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:09:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4baf3a0c-387b-4215-b490-5edd972623e9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:09:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 01:09:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fba6bf07-65a1-432e-9a3d-72209557a21d is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:09:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask eddb1a5f-448f-4517-b064-c30da006d691 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:09:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a29489db-95ad-418c-bc67-f87a0737c252 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:09:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4baf3a0c-387b-4215-b490-5edd972623e9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:09:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 01:09:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fba6bf07-65a1-432e-9a3d-72209557a21d is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:09:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask eddb1a5f-448f-4517-b064-c30da006d691 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:09:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a29489db-95ad-418c-bc67-f87a0737c252 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:09:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4baf3a0c-387b-4215-b490-5edd972623e9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:09:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 01:09:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fba6bf07-65a1-432e-9a3d-72209557a21d is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:09:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask eddb1a5f-448f-4517-b064-c30da006d691 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:09:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a29489db-95ad-418c-bc67-f87a0737c252 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:09:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4baf3a0c-387b-4215-b490-5edd972623e9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:09:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 01:09:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fba6bf07-65a1-432e-9a3d-72209557a21d is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:09:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask eddb1a5f-448f-4517-b064-c30da006d691 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:09:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a29489db-95ad-418c-bc67-f87a0737c252 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:09:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4baf3a0c-387b-4215-b490-5edd972623e9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:09:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 01:09:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fba6bf07-65a1-432e-9a3d-72209557a21d is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:09:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask eddb1a5f-448f-4517-b064-c30da006d691 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:09:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a29489db-95ad-418c-bc67-f87a0737c252 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:09:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4baf3a0c-387b-4215-b490-5edd972623e9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:09:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 01:09:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fba6bf07-65a1-432e-9a3d-72209557a21d is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:09:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask eddb1a5f-448f-4517-b064-c30da006d691 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:09:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask de6aa5ad-1dfd-4a58-bf00-4755d79ef102 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:09:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask a29489db-95ad-418c-bc67-f87a0737c252 is in state SUCCESS\u001b[0m",
                                "",
                                "",
                                "PLAY [Group hosts based on configuration] **************************************",
                                "",
                                "TASK [Group hosts based on Kolla action] ***************************************",
                                "Tuesday 24 February 2026  01:06:05 +0000 (0:00:00.293)       0:00:00.293 ****** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Group hosts based on enabled services] ***********************************",
                                "Tuesday 24 February 2026  01:06:06 +0000 (0:00:01.006)       0:00:01.300 ****** ",
                                "\u001b[0;32mok: [testbed-manager] => (item=enable_prometheus_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0] => (item=enable_prometheus_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item=enable_prometheus_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item=enable_prometheus_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item=enable_prometheus_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item=enable_prometheus_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item=enable_prometheus_True)\u001b[0m",
                                "",
                                "PLAY [Apply role prometheus] ***************************************************",
                                "",
                                "TASK [prometheus : include_tasks] **********************************************",
                                "Tuesday 24 February 2026  01:06:06 +0000 (0:00:00.776)       0:00:02.077 ****** ",
                                "\u001b[0;36mincluded: /ansible/roles/prometheus/tasks/deploy.yml for testbed-manager, testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m",
                                "",
                                "TASK [prometheus : Ensuring config directories exist] **************************",
                                "Tuesday 24 February 2026  01:06:08 +0000 (0:00:01.785)       0:00:03.863 ****** ",
                                "\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'prometheus-server', 'value': {'container_name': 'prometheus_server', 'group': 'prometheus', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-v2-server:2024.2', 'volumes': ['/etc/kolla/prometheus-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'prometheus_v2:/var/lib/prometheus', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'prometheus_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9091', 'active_passive': True}, 'prometheus_server_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9091', 'listen_port': '9091', 'active_passive': True}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2024.2', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2024.2', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2024.2', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2024.2', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'prometheus-mysqld-exporter', 'value': {'container_name': 'prometheus_mysqld_exporter', 'group': 'prometheus-mysqld-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-mysqld-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-mysqld-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'prometheus-mysqld-exporter', 'value': {'container_name': 'prometheus_mysqld_exporter', 'group': 'prometheus-mysqld-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-mysqld-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-mysqld-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2024.2', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'prometheus-mysqld-exporter', 'value': {'container_name': 'prometheus_mysqld_exporter', 'group': 'prometheus-mysqld-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-mysqld-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-mysqld-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2024.2', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'prometheus-memcached-exporter', 'value': {'container_name': 'prometheus_memcached_exporter', 'group': 'prometheus-memcached-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-memcached-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-memcached-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2024.2', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'prometheus-memcached-exporter', 'value': {'container_name': 'prometheus_memcached_exporter', 'group': 'prometheus-memcached-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-memcached-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-memcached-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2024.2', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'prometheus-alertmanager', 'value': {'container_name': 'prometheus_alertmanager', 'group': 'prometheus-alertmanager', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-alertmanager:2024.2', 'volumes': ['/etc/kolla/prometheus-alertmanager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'prometheus:/var/lib/prometheus'], 'dimensions': {}, 'haproxy': {'prometheus_alertmanager': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True}, 'prometheus_alertmanager_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9093', 'listen_port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2024.2', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'prometheus-memcached-exporter', 'value': {'container_name': 'prometheus_memcached_exporter', 'group': 'prometheus-memcached-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-memcached-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-memcached-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2024.2', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2024.2', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2024.2', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'prometheus-libvirt-exporter', 'value': {'container_name': 'prometheus_libvirt_exporter', 'group': 'prometheus-libvirt-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-libvirt-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-libvirt-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run/libvirt:/run/libvirt:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'prometheus-blackbox-exporter', 'value': {'container_name': 'prometheus_blackbox_exporter', 'group': 'prometheus-blackbox-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-blackbox-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-blackbox-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2024.2', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2024.2', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'prometheus-elasticsearch-exporter', 'value': {'container_name': 'prometheus_elasticsearch_exporter', 'group': 'prometheus-elasticsearch-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-elasticsearch-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-elasticsearch-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'prometheus-elasticsearch-exporter', 'value': {'container_name': 'prometheus_elasticsearch_exporter', 'group': 'prometheus-elasticsearch-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-elasticsearch-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-elasticsearch-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'prometheus-libvirt-exporter', 'value': {'container_name': 'prometheus_libvirt_exporter', 'group': 'prometheus-libvirt-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-libvirt-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-libvirt-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run/libvirt:/run/libvirt:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'prometheus-libvirt-exporter', 'value': {'container_name': 'prometheus_libvirt_exporter', 'group': 'prometheus-libvirt-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-libvirt-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-libvirt-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run/libvirt:/run/libvirt:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'prometheus-elasticsearch-exporter', 'value': {'container_name': 'prometheus_elasticsearch_exporter', 'group': 'prometheus-elasticsearch-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-elasticsearch-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-elasticsearch-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "",
                                "TASK [prometheus : include_tasks] **********************************************",
                                "Tuesday 24 February 2026  01:06:11 +0000 (0:00:03.323)       0:00:07.187 ****** ",
                                "\u001b[0;36mincluded: /ansible/roles/prometheus/tasks/copy-certs.yml for testbed-manager, testbed-node-0, testbed-node-1, testbed-node-2, testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m",
                                "",
                                "TASK [service-cert-copy : prometheus | Copying over extra CA certificates] *****",
                                "Tuesday 24 February 2026  01:06:13 +0000 (0:00:01.666)       0:00:08.854 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2024.2', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'prometheus-server', 'value': {'container_name': 'prometheus_server', 'group': 'prometheus', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-v2-server:2024.2', 'volumes': ['/etc/kolla/prometheus-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'prometheus_v2:/var/lib/prometheus', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'prometheus_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9091', 'active_passive': True}, 'prometheus_server_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9091', 'listen_port': '9091', 'active_passive': True}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2024.2', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2024.2', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2024.2', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2024.2', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2024.2', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'prometheus-mysqld-exporter', 'value': {'container_name': 'prometheus_mysqld_exporter', 'group': 'prometheus-mysqld-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-mysqld-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-mysqld-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2024.2', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2024.2', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'prometheus-mysqld-exporter', 'value': {'container_name': 'prometheus_mysqld_exporter', 'group': 'prometheus-mysqld-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-mysqld-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-mysqld-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'prometheus-mysqld-exporter', 'value': {'container_name': 'prometheus_mysqld_exporter', 'group': 'prometheus-mysqld-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-mysqld-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-mysqld-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2024.2', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2024.2', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'prometheus-memcached-exporter', 'value': {'container_name': 'prometheus_memcached_exporter', 'group': 'prometheus-memcached-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-memcached-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-memcached-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'prometheus-libvirt-exporter', 'value': {'container_name': 'prometheus_libvirt_exporter', 'group': 'prometheus-libvirt-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-libvirt-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-libvirt-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run/libvirt:/run/libvirt:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'prometheus-memcached-exporter', 'value': {'container_name': 'prometheus_memcached_exporter', 'group': 'prometheus-memcached-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-memcached-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-memcached-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2024.2', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'prometheus-memcached-exporter', 'value': {'container_name': 'prometheus_memcached_exporter', 'group': 'prometheus-memcached-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-memcached-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-memcached-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'prometheus-libvirt-exporter', 'value': {'container_name': 'prometheus_libvirt_exporter', 'group': 'prometheus-libvirt-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-libvirt-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-libvirt-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run/libvirt:/run/libvirt:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'prometheus-alertmanager', 'value': {'container_name': 'prometheus_alertmanager', 'group': 'prometheus-alertmanager', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-alertmanager:2024.2', 'volumes': ['/etc/kolla/prometheus-alertmanager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'prometheus:/var/lib/prometheus'], 'dimensions': {}, 'haproxy': {'prometheus_alertmanager': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True}, 'prometheus_alertmanager_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9093', 'listen_port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2024.2', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'prometheus-libvirt-exporter', 'value': {'container_name': 'prometheus_libvirt_exporter', 'group': 'prometheus-libvirt-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-libvirt-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-libvirt-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run/libvirt:/run/libvirt:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2024.2', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2024.2', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'prometheus-blackbox-exporter', 'value': {'container_name': 'prometheus_blackbox_exporter', 'group': 'prometheus-blackbox-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-blackbox-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-blackbox-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'prometheus-elasticsearch-exporter', 'value': {'container_name': 'prometheus_elasticsearch_exporter', 'group': 'prometheus-elasticsearch-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-elasticsearch-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-elasticsearch-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'prometheus-elasticsearch-exporter', 'value': {'container_name': 'prometheus_elasticsearch_exporter', 'group': 'prometheus-elasticsearch-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-elasticsearch-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-elasticsearch-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'prometheus-elasticsearch-exporter', 'value': {'container_name': 'prometheus_elasticsearch_exporter', 'group': 'prometheus-elasticsearch-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-elasticsearch-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-elasticsearch-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "",
                                "TASK [service-cert-copy : prometheus | Copying over backend internal TLS certificate] ***",
                                "Tuesday 24 February 2026  01:06:20 +0000 (0:00:07.061)       0:00:15.915 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2024.2', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-mysqld-exporter', 'value': {'container_name': 'prometheus_mysqld_exporter', 'group': 'prometheus-mysqld-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-mysqld-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-mysqld-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-memcached-exporter', 'value': {'container_name': 'prometheus_memcached_exporter', 'group': 'prometheus-memcached-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-memcached-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-memcached-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2024.2', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-elasticsearch-exporter', 'value': {'container_name': 'prometheus_elasticsearch_exporter', 'group': 'prometheus-elasticsearch-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-elasticsearch-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-elasticsearch-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager] => (item={'key': 'prometheus-server', 'value': {'container_name': 'prometheus_server', 'group': 'prometheus', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-v2-server:2024.2', 'volumes': ['/etc/kolla/prometheus-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'prometheus_v2:/var/lib/prometheus', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'prometheus_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9091', 'active_passive': True}, 'prometheus_server_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9091', 'listen_port': '9091', 'active_passive': True}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2024.2', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2024.2', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager] => (item={'key': 'prometheus-alertmanager', 'value': {'container_name': 'prometheus_alertmanager', 'group': 'prometheus-alertmanager', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-alertmanager:2024.2', 'volumes': ['/etc/kolla/prometheus-alertmanager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'prometheus:/var/lib/prometheus'], 'dimensions': {}, 'haproxy': {'prometheus_alertmanager': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True}, 'prometheus_alertmanager_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9093', 'listen_port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager] => (item={'key': 'prometheus-blackbox-exporter', 'value': {'container_name': 'prometheus_blackbox_exporter', 'group': 'prometheus-blackbox-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-blackbox-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-blackbox-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2024.2', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-mysqld-exporter', 'value': {'container_name': 'prometheus_mysqld_exporter', 'group': 'prometheus-mysqld-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-mysqld-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-mysqld-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-memcached-exporter', 'value': {'container_name': 'prometheus_memcached_exporter', 'group': 'prometheus-memcached-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-memcached-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-memcached-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2024.2', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-elasticsearch-exporter', 'value': {'container_name': 'prometheus_elasticsearch_exporter', 'group': 'prometheus-elasticsearch-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-elasticsearch-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-elasticsearch-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2024.2', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-mysqld-exporter', 'value': {'container_name': 'prometheus_mysqld_exporter', 'group': 'prometheus-mysqld-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-mysqld-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-mysqld-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-memcached-exporter', 'value': {'container_name': 'prometheus_memcached_exporter', 'group': 'prometheus-memcached-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-memcached-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-memcached-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2024.2', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-elasticsearch-exporter', 'value': {'container_name': 'prometheus_elasticsearch_exporter', 'group': 'prometheus-elasticsearch-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-elasticsearch-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-elasticsearch-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2024.2', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2024.2', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'prometheus-libvirt-exporter', 'value': {'container_name': 'prometheus_libvirt_exporter', 'group': 'prometheus-libvirt-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-libvirt-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-libvirt-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run/libvirt:/run/libvirt:ro'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2024.2', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2024.2', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'prometheus-libvirt-exporter', 'value': {'container_name': 'prometheus_libvirt_exporter', 'group': 'prometheus-libvirt-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-libvirt-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-libvirt-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run/libvirt:/run/libvirt:ro'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2024.2', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2024.2', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'prometheus-libvirt-exporter', 'value': {'container_name': 'prometheus_libvirt_exporter', 'group': 'prometheus-libvirt-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-libvirt-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-libvirt-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run/libvirt:/run/libvirt:ro'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [service-cert-copy : prometheus | Copying over backend internal TLS key] ***",
                                "Tuesday 24 February 2026  01:06:24 +0000 (0:00:03.620)       0:00:19.536 ****** ",
                                "\u001b[0;36mskipping: [testbed-manager] => (item={'key': 'prometheus-server', 'value': {'container_name': 'prometheus_server', 'group': 'prometheus', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-v2-server:2024.2', 'volumes': ['/etc/kolla/prometheus-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'prometheus_v2:/var/lib/prometheus', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'prometheus_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9091', 'active_passive': True}, 'prometheus_server_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9091', 'listen_port': '9091', 'active_passive': True}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2024.2', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2024.2', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager] => (item={'key': 'prometheus-alertmanager', 'value': {'container_name': 'prometheus_alertmanager', 'group': 'prometheus-alertmanager', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-alertmanager:2024.2', 'volumes': ['/etc/kolla/prometheus-alertmanager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'prometheus:/var/lib/prometheus'], 'dimensions': {}, 'haproxy': {'prometheus_alertmanager': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True}, 'prometheus_alertmanager_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9093', 'listen_port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager] => (item={'key': 'prometheus-blackbox-exporter', 'value': {'container_name': 'prometheus_blackbox_exporter', 'group': 'prometheus-blackbox-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-blackbox-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-blackbox-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2024.2', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-mysqld-exporter', 'value': {'container_name': 'prometheus_mysqld_exporter', 'group': 'prometheus-mysqld-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-mysqld-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-mysqld-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-memcached-exporter', 'value': {'container_name': 'prometheus_memcached_exporter', 'group': 'prometheus-memcached-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-memcached-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-memcached-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2024.2', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'prometheus-elasticsearch-exporter', 'value': {'container_name': 'prometheus_elasticsearch_exporter', 'group': 'prometheus-elasticsearch-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-elasticsearch-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-elasticsearch-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2024.2', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-mysqld-exporter', 'value': {'container_name': 'prometheus_mysqld_exporter', 'group': 'prometheus-mysqld-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-mysqld-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-mysqld-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-memcached-exporter', 'value': {'container_name': 'prometheus_memcached_exporter', 'group': 'prometheus-memcached-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-memcached-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-memcached-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2024.2', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'prometheus-elasticsearch-exporter', 'value': {'container_name': 'prometheus_elasticsearch_exporter', 'group': 'prometheus-elasticsearch-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-elasticsearch-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-elasticsearch-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2024.2', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-mysqld-exporter', 'value': {'container_name': 'prometheus_mysqld_exporter', 'group': 'prometheus-mysqld-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-mysqld-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-mysqld-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-memcached-exporter', 'value': {'container_name': 'prometheus_memcached_exporter', 'group': 'prometheus-memcached-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-memcached-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-memcached-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2024.2', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'prometheus-elasticsearch-exporter', 'value': {'container_name': 'prometheus_elasticsearch_exporter', 'group': 'prometheus-elasticsearch-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-elasticsearch-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-elasticsearch-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2024.2', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2024.2', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'prometheus-libvirt-exporter', 'value': {'container_name': 'prometheus_libvirt_exporter', 'group': 'prometheus-libvirt-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-libvirt-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-libvirt-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run/libvirt:/run/libvirt:ro'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2024.2', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2024.2', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'prometheus-libvirt-exporter', 'value': {'container_name': 'prometheus_libvirt_exporter', 'group': 'prometheus-libvirt-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-libvirt-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-libvirt-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run/libvirt:/run/libvirt:ro'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2024.2', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2024.2', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'prometheus-libvirt-exporter', 'value': {'container_name': 'prometheus_libvirt_exporter', 'group': 'prometheus-libvirt-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-libvirt-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-libvirt-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run/libvirt:/run/libvirt:ro'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [prometheus : Copying over config.json files] *****************************",
                                "Tuesday 24 February 2026  01:06:28 +0000 (0:00:04.616)       0:00:24.153 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2024.2', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2024.2', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2024.2', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'prometheus-server', 'value': {'container_name': 'prometheus_server', 'group': 'prometheus', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-v2-server:2024.2', 'volumes': ['/etc/kolla/prometheus-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'prometheus_v2:/var/lib/prometheus', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'prometheus_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9091', 'active_passive': True}, 'prometheus_server_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9091', 'listen_port': '9091', 'active_passive': True}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'prometheus-mysqld-exporter', 'value': {'container_name': 'prometheus_mysqld_exporter', 'group': 'prometheus-mysqld-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-mysqld-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-mysqld-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2024.2', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2024.2', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'prometheus-mysqld-exporter', 'value': {'container_name': 'prometheus_mysqld_exporter', 'group': 'prometheus-mysqld-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-mysqld-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-mysqld-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2024.2', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'prometheus-memcached-exporter', 'value': {'container_name': 'prometheus_memcached_exporter', 'group': 'prometheus-memcached-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-memcached-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-memcached-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2024.2', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2024.2', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2024.2', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'prometheus-memcached-exporter', 'value': {'container_name': 'prometheus_memcached_exporter', 'group': 'prometheus-memcached-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-memcached-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-memcached-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'prometheus-mysqld-exporter', 'value': {'container_name': 'prometheus_mysqld_exporter', 'group': 'prometheus-mysqld-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-mysqld-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-mysqld-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'prometheus-libvirt-exporter', 'value': {'container_name': 'prometheus_libvirt_exporter', 'group': 'prometheus-libvirt-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-libvirt-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-libvirt-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run/libvirt:/run/libvirt:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2024.2', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2024.2', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2024.2', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2024.2', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'prometheus-alertmanager', 'value': {'container_name': 'prometheus_alertmanager', 'group': 'prometheus-alertmanager', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-alertmanager:2024.2', 'volumes': ['/etc/kolla/prometheus-alertmanager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'prometheus:/var/lib/prometheus'], 'dimensions': {}, 'haproxy': {'prometheus_alertmanager': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True}, 'prometheus_alertmanager_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9093', 'listen_port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'prometheus-memcached-exporter', 'value': {'container_name': 'prometheus_memcached_exporter', 'group': 'prometheus-memcached-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-memcached-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-memcached-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'prometheus-libvirt-exporter', 'value': {'container_name': 'prometheus_libvirt_exporter', 'group': 'prometheus-libvirt-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-libvirt-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-libvirt-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run/libvirt:/run/libvirt:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'prometheus-libvirt-exporter', 'value': {'container_name': 'prometheus_libvirt_exporter', 'group': 'prometheus-libvirt-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-libvirt-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-libvirt-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run/libvirt:/run/libvirt:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'prometheus-elasticsearch-exporter', 'value': {'container_name': 'prometheus_elasticsearch_exporter', 'group': 'prometheus-elasticsearch-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-elasticsearch-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-elasticsearch-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'prometheus-elasticsearch-exporter', 'value': {'container_name': 'prometheus_elasticsearch_exporter', 'group': 'prometheus-elasticsearch-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-elasticsearch-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-elasticsearch-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'prometheus-blackbox-exporter', 'value': {'container_name': 'prometheus_blackbox_exporter', 'group': 'prometheus-blackbox-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-blackbox-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-blackbox-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2024.2', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'prometheus-elasticsearch-exporter', 'value': {'container_name': 'prometheus_elasticsearch_exporter', 'group': 'prometheus-elasticsearch-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-elasticsearch-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-elasticsearch-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "",
                                "TASK [prometheus : Find custom prometheus alert rules files] *******************",
                                "Tuesday 24 February 2026  01:06:38 +0000 (0:00:09.707)       0:00:33.860 ****** ",
                                "\u001b[0;32mok: [testbed-manager -> localhost]\u001b[0m",
                                "",
                                "TASK [prometheus : Copying over custom prometheus alert rules files] ***********",
                                "Tuesday 24 February 2026  01:06:40 +0000 (0:00:01.708)       0:00:35.568 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'path': '/operations/prometheus/fluentd-aggregator.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 996, 'inode': 1317250, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.3637931, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'path': '/operations/prometheus/fluentd-aggregator.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 996, 'inode': 1317250, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.3637931, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'path': '/operations/prometheus/fluentd-aggregator.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 996, 'inode': 1317250, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.3637931, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'path': '/operations/prometheus/prometheus.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 12944, 'inode': 1317263, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.368886, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'path': '/operations/prometheus/prometheus.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 12944, 'inode': 1317263, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.368886, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'path': '/operations/prometheus/ceph.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 56929, 'inode': 1317246, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.362469, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'path': '/operations/prometheus/prometheus.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 12944, 'inode': 1317263, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.368886, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item={'path': '/operations/prometheus/fluentd-aggregator.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 996, 'inode': 1317250, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.3637931, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'path': '/operations/prometheus/fluentd-aggregator.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 996, 'inode': 1317250, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.3637931, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'path': '/operations/prometheus/openstack.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 12293, 'inode': 1317258, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.3675466, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'path': '/operations/prometheus/ceph.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 56929, 'inode': 1317246, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.362469, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'path': '/operations/prometheus/fluentd-aggregator.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 996, 'inode': 1317250, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.3637931, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'path': '/operations/prometheus/fluentd-aggregator.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 996, 'inode': 1317250, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.3637931, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'path': '/operations/prometheus/ceph.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 56929, 'inode': 1317246, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.362469, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'path': '/operations/prometheus/cadvisor.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3900, 'inode': 1317244, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.3611212, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'path': '/operations/prometheus/haproxy.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 7933, 'inode': 1317251, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.3647292, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'path': '/operations/prometheus/openstack.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 12293, 'inode': 1317258, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.3675466, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'path': '/operations/prometheus/prometheus.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 12944, 'inode': 1317263, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.368886, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'path': '/operations/prometheus/prometheus.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 12944, 'inode': 1317263, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.368886, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'path': '/operations/prometheus/openstack.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 12293, 'inode': 1317258, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.3675466, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'path': '/operations/prometheus/prometheus.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 12944, 'inode': 1317263, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.368886, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'path': '/operations/prometheus/cadvisor.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3900, 'inode': 1317244, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.3611212, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'path': '/operations/prometheus/ceph.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 56929, 'inode': 1317246, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.362469, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'path': '/operations/prometheus/node.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 14018, 'inode': 1317257, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.366479, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'path': '/operations/prometheus/ceph.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 56929, 'inode': 1317246, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.362469, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'path': '/operations/prometheus/cadvisor.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3900, 'inode': 1317244, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.3611212, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'path': '/operations/prometheus/haproxy.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 7933, 'inode': 1317251, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.3647292, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item={'path': '/operations/prometheus/prometheus.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 12944, 'inode': 1317263, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.368886, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'path': '/operations/prometheus/ceph.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 56929, 'inode': 1317246, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.362469, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'path': '/operations/prometheus/hardware.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 5593, 'inode': 1317253, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.36546, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'path': '/operations/prometheus/openstack.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 12293, 'inode': 1317258, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.3675466, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'path': '/operations/prometheus/openstack.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 12293, 'inode': 1317258, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.3675466, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'path': '/operations/prometheus/haproxy.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 7933, 'inode': 1317251, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.3647292, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'path': '/operations/prometheus/cadvisor.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3900, 'inode': 1317244, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.3611212, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'path': '/operations/prometheus/node.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 14018, 'inode': 1317257, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.366479, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'path': '/operations/prometheus/elasticsearch.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 5987, 'inode': 1317249, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.3633637, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'path': '/operations/prometheus/cadvisor.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3900, 'inode': 1317244, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.3611212, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'path': '/operations/prometheus/prometheus.rec.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3, 'inode': 1317262, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.368623, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'path': '/operations/prometheus/openstack.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 12293, 'inode': 1317258, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.3675466, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'path': '/operations/prometheus/haproxy.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 7933, 'inode': 1317251, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.3647292, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'path': '/operations/prometheus/cadvisor.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3900, 'inode': 1317244, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.3611212, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item={'path': '/operations/prometheus/ceph.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 56929, 'inode': 1317246, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.362469, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'path': '/operations/prometheus/alertmanager.rec.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3, 'inode': 1317241, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.3604016, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'path': '/operations/prometheus/haproxy.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 7933, 'inode': 1317251, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.3647292, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'path': '/operations/prometheus/redfish.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 334, 'inode': 1317272, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.371548, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'path': '/operations/prometheus/haproxy.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 7933, 'inode': 1317251, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.3647292, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'path': '/operations/prometheus/node.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 14018, 'inode': 1317257, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.366479, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'path': '/operations/prometheus/node.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 14018, 'inode': 1317257, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.366479, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'path': '/operations/prometheus/hardware.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 5593, 'inode': 1317253, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.36546, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'path': '/operations/prometheus/node.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 14018, 'inode': 1317257, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.366479, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'path': '/operations/prometheus/node.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 14018, 'inode': 1317257, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.366479, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'path': '/operations/prometheus/hardware.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 5593, 'inode': 1317253, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.36546, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'path': '/operations/prometheus/prometheus-extra.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 7408, 'inode': 1317260, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.3682377, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'path': '/operations/prometheus/elasticsearch.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 5987, 'inode': 1317249, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.3633637, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'path': '/operations/prometheus/hardware.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 5593, 'inode': 1317253, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.36546, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'path': '/operations/prometheus/hardware.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 5593, 'inode': 1317253, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.36546, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'path': '/operations/prometheus/hardware.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 5593, 'inode': 1317253, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.36546, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'path': '/operations/prometheus/ceph.rec.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3, 'inode': 1317245, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.361316, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'path': '/operations/prometheus/elasticsearch.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 5987, 'inode': 1317249, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.3633637, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'path': '/operations/prometheus/elasticsearch.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 5987, 'inode': 1317249, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.3633637, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'path': '/operations/prometheus/elasticsearch.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 5987, 'inode': 1317249, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.3633637, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'path': '/operations/prometheus/elasticsearch.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 5987, 'inode': 1317249, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.3633637, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'path': '/operations/prometheus/prometheus.rec.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3, 'inode': 1317262, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.368623, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item={'path': '/operations/prometheus/openstack.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 12293, 'inode': 1317258, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.3675466, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'path': '/operations/prometheus/prometheus.rec.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3, 'inode': 1317262, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.368623, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'path': '/operations/prometheus/prometheus.rec.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3, 'inode': 1317262, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.368623, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'path': '/operations/prometheus/alertmanager.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 5065, 'inode': 1317242, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.3607938, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'path': '/operations/prometheus/prometheus.rec.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3, 'inode': 1317262, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.368623, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'path': '/operations/prometheus/alertmanager.rec.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3, 'inode': 1317241, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.3604016, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'path': '/operations/prometheus/prometheus.rec.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3, 'inode': 1317262, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.368623, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'path': '/operations/prometheus/alertmanager.rec.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3, 'inode': 1317241, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.3604016, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'path': '/operations/prometheus/alertmanager.rec.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3, 'inode': 1317241, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.3604016, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'path': '/operations/prometheus/alertmanager.rec.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3, 'inode': 1317241, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.3604016, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'path': '/operations/prometheus/redfish.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 334, 'inode': 1317272, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.371548, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'path': '/operations/prometheus/node.rec.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 2309, 'inode': 1317256, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.3662472, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'path': '/operations/prometheus/alertmanager.rec.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3, 'inode': 1317241, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.3604016, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'path': '/operations/prometheus/redfish.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 334, 'inode': 1317272, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.371548, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'path': '/operations/prometheus/mysql.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3792, 'inode': 1317255, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.3659542, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'path': '/operations/prometheus/redfish.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 334, 'inode': 1317272, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.371548, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item={'path': '/operations/prometheus/cadvisor.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3900, 'inode': 1317244, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.3611212, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'path': '/operations/prometheus/redfish.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 334, 'inode': 1317272, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.371548, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'path': '/operations/prometheus/rabbitmq.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3539, 'inode': 1317271, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.3709881, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'path': '/operations/prometheus/prometheus-extra.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 7408, 'inode': 1317260, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.3682377, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'path': '/operations/prometheus/prometheus-extra.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 7408, 'inode': 1317260, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.3682377, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'path': '/operations/prometheus/redfish.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 334, 'inode': 1317272, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.371548, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'path': '/operations/prometheus/prometheus-extra.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 7408, 'inode': 1317260, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.3682377, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'path': '/operations/prometheus/prometheus-extra.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 7408, 'inode': 1317260, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.3682377, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'path': '/operations/prometheus/ceph.rec.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3, 'inode': 1317245, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.361316, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'path': '/operations/prometheus/ceph.rec.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3, 'inode': 1317245, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.361316, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'path': '/operations/prometheus/ceph.rec.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3, 'inode': 1317245, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.361316, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'path': '/operations/prometheus/alertmanager.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 5065, 'inode': 1317242, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.3607938, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'path': '/operations/prometheus/prometheus-extra.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 7408, 'inode': 1317260, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.3682377, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'path': '/operations/prometheus/ceph.rec.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3, 'inode': 1317245, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.361316, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'path': '/operations/prometheus/alertmanager.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 5065, 'inode': 1317242, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.3607938, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item={'path': '/operations/prometheus/haproxy.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 7933, 'inode': 1317251, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.3647292, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'path': '/operations/prometheus/alertmanager.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 5065, 'inode': 1317242, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.3607938, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'path': '/operations/prometheus/node.rec.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 2309, 'inode': 1317256, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.3662472, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'path': '/operations/prometheus/node.rec.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 2309, 'inode': 1317256, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.3662472, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'path': '/operations/prometheus/ceph.rec.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3, 'inode': 1317245, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.361316, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'path': '/operations/prometheus/node.rec.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 2309, 'inode': 1317256, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.3662472, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'path': '/operations/prometheus/alertmanager.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 5065, 'inode': 1317242, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.3607938, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'path': '/operations/prometheus/mysql.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3792, 'inode': 1317255, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.3659542, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'path': '/operations/prometheus/mysql.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3792, 'inode': 1317255, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.3659542, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'path': '/operations/prometheus/alertmanager.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 5065, 'inode': 1317242, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.3607938, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'path': '/operations/prometheus/mysql.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3792, 'inode': 1317255, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.3659542, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'path': '/operations/prometheus/node.rec.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 2309, 'inode': 1317256, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.3662472, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item={'path': '/operations/prometheus/node.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 14018, 'inode': 1317257, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.366479, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'path': '/operations/prometheus/rabbitmq.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3539, 'inode': 1317271, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.3709881, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'path': '/operations/prometheus/node.rec.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 2309, 'inode': 1317256, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.3662472, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'path': '/operations/prometheus/mysql.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3792, 'inode': 1317255, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.3659542, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'path': '/operations/prometheus/rabbitmq.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3539, 'inode': 1317271, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.3709881, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'path': '/operations/prometheus/rabbitmq.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3539, 'inode': 1317271, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.3709881, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'path': '/operations/prometheus/rabbitmq.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3539, 'inode': 1317271, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.3709881, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'path': '/operations/prometheus/mysql.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3792, 'inode': 1317255, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.3659542, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'path': '/operations/prometheus/rabbitmq.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3539, 'inode': 1317271, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.3709881, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item={'path': '/operations/prometheus/hardware.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 5593, 'inode': 1317253, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.36546, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item={'path': '/operations/prometheus/elasticsearch.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 5987, 'inode': 1317249, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.3633637, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item={'path': '/operations/prometheus/prometheus.rec.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3, 'inode': 1317262, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.368623, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item={'path': '/operations/prometheus/alertmanager.rec.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3, 'inode': 1317241, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.3604016, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item={'path': '/operations/prometheus/redfish.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 334, 'inode': 1317272, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.371548, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item={'path': '/operations/prometheus/prometheus-extra.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 7408, 'inode': 1317260, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.3682377, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item={'path': '/operations/prometheus/ceph.rec.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3, 'inode': 1317245, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.361316, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item={'path': '/operations/prometheus/alertmanager.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 5065, 'inode': 1317242, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.3607938, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item={'path': '/operations/prometheus/node.rec.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 2309, 'inode': 1317256, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.3662472, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item={'path': '/operations/prometheus/mysql.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3792, 'inode': 1317255, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.3659542, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item={'path': '/operations/prometheus/rabbitmq.rules', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 3539, 'inode': 1317271, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.3709881, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False})\u001b[0m",
                                "",
                                "TASK [prometheus : Find prometheus common config overrides] ********************",
                                "Tuesday 24 February 2026  01:07:16 +0000 (0:00:35.749)       0:01:11.318 ****** ",
                                "\u001b[0;32mok: [testbed-manager -> localhost]\u001b[0m",
                                "",
                                "TASK [prometheus : Find prometheus host config overrides] **********************",
                                "Tuesday 24 February 2026  01:07:16 +0000 (0:00:00.865)       0:01:12.184 ****** ",
                                "\u001b[1;35m[WARNING]: Skipped\u001b[0m",
                                "\u001b[1;35m'/opt/configuration/environments/kolla/files/overlays/prometheus/testbed-\u001b[0m",
                                "\u001b[1;35mmanager/prometheus.yml.d' path due to this access issue:\u001b[0m",
                                "\u001b[1;35m'/opt/configuration/environments/kolla/files/overlays/prometheus/testbed-\u001b[0m",
                                "\u001b[1;35mmanager/prometheus.yml.d' is not a directory\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager -> localhost]\u001b[0m",
                                "\u001b[1;35m[WARNING]: Skipped\u001b[0m",
                                "\u001b[1;35m'/opt/configuration/environments/kolla/files/overlays/prometheus/testbed-\u001b[0m",
                                "\u001b[1;35mnode-0/prometheus.yml.d' path due to this access issue:\u001b[0m",
                                "\u001b[1;35m'/opt/configuration/environments/kolla/files/overlays/prometheus/testbed-\u001b[0m",
                                "\u001b[1;35mnode-0/prometheus.yml.d' is not a directory\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0 -> localhost]\u001b[0m",
                                "\u001b[1;35m[WARNING]: Skipped\u001b[0m",
                                "\u001b[1;35m'/opt/configuration/environments/kolla/files/overlays/prometheus/testbed-\u001b[0m",
                                "\u001b[1;35mnode-1/prometheus.yml.d' path due to this access issue:\u001b[0m",
                                "\u001b[1;35m'/opt/configuration/environments/kolla/files/overlays/prometheus/testbed-\u001b[0m",
                                "\u001b[1;35mnode-1/prometheus.yml.d' is not a directory\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1 -> localhost]\u001b[0m",
                                "\u001b[1;35m[WARNING]: Skipped\u001b[0m",
                                "\u001b[1;35m'/opt/configuration/environments/kolla/files/overlays/prometheus/testbed-\u001b[0m",
                                "\u001b[1;35mnode-2/prometheus.yml.d' path due to this access issue:\u001b[0m",
                                "\u001b[1;35m'/opt/configuration/environments/kolla/files/overlays/prometheus/testbed-\u001b[0m",
                                "\u001b[1;35mnode-2/prometheus.yml.d' is not a directory\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2 -> localhost]\u001b[0m",
                                "\u001b[1;35m[WARNING]: Skipped\u001b[0m",
                                "\u001b[1;35m'/opt/configuration/environments/kolla/files/overlays/prometheus/testbed-\u001b[0m",
                                "\u001b[1;35mnode-3/prometheus.yml.d' path due to this access issue:\u001b[0m",
                                "\u001b[1;35m'/opt/configuration/environments/kolla/files/overlays/prometheus/testbed-\u001b[0m",
                                "\u001b[1;35mnode-3/prometheus.yml.d' is not a directory\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3 -> localhost]\u001b[0m",
                                "\u001b[1;35m[WARNING]: Skipped\u001b[0m",
                                "\u001b[1;35m'/opt/configuration/environments/kolla/files/overlays/prometheus/testbed-\u001b[0m",
                                "\u001b[1;35mnode-5/prometheus.yml.d' path due to this access issue:\u001b[0m",
                                "\u001b[1;35m'/opt/configuration/environments/kolla/files/overlays/prometheus/testbed-\u001b[0m",
                                "\u001b[1;35mnode-5/prometheus.yml.d' is not a directory\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5 -> localhost]\u001b[0m",
                                "\u001b[1;35m[WARNING]: Skipped\u001b[0m",
                                "\u001b[1;35m'/opt/configuration/environments/kolla/files/overlays/prometheus/testbed-\u001b[0m",
                                "\u001b[1;35mnode-4/prometheus.yml.d' path due to this access issue:\u001b[0m",
                                "\u001b[1;35m'/opt/configuration/environments/kolla/files/overlays/prometheus/testbed-\u001b[0m",
                                "\u001b[1;35mnode-4/prometheus.yml.d' is not a directory\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4 -> localhost]\u001b[0m",
                                "",
                                "TASK [prometheus : Copying over prometheus config file] ************************",
                                "Tuesday 24 February 2026  01:07:18 +0000 (0:00:01.910)       0:01:14.094 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=/ansible/roles/prometheus/templates/prometheus.yml.j2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=/ansible/roles/prometheus/templates/prometheus.yml.j2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=/ansible/roles/prometheus/templates/prometheus.yml.j2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=/ansible/roles/prometheus/templates/prometheus.yml.j2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=/ansible/roles/prometheus/templates/prometheus.yml.j2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=/ansible/roles/prometheus/templates/prometheus.yml.j2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=/ansible/roles/prometheus/templates/prometheus.yml.j2)\u001b[0m",
                                "",
                                "TASK [prometheus : Copying over prometheus web config file] ********************",
                                "Tuesday 24 February 2026  01:07:37 +0000 (0:00:18.746)       0:01:32.840 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=/ansible/roles/prometheus/templates/prometheus-web.yml.j2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=/ansible/roles/prometheus/templates/prometheus-web.yml.j2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=/ansible/roles/prometheus/templates/prometheus-web.yml.j2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=/ansible/roles/prometheus/templates/prometheus-web.yml.j2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=/ansible/roles/prometheus/templates/prometheus-web.yml.j2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=/ansible/roles/prometheus/templates/prometheus-web.yml.j2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=/ansible/roles/prometheus/templates/prometheus-web.yml.j2)\u001b[0m",
                                "",
                                "TASK [prometheus : Copying over prometheus alertmanager config file] ***********",
                                "Tuesday 24 February 2026  01:07:42 +0000 (0:00:04.630)       0:01:37.471 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=/opt/configuration/environments/kolla/files/overlays/prometheus/prometheus-alertmanager.yml) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=/opt/configuration/environments/kolla/files/overlays/prometheus/prometheus-alertmanager.yml) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=/opt/configuration/environments/kolla/files/overlays/prometheus/prometheus-alertmanager.yml) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=/opt/configuration/environments/kolla/files/overlays/prometheus/prometheus-alertmanager.yml) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=/opt/configuration/environments/kolla/files/overlays/prometheus/prometheus-alertmanager.yml) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=/opt/configuration/environments/kolla/files/overlays/prometheus/prometheus-alertmanager.yml)\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=/opt/configuration/environments/kolla/files/overlays/prometheus/prometheus-alertmanager.yml) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [prometheus : Find custom Alertmanager alert notification templates] ******",
                                "Tuesday 24 February 2026  01:07:45 +0000 (0:00:03.081)       0:01:40.553 ****** ",
                                "\u001b[0;32mok: [testbed-manager -> localhost]\u001b[0m",
                                "",
                                "TASK [prometheus : Copying over custom Alertmanager alert notification templates] ***",
                                "Tuesday 24 February 2026  01:07:46 +0000 (0:00:01.367)       0:01:41.920 ****** ",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [prometheus : Copying over my.cnf for mysqld_exporter] ********************",
                                "Tuesday 24 February 2026  01:07:47 +0000 (0:00:01.191)       0:01:43.111 ****** ",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [prometheus : Copying cloud config file for openstack exporter] ***********",
                                "Tuesday 24 February 2026  01:07:51 +0000 (0:00:03.254)       0:01:46.366 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=/ansible/roles/prometheus/templates/clouds.yml.j2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=/ansible/roles/prometheus/templates/clouds.yml.j2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager] => (item=/ansible/roles/prometheus/templates/clouds.yml.j2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=/ansible/roles/prometheus/templates/clouds.yml.j2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=/ansible/roles/prometheus/templates/clouds.yml.j2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=/ansible/roles/prometheus/templates/clouds.yml.j2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=/ansible/roles/prometheus/templates/clouds.yml.j2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [prometheus : Copying config file for blackbox exporter] ******************",
                                "Tuesday 24 February 2026  01:07:53 +0000 (0:00:02.360)       0:01:48.726 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=/ansible/roles/prometheus/templates/prometheus-blackbox-exporter.yml.j2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=/ansible/roles/prometheus/templates/prometheus-blackbox-exporter.yml.j2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=/ansible/roles/prometheus/templates/prometheus-blackbox-exporter.yml.j2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=/ansible/roles/prometheus/templates/prometheus-blackbox-exporter.yml.j2)\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=/ansible/roles/prometheus/templates/prometheus-blackbox-exporter.yml.j2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=/ansible/roles/prometheus/templates/prometheus-blackbox-exporter.yml.j2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=/ansible/roles/prometheus/templates/prometheus-blackbox-exporter.yml.j2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [prometheus : Find extra prometheus server config files] ******************",
                                "Tuesday 24 February 2026  01:07:55 +0000 (0:00:02.105)       0:01:50.832 ****** ",
                                "\u001b[1;35m[WARNING]: Skipped\u001b[0m",
                                "\u001b[1;35m'/opt/configuration/environments/kolla/files/overlays/prometheus/extras/' path\u001b[0m",
                                "\u001b[1;35mdue to this access issue:\u001b[0m",
                                "\u001b[1;35m'/opt/configuration/environments/kolla/files/overlays/prometheus/extras/' is\u001b[0m",
                                "\u001b[1;35mnot a directory\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager -> localhost]\u001b[0m",
                                "",
                                "TASK [prometheus : Create subdirectories for extra config files] ***************",
                                "Tuesday 24 February 2026  01:07:57 +0000 (0:00:01.599)       0:01:52.432 ****** ",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [prometheus : Template extra prometheus server config files] **************",
                                "Tuesday 24 February 2026  01:07:59 +0000 (0:00:02.145)       0:01:54.577 ****** ",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [prometheus : Check prometheus containers] ********************************",
                                "Tuesday 24 February 2026  01:08:00 +0000 (0:00:00.925)       0:01:55.502 ****** ",
                                "\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'prometheus-server', 'value': {'container_name': 'prometheus_server', 'group': 'prometheus', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-v2-server:2024.2', 'volumes': ['/etc/kolla/prometheus-server/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'prometheus_v2:/var/lib/prometheus', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'prometheus_server': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9091', 'active_passive': True}, 'prometheus_server_external': {'enabled': False, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9091', 'listen_port': '9091', 'active_passive': True}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2024.2', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2024.2', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2024.2', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2024.2', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2024.2', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2024.2', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'prometheus-node-exporter', 'value': {'container_name': 'prometheus_node_exporter', 'group': 'prometheus-node-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2024.2', 'pid_mode': 'host', 'volumes': ['/etc/kolla/prometheus-node-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/host:ro,rslave'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2024.2', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'prometheus-mysqld-exporter', 'value': {'container_name': 'prometheus_mysqld_exporter', 'group': 'prometheus-mysqld-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-mysqld-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-mysqld-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'prometheus-mysqld-exporter', 'value': {'container_name': 'prometheus_mysqld_exporter', 'group': 'prometheus-mysqld-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-mysqld-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-mysqld-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'prometheus-mysqld-exporter', 'value': {'container_name': 'prometheus_mysqld_exporter', 'group': 'prometheus-mysqld-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-mysqld-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-mysqld-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2024.2', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2024.2', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2024.2', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'prometheus-libvirt-exporter', 'value': {'container_name': 'prometheus_libvirt_exporter', 'group': 'prometheus-libvirt-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-libvirt-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-libvirt-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run/libvirt:/run/libvirt:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'prometheus-alertmanager', 'value': {'container_name': 'prometheus_alertmanager', 'group': 'prometheus-alertmanager', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-alertmanager:2024.2', 'volumes': ['/etc/kolla/prometheus-alertmanager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', 'prometheus:/var/lib/prometheus'], 'dimensions': {}, 'haproxy': {'prometheus_alertmanager': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True}, 'prometheus_alertmanager_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9093', 'listen_port': '9093', 'auth_user': 'admin', 'auth_pass': 'BXo64rLqmF7bTbWLDOnNJlD0qJ4BSTWocNHVNKU2', 'active_passive': True}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'prometheus-memcached-exporter', 'value': {'container_name': 'prometheus_memcached_exporter', 'group': 'prometheus-memcached-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-memcached-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-memcached-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'prometheus-memcached-exporter', 'value': {'container_name': 'prometheus_memcached_exporter', 'group': 'prometheus-memcached-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-memcached-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-memcached-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'prometheus-libvirt-exporter', 'value': {'container_name': 'prometheus_libvirt_exporter', 'group': 'prometheus-libvirt-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-libvirt-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-libvirt-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run/libvirt:/run/libvirt:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'prometheus-memcached-exporter', 'value': {'container_name': 'prometheus_memcached_exporter', 'group': 'prometheus-memcached-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-memcached-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-memcached-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'prometheus-libvirt-exporter', 'value': {'container_name': 'prometheus_libvirt_exporter', 'group': 'prometheus-libvirt-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-libvirt-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-libvirt-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/run/libvirt:/run/libvirt:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager] => (item={'key': 'prometheus-blackbox-exporter', 'value': {'container_name': 'prometheus_blackbox_exporter', 'group': 'prometheus-blackbox-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-blackbox-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-blackbox-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2024.2', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2024.2', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'prometheus-cadvisor', 'value': {'container_name': 'prometheus_cadvisor', 'group': 'prometheus-cadvisor', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2024.2', 'volumes': ['/etc/kolla/prometheus-cadvisor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '/:/rootfs:ro', '/var/run:/var/run:rw', '/sys:/sys:ro', '/var/lib/docker/:/var/lib/docker:ro', '/dev/disk/:/dev/disk:ro'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'prometheus-elasticsearch-exporter', 'value': {'container_name': 'prometheus_elasticsearch_exporter', 'group': 'prometheus-elasticsearch-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-elasticsearch-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-elasticsearch-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'prometheus-elasticsearch-exporter', 'value': {'container_name': 'prometheus_elasticsearch_exporter', 'group': 'prometheus-elasticsearch-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-elasticsearch-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-elasticsearch-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'prometheus-elasticsearch-exporter', 'value': {'container_name': 'prometheus_elasticsearch_exporter', 'group': 'prometheus-elasticsearch-exporter', 'enabled': True, 'image': 'registry.osism.tech/kolla/prometheus-elasticsearch-exporter:2024.2', 'volumes': ['/etc/kolla/prometheus-elasticsearch-exporter/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}}})\u001b[0m",
                                "",
                                "TASK [prometheus : Creating prometheus database user and setting permissions] ***",
                                "Tuesday 24 February 2026  01:08:04 +0000 (0:00:04.587)       0:02:00.090 ****** ",
                                "\u001b[0;36mskipping: [testbed-manager] => (item=testbed-node-0) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [prometheus : Flush handlers] *********************************************",
                                "Tuesday 24 February 2026  01:08:06 +0000 (0:00:01.662)       0:02:01.753 ****** ",
                                "",
                                "TASK [prometheus : Flush handlers] *********************************************",
                                "Tuesday 24 February 2026  01:08:06 +0000 (0:00:00.087)       0:02:01.840 ****** ",
                                "",
                                "TASK [prometheus : Flush handlers] *********************************************",
                                "Tuesday 24 February 2026  01:08:06 +0000 (0:00:00.069)       0:02:01.910 ****** ",
                                "",
                                "TASK [prometheus : Flush handlers] *********************************************",
                                "Tuesday 24 February 2026  01:08:06 +0000 (0:00:00.071)       0:02:01.982 ****** ",
                                "",
                                "TASK [prometheus : Flush handlers] *********************************************",
                                "Tuesday 24 February 2026  01:08:07 +0000 (0:00:00.405)       0:02:02.387 ****** ",
                                "",
                                "TASK [prometheus : Flush handlers] *********************************************",
                                "Tuesday 24 February 2026  01:08:07 +0000 (0:00:00.079)       0:02:02.467 ****** ",
                                "",
                                "TASK [prometheus : Flush handlers] *********************************************",
                                "Tuesday 24 February 2026  01:08:07 +0000 (0:00:00.072)       0:02:02.539 ****** ",
                                "",
                                "RUNNING HANDLER [prometheus : Restart prometheus-server container] *************",
                                "Tuesday 24 February 2026  01:08:07 +0000 (0:00:00.111)       0:02:02.651 ****** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "RUNNING HANDLER [prometheus : Restart prometheus-node-exporter container] ******",
                                "Tuesday 24 February 2026  01:08:23 +0000 (0:00:15.973)       0:02:18.625 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "",
                                "RUNNING HANDLER [prometheus : Restart prometheus-mysqld-exporter container] ****",
                                "Tuesday 24 February 2026  01:08:38 +0000 (0:00:15.586)       0:02:34.211 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [prometheus : Restart prometheus-memcached-exporter container] ***",
                                "Tuesday 24 February 2026  01:08:50 +0000 (0:00:11.283)       0:02:45.494 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [prometheus : Restart prometheus-cadvisor container] ***********",
                                "Tuesday 24 February 2026  01:08:57 +0000 (0:00:07.008)       0:02:52.503 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "",
                                "RUNNING HANDLER [prometheus : Restart prometheus-alertmanager container] *******",
                                "Tuesday 24 February 2026  01:09:13 +0000 (0:00:16.517)       0:03:09.021 ****** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "RUNNING HANDLER [prometheus : Restart prometheus-elasticsearch-exporter container] ***",
                                "Tuesday 24 February 2026  01:09:20 +0000 (0:00:06.976)       0:03:15.997 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "RUNNING HANDLER [prometheus : Restart prometheus-blackbox-exporter container] ***",
                                "Tuesday 24 February 2026  01:09:30 +0000 (0:00:09.898)       0:03:25.895 ****** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "RUNNING HANDLER [prometheus : Restart prometheus-libvirt-exporter container] ***",
                                "Tuesday 24 February 2026  01:09:35 +0000 (0:00:05.162)       0:03:31.058 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;33mtestbed-manager\u001b[0m            : \u001b[0;32mok=23  \u001b[0m \u001b[0;33mchanged=14  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=8   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=15  \u001b[0m \u001b[0;33mchanged=10  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=11  \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=15  \u001b[0m \u001b[0;33mchanged=10  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=11  \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=15  \u001b[0m \u001b[0;33mchanged=10  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=11  \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-3\u001b[0m             : \u001b[0;32mok=12  \u001b[0m \u001b[0;33mchanged=7   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=12  \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-4\u001b[0m             : \u001b[0;32mok=12  \u001b[0m \u001b[0;33mchanged=7   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=12  \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-5\u001b[0m             : \u001b[0;32mok=12  \u001b[0m \u001b[0;33mchanged=7   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=12  \u001b[0m rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Tuesday 24 February 2026  01:09:46 +0000 (0:00:10.244)       0:03:41.303 ****** ",
                                "=============================================================================== ",
                                "prometheus : Copying over custom prometheus alert rules files ---------- 35.75s",
                                "prometheus : Copying over prometheus config file ----------------------- 18.75s",
                                "prometheus : Restart prometheus-cadvisor container --------------------- 16.52s",
                                "prometheus : Restart prometheus-server container ----------------------- 15.97s",
                                "prometheus : Restart prometheus-node-exporter container ---------------- 15.59s",
                                "prometheus : Restart prometheus-mysqld-exporter container -------------- 11.28s",
                                "prometheus : Restart prometheus-libvirt-exporter container ------------- 10.24s",
                                "prometheus : Restart prometheus-elasticsearch-exporter container -------- 9.90s",
                                "prometheus : Copying over config.json files ----------------------------- 9.71s",
                                "service-cert-copy : prometheus | Copying over extra CA certificates ----- 7.06s",
                                "prometheus : Restart prometheus-memcached-exporter container ------------ 7.01s",
                                "prometheus : Restart prometheus-alertmanager container ------------------ 6.98s",
                                "prometheus : Restart prometheus-blackbox-exporter container ------------- 5.16s",
                                "prometheus : Copying over prometheus web config file -------------------- 4.63s",
                                "service-cert-copy : prometheus | Copying over backend internal TLS key --- 4.62s",
                                "prometheus : Check prometheus containers -------------------------------- 4.59s",
                                "service-cert-copy : prometheus | Copying over backend internal TLS certificate --- 3.62s",
                                "prometheus : Ensuring config directories exist -------------------------- 3.32s",
                                "prometheus : Copying over my.cnf for mysqld_exporter -------------------- 3.25s",
                                "prometheus : Copying over prometheus alertmanager config file ----------- 3.08s",
                                "\u001b[32m2026-02-24 01:09:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4baf3a0c-387b-4215-b490-5edd972623e9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:09:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 01:09:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fba6bf07-65a1-432e-9a3d-72209557a21d is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:09:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask eddb1a5f-448f-4517-b064-c30da006d691 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:09:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask de6aa5ad-1dfd-4a58-bf00-4755d79ef102 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:09:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4baf3a0c-387b-4215-b490-5edd972623e9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:09:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 01:09:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fba6bf07-65a1-432e-9a3d-72209557a21d is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:09:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask eddb1a5f-448f-4517-b064-c30da006d691 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:09:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask de6aa5ad-1dfd-4a58-bf00-4755d79ef102 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:09:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4baf3a0c-387b-4215-b490-5edd972623e9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:09:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 01:09:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fba6bf07-65a1-432e-9a3d-72209557a21d is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:09:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask eddb1a5f-448f-4517-b064-c30da006d691 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:09:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask de6aa5ad-1dfd-4a58-bf00-4755d79ef102 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:09:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4baf3a0c-387b-4215-b490-5edd972623e9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:09:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 01:10:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fba6bf07-65a1-432e-9a3d-72209557a21d is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:10:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask eddb1a5f-448f-4517-b064-c30da006d691 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:10:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask de6aa5ad-1dfd-4a58-bf00-4755d79ef102 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:10:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4baf3a0c-387b-4215-b490-5edd972623e9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:10:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 01:10:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fba6bf07-65a1-432e-9a3d-72209557a21d is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:10:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask eddb1a5f-448f-4517-b064-c30da006d691 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:10:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask de6aa5ad-1dfd-4a58-bf00-4755d79ef102 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:10:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4baf3a0c-387b-4215-b490-5edd972623e9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:10:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 01:10:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fba6bf07-65a1-432e-9a3d-72209557a21d is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:10:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask eddb1a5f-448f-4517-b064-c30da006d691 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:10:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask de6aa5ad-1dfd-4a58-bf00-4755d79ef102 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:10:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4baf3a0c-387b-4215-b490-5edd972623e9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:10:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 01:10:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fba6bf07-65a1-432e-9a3d-72209557a21d is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:10:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask eddb1a5f-448f-4517-b064-c30da006d691 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:10:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask de6aa5ad-1dfd-4a58-bf00-4755d79ef102 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:10:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4baf3a0c-387b-4215-b490-5edd972623e9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:10:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 01:10:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fba6bf07-65a1-432e-9a3d-72209557a21d is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:10:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask eddb1a5f-448f-4517-b064-c30da006d691 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:10:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask de6aa5ad-1dfd-4a58-bf00-4755d79ef102 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:10:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4baf3a0c-387b-4215-b490-5edd972623e9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:10:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 01:10:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fba6bf07-65a1-432e-9a3d-72209557a21d is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:10:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask eddb1a5f-448f-4517-b064-c30da006d691 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:10:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask de6aa5ad-1dfd-4a58-bf00-4755d79ef102 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:10:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4baf3a0c-387b-4215-b490-5edd972623e9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:10:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 01:10:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fba6bf07-65a1-432e-9a3d-72209557a21d is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:10:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask eddb1a5f-448f-4517-b064-c30da006d691 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:10:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask de6aa5ad-1dfd-4a58-bf00-4755d79ef102 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:10:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4baf3a0c-387b-4215-b490-5edd972623e9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:10:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 01:10:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fba6bf07-65a1-432e-9a3d-72209557a21d is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:10:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask eddb1a5f-448f-4517-b064-c30da006d691 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:10:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask de6aa5ad-1dfd-4a58-bf00-4755d79ef102 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:10:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4baf3a0c-387b-4215-b490-5edd972623e9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:10:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 01:10:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fba6bf07-65a1-432e-9a3d-72209557a21d is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:10:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask eddb1a5f-448f-4517-b064-c30da006d691 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:10:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask de6aa5ad-1dfd-4a58-bf00-4755d79ef102 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:10:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4baf3a0c-387b-4215-b490-5edd972623e9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:10:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 01:10:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fba6bf07-65a1-432e-9a3d-72209557a21d is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:10:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask eddb1a5f-448f-4517-b064-c30da006d691 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:10:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask de6aa5ad-1dfd-4a58-bf00-4755d79ef102 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:10:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4baf3a0c-387b-4215-b490-5edd972623e9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:10:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 01:10:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask fba6bf07-65a1-432e-9a3d-72209557a21d is in state SUCCESS\u001b[0m",
                                "",
                                "",
                                "PLAY [Group hosts based on configuration] **************************************",
                                "",
                                "TASK [Group hosts based on Kolla action] ***************************************",
                                "Tuesday 24 February 2026  01:07:05 +0000 (0:00:00.529)       0:00:00.529 ****** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [Group hosts based on enabled services] ***********************************",
                                "Tuesday 24 February 2026  01:07:06 +0000 (0:00:00.606)       0:00:01.135 ****** ",
                                "\u001b[0;32mok: [testbed-node-0] => (item=enable_glance_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item=enable_glance_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item=enable_glance_True)\u001b[0m",
                                "",
                                "PLAY [Apply role glance] *******************************************************",
                                "",
                                "TASK [glance : include_tasks] **************************************************",
                                "Tuesday 24 February 2026  01:07:07 +0000 (0:00:01.432)       0:00:02.568 ****** ",
                                "\u001b[0;36mincluded: /ansible/roles/glance/tasks/deploy.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [service-ks-register : glance | Creating services] ************************",
                                "Tuesday 24 February 2026  01:07:09 +0000 (0:00:01.471)       0:00:04.040 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=glance (image))\u001b[0m",
                                "",
                                "TASK [service-ks-register : glance | Creating endpoints] ***********************",
                                "Tuesday 24 February 2026  01:07:13 +0000 (0:00:04.321)       0:00:08.362 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=glance -> https://api-int.testbed.osism.xyz:9292 -> internal)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=glance -> https://api.testbed.osism.xyz:9292 -> public)\u001b[0m",
                                "",
                                "TASK [service-ks-register : glance | Creating projects] ************************",
                                "Tuesday 24 February 2026  01:07:20 +0000 (0:00:06.822)       0:00:15.184 ****** ",
                                "\u001b[0;32mok: [testbed-node-0] => (item=service)\u001b[0m",
                                "",
                                "TASK [service-ks-register : glance | Creating users] ***************************",
                                "Tuesday 24 February 2026  01:07:23 +0000 (0:00:03.335)       0:00:18.520 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=glance -> service)\u001b[0m",
                                "\u001b[1;35m[WARNING]: Module did not set no_log for update_password\u001b[0m",
                                "",
                                "TASK [service-ks-register : glance | Creating roles] ***************************",
                                "Tuesday 24 February 2026  01:07:27 +0000 (0:00:04.402)       0:00:22.922 ****** ",
                                "\u001b[0;32mok: [testbed-node-0] => (item=admin)\u001b[0m",
                                "",
                                "TASK [service-ks-register : glance | Granting user roles] **********************",
                                "Tuesday 24 February 2026  01:07:31 +0000 (0:00:03.688)       0:00:26.611 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=glance -> service -> admin)\u001b[0m",
                                "",
                                "TASK [glance : Ensuring config directories exist] ******************************",
                                "Tuesday 24 February 2026  01:07:36 +0000 (0:00:04.453)       0:00:31.064 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'glance-api', 'value': {'container_name': 'glance_api', 'group': 'glance-api', 'host_in_groups': True, 'enabled': True, 'image': 'registry.osism.tech/kolla/glance-api:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.11,192.168.16.9'}, 'privileged': True, 'volumes': ['/etc/kolla/glance-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'glance:/var/lib/glance/', '', 'kolla_logs:/var/log/kolla/', '', '', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9292'], 'timeout': '30'}, 'haproxy': {'glance_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}, 'glance_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'glance-api', 'value': {'container_name': 'glance_api', 'group': 'glance-api', 'host_in_groups': True, 'enabled': True, 'image': 'registry.osism.tech/kolla/glance-api:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.10,192.168.16.9'}, 'privileged': True, 'volumes': ['/etc/kolla/glance-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'glance:/var/lib/glance/', '', 'kolla_logs:/var/log/kolla/', '', '', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9292'], 'timeout': '30'}, 'haproxy': {'glance_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}, 'glance_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'glance-api', 'value': {'container_name': 'glance_api', 'group': 'glance-api', 'host_in_groups': True, 'enabled': True, 'image': 'registry.osism.tech/kolla/glance-api:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.12,192.168.16.9'}, 'privileged': True, 'volumes': ['/etc/kolla/glance-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'glance:/var/lib/glance/', '', 'kolla_logs:/var/log/kolla/', '', '', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9292'], 'timeout': '30'}, 'haproxy': {'glance_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}, 'glance_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}}})\u001b[0m",
                                "",
                                "TASK [glance : include_tasks] **************************************************",
                                "Tuesday 24 February 2026  01:07:42 +0000 (0:00:05.994)       0:00:37.059 ****** ",
                                "\u001b[0;36mincluded: /ansible/roles/glance/tasks/external_ceph.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [glance : Ensuring glance service ceph config subdir exists] **************",
                                "Tuesday 24 February 2026  01:07:43 +0000 (0:00:01.396)       0:00:38.456 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [glance : Copy over multiple ceph configs for Glance] *********************",
                                "Tuesday 24 February 2026  01:07:49 +0000 (0:00:05.557)       0:00:44.014 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'name': 'rbd', 'type': 'rbd', 'cluster': 'ceph', 'enabled': True})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'name': 'rbd', 'type': 'rbd', 'cluster': 'ceph', 'enabled': True})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'name': 'rbd', 'type': 'rbd', 'cluster': 'ceph', 'enabled': True})\u001b[0m",
                                "",
                                "TASK [glance : Copy over ceph Glance keyrings] *********************************",
                                "Tuesday 24 February 2026  01:07:51 +0000 (0:00:02.304)       0:00:46.319 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'name': 'rbd', 'type': 'rbd', 'cluster': 'ceph', 'enabled': True})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'name': 'rbd', 'type': 'rbd', 'cluster': 'ceph', 'enabled': True})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'name': 'rbd', 'type': 'rbd', 'cluster': 'ceph', 'enabled': True})\u001b[0m",
                                "",
                                "TASK [glance : Ensuring config directory has correct owner and permission] *****",
                                "Tuesday 24 February 2026  01:07:53 +0000 (0:00:01.724)       0:00:48.043 ****** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [glance : Check if policies shall be overwritten] *************************",
                                "Tuesday 24 February 2026  01:07:54 +0000 (0:00:01.343)       0:00:49.387 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [glance : Set glance policy file] *****************************************",
                                "Tuesday 24 February 2026  01:07:54 +0000 (0:00:00.366)       0:00:49.754 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [glance : include_tasks] **************************************************",
                                "Tuesday 24 February 2026  01:07:55 +0000 (0:00:00.616)       0:00:50.371 ****** ",
                                "\u001b[0;36mincluded: /ansible/roles/glance/tasks/copy-certs.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [service-cert-copy : glance | Copying over extra CA certificates] *********",
                                "Tuesday 24 February 2026  01:07:55 +0000 (0:00:00.614)       0:00:50.985 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'glance-api', 'value': {'container_name': 'glance_api', 'group': 'glance-api', 'host_in_groups': True, 'enabled': True, 'image': 'registry.osism.tech/kolla/glance-api:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.10,192.168.16.9'}, 'privileged': True, 'volumes': ['/etc/kolla/glance-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'glance:/var/lib/glance/', '', 'kolla_logs:/var/log/kolla/', '', '', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9292'], 'timeout': '30'}, 'haproxy': {'glance_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}, 'glance_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'glance-api', 'value': {'container_name': 'glance_api', 'group': 'glance-api', 'host_in_groups': True, 'enabled': True, 'image': 'registry.osism.tech/kolla/glance-api:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.11,192.168.16.9'}, 'privileged': True, 'volumes': ['/etc/kolla/glance-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'glance:/var/lib/glance/', '', 'kolla_logs:/var/log/kolla/', '', '', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9292'], 'timeout': '30'}, 'haproxy': {'glance_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}, 'glance_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'glance-api', 'value': {'container_name': 'glance_api', 'group': 'glance-api', 'host_in_groups': True, 'enabled': True, 'image': 'registry.osism.tech/kolla/glance-api:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.12,192.168.16.9'}, 'privileged': True, 'volumes': ['/etc/kolla/glance-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'glance:/var/lib/glance/', '', 'kolla_logs:/var/log/kolla/', '', '', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9292'], 'timeout': '30'}, 'haproxy': {'glance_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}, 'glance_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}}})\u001b[0m",
                                "",
                                "TASK [service-cert-copy : glance | Copying over backend internal TLS certificate] ***",
                                "Tuesday 24 February 2026  01:08:01 +0000 (0:00:05.846)       0:00:56.831 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'glance-api', 'value': {'container_name': 'glance_api', 'group': 'glance-api', 'host_in_groups': True, 'enabled': True, 'image': 'registry.osism.tech/kolla/glance-api:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.10,192.168.16.9'}, 'privileged': True, 'volumes': ['/etc/kolla/glance-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'glance:/var/lib/glance/', '', 'kolla_logs:/var/log/kolla/', '', '', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9292'], 'timeout': '30'}, 'haproxy': {'glance_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}, 'glance_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'glance-api', 'value': {'container_name': 'glance_api', 'group': 'glance-api', 'host_in_groups': True, 'enabled': True, 'image': 'registry.osism.tech/kolla/glance-api:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.11,192.168.16.9'}, 'privileged': True, 'volumes': ['/etc/kolla/glance-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'glance:/var/lib/glance/', '', 'kolla_logs:/var/log/kolla/', '', '', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9292'], 'timeout': '30'}, 'haproxy': {'glance_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}, 'glance_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'glance-api', 'value': {'container_name': 'glance_api', 'group': 'glance-api', 'host_in_groups': True, 'enabled': True, 'image': 'registry.osism.tech/kolla/glance-api:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.12,192.168.16.9'}, 'privileged': True, 'volumes': ['/etc/kolla/glance-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'glance:/var/lib/glance/', '', 'kolla_logs:/var/log/kolla/', '', '', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9292'], 'timeout': '30'}, 'haproxy': {'glance_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}, 'glance_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [service-cert-copy : glance | Copying over backend internal TLS key] ******",
                                "Tuesday 24 February 2026  01:08:04 +0000 (0:00:03.120)       0:00:59.952 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'glance-api', 'value': {'container_name': 'glance_api', 'group': 'glance-api', 'host_in_groups': True, 'enabled': True, 'image': 'registry.osism.tech/kolla/glance-api:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.10,192.168.16.9'}, 'privileged': True, 'volumes': ['/etc/kolla/glance-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'glance:/var/lib/glance/', '', 'kolla_logs:/var/log/kolla/', '', '', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9292'], 'timeout': '30'}, 'haproxy': {'glance_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}, 'glance_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'glance-api', 'value': {'container_name': 'glance_api', 'group': 'glance-api', 'host_in_groups': True, 'enabled': True, 'image': 'registry.osism.tech/kolla/glance-api:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.11,192.168.16.9'}, 'privileged': True, 'volumes': ['/etc/kolla/glance-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'glance:/var/lib/glance/', '', 'kolla_logs:/var/log/kolla/', '', '', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9292'], 'timeout': '30'}, 'haproxy': {'glance_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}, 'glance_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'glance-api', 'value': {'container_name': 'glance_api', 'group': 'glance-api', 'host_in_groups': True, 'enabled': True, 'image': 'registry.osism.tech/kolla/glance-api:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.12,192.168.16.9'}, 'privileged': True, 'volumes': ['/etc/kolla/glance-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'glance:/var/lib/glance/', '', 'kolla_logs:/var/log/kolla/', '', '', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9292'], 'timeout': '30'}, 'haproxy': {'glance_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}, 'glance_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [glance : Creating TLS backend PEM File] **********************************",
                                "Tuesday 24 February 2026  01:08:08 +0000 (0:00:03.603)       0:01:03.556 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [glance : Copying over config.json files for services] ********************",
                                "Tuesday 24 February 2026  01:08:13 +0000 (0:00:05.412)       0:01:08.968 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'glance-api', 'value': {'container_name': 'glance_api', 'group': 'glance-api', 'host_in_groups': True, 'enabled': True, 'image': 'registry.osism.tech/kolla/glance-api:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.11,192.168.16.9'}, 'privileged': True, 'volumes': ['/etc/kolla/glance-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'glance:/var/lib/glance/', '', 'kolla_logs:/var/log/kolla/', '', '', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9292'], 'timeout': '30'}, 'haproxy': {'glance_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}, 'glance_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'glance-api', 'value': {'container_name': 'glance_api', 'group': 'glance-api', 'host_in_groups': True, 'enabled': True, 'image': 'registry.osism.tech/kolla/glance-api:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.10,192.168.16.9'}, 'privileged': True, 'volumes': ['/etc/kolla/glance-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'glance:/var/lib/glance/', '', 'kolla_logs:/var/log/kolla/', '', '', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9292'], 'timeout': '30'}, 'haproxy': {'glance_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}, 'glance_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'glance-api', 'value': {'container_name': 'glance_api', 'group': 'glance-api', 'host_in_groups': True, 'enabled': True, 'image': 'registry.osism.tech/kolla/glance-api:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.12,192.168.16.9'}, 'privileged': True, 'volumes': ['/etc/kolla/glance-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'glance:/var/lib/glance/', '', 'kolla_logs:/var/log/kolla/', '', '', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9292'], 'timeout': '30'}, 'haproxy': {'glance_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}, 'glance_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}}})\u001b[0m",
                                "",
                                "TASK [glance : Copying over glance-api.conf] ***********************************",
                                "Tuesday 24 February 2026  01:08:19 +0000 (0:00:05.947)       0:01:14.916 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [glance : Copying over glance-cache.conf for glance_api] ******************",
                                "Tuesday 24 February 2026  01:08:31 +0000 (0:00:11.398)       0:01:26.314 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [glance : Copying over glance-swift.conf for glance_api] ******************",
                                "Tuesday 24 February 2026  01:08:35 +0000 (0:00:04.409)       0:01:30.724 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [glance : Copying over glance-image-import.conf] **************************",
                                "Tuesday 24 February 2026  01:08:40 +0000 (0:00:04.911)       0:01:35.635 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [glance : Copying over property-protections-rules.conf] *******************",
                                "Tuesday 24 February 2026  01:08:45 +0000 (0:00:04.503)       0:01:40.139 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [glance : Copying over existing policy file] ******************************",
                                "Tuesday 24 February 2026  01:08:49 +0000 (0:00:04.070)       0:01:44.209 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [glance : Copying over glance-haproxy-tls.cfg] ****************************",
                                "Tuesday 24 February 2026  01:08:49 +0000 (0:00:00.425)       0:01:44.634 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=/ansible/roles/glance/templates/glance-tls-proxy.cfg.j2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=/ansible/roles/glance/templates/glance-tls-proxy.cfg.j2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=/ansible/roles/glance/templates/glance-tls-proxy.cfg.j2) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [glance : Generating 'hostnqn' file for glance_api] ***********************",
                                "Tuesday 24 February 2026  01:08:55 +0000 (0:00:05.794)       0:01:50.429 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [glance : Check glance containers] ****************************************",
                                "Tuesday 24 February 2026  01:09:02 +0000 (0:00:07.563)       0:01:57.992 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'glance-api', 'value': {'container_name': 'glance_api', 'group': 'glance-api', 'host_in_groups': True, 'enabled': True, 'image': 'registry.osism.tech/kolla/glance-api:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.10,192.168.16.9'}, 'privileged': True, 'volumes': ['/etc/kolla/glance-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'glance:/var/lib/glance/', '', 'kolla_logs:/var/log/kolla/', '', '', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9292'], 'timeout': '30'}, 'haproxy': {'glance_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}, 'glance_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'glance-api', 'value': {'container_name': 'glance_api', 'group': 'glance-api', 'host_in_groups': True, 'enabled': True, 'image': 'registry.osism.tech/kolla/glance-api:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.11,192.168.16.9'}, 'privileged': True, 'volumes': ['/etc/kolla/glance-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'glance:/var/lib/glance/', '', 'kolla_logs:/var/log/kolla/', '', '', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9292'], 'timeout': '30'}, 'haproxy': {'glance_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}, 'glance_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'glance-api', 'value': {'container_name': 'glance_api', 'group': 'glance-api', 'host_in_groups': True, 'enabled': True, 'image': 'registry.osism.tech/kolla/glance-api:2024.2', 'environment': {'http_proxy': '', 'https_proxy': '', 'no_proxy': 'localhost,127.0.0.1,192.168.16.12,192.168.16.9'}, 'privileged': True, 'volumes': ['/etc/kolla/glance-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'glance:/var/lib/glance/', '', 'kolla_logs:/var/log/kolla/', '', '', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9292'], 'timeout': '30'}, 'haproxy': {'glance_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}, 'glance_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9292', 'frontend_http_extra': ['timeout client 6h'], 'backend_http_extra': ['timeout server 6h'], 'custom_member_list': ['server testbed-node-0 192.168.16.10:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-1 192.168.16.11:9292 check inter 2000 rise 2 fall 5', 'server testbed-node-2 192.168.16.12:9292 check inter 2000 rise 2 fall 5', '']}}}})\u001b[0m",
                                "",
                                "TASK [glance : include_tasks] **************************************************",
                                "Tuesday 24 February 2026  01:09:09 +0000 (0:00:06.758)       0:02:04.751 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [glance : Creating Glance database] ***************************************",
                                "Tuesday 24 February 2026  01:09:10 +0000 (0:00:00.327)       0:02:05.079 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [glance : Creating Glance database user and setting permissions] **********",
                                "Tuesday 24 February 2026  01:09:12 +0000 (0:00:02.019)       0:02:07.099 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [glance : Enable log_bin_trust_function_creators function] ****************",
                                "Tuesday 24 February 2026  01:09:14 +0000 (0:00:02.366)       0:02:09.465 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [glance : Running Glance bootstrap container] *****************************",
                                "Tuesday 24 February 2026  01:09:16 +0000 (0:00:02.433)       0:02:11.899 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [glance : Disable log_bin_trust_function_creators function] ***************",
                                "Tuesday 24 February 2026  01:09:51 +0000 (0:00:34.287)       0:02:46.187 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [glance : Flush handlers] *************************************************",
                                "Tuesday 24 February 2026  01:09:53 +0000 (0:00:02.220)       0:02:48.407 ****** ",
                                "",
                                "TASK [glance : Flush handlers] *************************************************",
                                "Tuesday 24 February 2026  01:09:54 +0000 (0:00:00.689)       0:02:49.096 ****** ",
                                "",
                                "TASK [glance : Flush handlers] *************************************************",
                                "Tuesday 24 February 2026  01:09:54 +0000 (0:00:00.074)       0:02:49.171 ****** ",
                                "",
                                "RUNNING HANDLER [glance : Restart glance-api container] ************************",
                                "Tuesday 24 February 2026  01:09:54 +0000 (0:00:00.086)       0:02:49.257 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=27  \u001b[0m \u001b[0;33mchanged=19  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=12  \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=16  \u001b[0m \u001b[0;33mchanged=10  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=11  \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=16  \u001b[0m \u001b[0;33mchanged=10  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=11  \u001b[0m rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Tuesday 24 February 2026  01:10:28 +0000 (0:00:33.819)       0:03:23.077 ****** ",
                                "=============================================================================== ",
                                "glance : Running Glance bootstrap container ---------------------------- 34.29s",
                                "glance : Restart glance-api container ---------------------------------- 33.82s",
                                "glance : Copying over glance-api.conf ---------------------------------- 11.40s",
                                "glance : Generating 'hostnqn' file for glance_api ----------------------- 7.56s",
                                "service-ks-register : glance | Creating endpoints ----------------------- 6.82s",
                                "glance : Check glance containers ---------------------------------------- 6.76s",
                                "glance : Ensuring config directories exist ------------------------------ 5.99s",
                                "glance : Copying over config.json files for services -------------------- 5.95s",
                                "service-cert-copy : glance | Copying over extra CA certificates --------- 5.85s",
                                "glance : Copying over glance-haproxy-tls.cfg ---------------------------- 5.79s",
                                "glance : Ensuring glance service ceph config subdir exists -------------- 5.56s",
                                "glance : Creating TLS backend PEM File ---------------------------------- 5.41s",
                                "glance : Copying over glance-swift.conf for glance_api ------------------ 4.91s",
                                "glance : Copying over glance-image-import.conf -------------------------- 4.50s",
                                "service-ks-register : glance | Granting user roles ---------------------- 4.45s",
                                "glance : Copying over glance-cache.conf for glance_api ------------------ 4.41s",
                                "service-ks-register : glance | Creating users --------------------------- 4.40s",
                                "service-ks-register : glance | Creating services ------------------------ 4.32s",
                                "glance : Copying over property-protections-rules.conf ------------------- 4.07s",
                                "service-ks-register : glance | Creating roles --------------------------- 3.69s",
                                "\u001b[32m2026-02-24 01:10:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask eddb1a5f-448f-4517-b064-c30da006d691 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:10:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask de6aa5ad-1dfd-4a58-bf00-4755d79ef102 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:10:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 918da279-e500-4c4b-bdd7-1bd26f3de1b1 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:10:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4baf3a0c-387b-4215-b490-5edd972623e9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:10:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 01:10:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask eddb1a5f-448f-4517-b064-c30da006d691 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:10:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask de6aa5ad-1dfd-4a58-bf00-4755d79ef102 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:10:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 918da279-e500-4c4b-bdd7-1bd26f3de1b1 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:10:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4baf3a0c-387b-4215-b490-5edd972623e9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:10:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 01:10:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask eddb1a5f-448f-4517-b064-c30da006d691 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:10:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask de6aa5ad-1dfd-4a58-bf00-4755d79ef102 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:10:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 918da279-e500-4c4b-bdd7-1bd26f3de1b1 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:10:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4baf3a0c-387b-4215-b490-5edd972623e9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:10:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 01:10:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask eddb1a5f-448f-4517-b064-c30da006d691 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:10:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask de6aa5ad-1dfd-4a58-bf00-4755d79ef102 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:10:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 918da279-e500-4c4b-bdd7-1bd26f3de1b1 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:10:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4baf3a0c-387b-4215-b490-5edd972623e9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:10:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 01:10:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask eddb1a5f-448f-4517-b064-c30da006d691 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:10:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask de6aa5ad-1dfd-4a58-bf00-4755d79ef102 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:10:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 918da279-e500-4c4b-bdd7-1bd26f3de1b1 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:10:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4baf3a0c-387b-4215-b490-5edd972623e9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:10:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 01:10:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask eddb1a5f-448f-4517-b064-c30da006d691 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:10:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask de6aa5ad-1dfd-4a58-bf00-4755d79ef102 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:10:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 918da279-e500-4c4b-bdd7-1bd26f3de1b1 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:10:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4baf3a0c-387b-4215-b490-5edd972623e9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:10:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 01:10:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask eddb1a5f-448f-4517-b064-c30da006d691 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:10:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask de6aa5ad-1dfd-4a58-bf00-4755d79ef102 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:10:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 918da279-e500-4c4b-bdd7-1bd26f3de1b1 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:10:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4baf3a0c-387b-4215-b490-5edd972623e9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:10:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 01:10:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask eddb1a5f-448f-4517-b064-c30da006d691 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:10:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask de6aa5ad-1dfd-4a58-bf00-4755d79ef102 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:10:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 918da279-e500-4c4b-bdd7-1bd26f3de1b1 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:10:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4baf3a0c-387b-4215-b490-5edd972623e9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:10:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 01:10:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask eddb1a5f-448f-4517-b064-c30da006d691 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:10:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask de6aa5ad-1dfd-4a58-bf00-4755d79ef102 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:10:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 918da279-e500-4c4b-bdd7-1bd26f3de1b1 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:10:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4baf3a0c-387b-4215-b490-5edd972623e9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:10:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 01:10:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask eddb1a5f-448f-4517-b064-c30da006d691 is in state SUCCESS\u001b[0m",
                                "",
                                "",
                                "PLAY [Group hosts based on configuration] **************************************",
                                "",
                                "TASK [Group hosts based on Kolla action] ***************************************",
                                "Tuesday 24 February 2026  01:07:44 +0000 (0:00:00.350)       0:00:00.350 ****** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [Group hosts based on enabled services] ***********************************",
                                "Tuesday 24 February 2026  01:07:44 +0000 (0:00:00.452)       0:00:00.802 ****** ",
                                "\u001b[0;32mok: [testbed-node-0] => (item=enable_cinder_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item=enable_cinder_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item=enable_cinder_True)\u001b[0m",
                                "",
                                "PLAY [Apply role cinder] *******************************************************",
                                "",
                                "TASK [cinder : include_tasks] **************************************************",
                                "Tuesday 24 February 2026  01:07:46 +0000 (0:00:01.080)       0:00:01.882 ****** ",
                                "\u001b[0;36mincluded: /ansible/roles/cinder/tasks/deploy.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [service-ks-register : cinder | Creating services] ************************",
                                "Tuesday 24 February 2026  01:07:47 +0000 (0:00:01.080)       0:00:02.964 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=cinderv3 (volumev3))\u001b[0m",
                                "",
                                "TASK [service-ks-register : cinder | Creating endpoints] ***********************",
                                "Tuesday 24 February 2026  01:07:51 +0000 (0:00:03.950)       0:00:06.914 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=cinderv3 -> https://api-int.testbed.osism.xyz:8776/v3/%(tenant_id)s -> internal)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=cinderv3 -> https://api.testbed.osism.xyz:8776/v3/%(tenant_id)s -> public)\u001b[0m",
                                "",
                                "TASK [service-ks-register : cinder | Creating projects] ************************",
                                "Tuesday 24 February 2026  01:07:58 +0000 (0:00:07.411)       0:00:14.325 ****** ",
                                "\u001b[0;32mok: [testbed-node-0] => (item=service)\u001b[0m",
                                "",
                                "TASK [service-ks-register : cinder | Creating users] ***************************",
                                "Tuesday 24 February 2026  01:08:02 +0000 (0:00:03.961)       0:00:18.287 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=cinder -> service)\u001b[0m",
                                "\u001b[1;35m[WARNING]: Module did not set no_log for update_password\u001b[0m",
                                "",
                                "TASK [service-ks-register : cinder | Creating roles] ***************************",
                                "Tuesday 24 February 2026  01:08:06 +0000 (0:00:04.491)       0:00:22.778 ****** ",
                                "\u001b[0;32mok: [testbed-node-0] => (item=admin)\u001b[0m",
                                "",
                                "TASK [service-ks-register : cinder | Granting user roles] **********************",
                                "Tuesday 24 February 2026  01:08:11 +0000 (0:00:04.385)       0:00:27.164 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=cinder -> service -> admin)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=cinder -> service -> service)\u001b[0m",
                                "",
                                "TASK [cinder : Ensuring config directories exist] ******************************",
                                "Tuesday 24 February 2026  01:08:19 +0000 (0:00:08.019)       0:00:35.183 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-api:2024.2', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8776'], 'timeout': '30'}, 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-api:2024.2', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8776'], 'timeout': '30'}, 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-api:2024.2', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8776'], 'timeout': '30'}, 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-scheduler:2024.2', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-scheduler:2024.2', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-scheduler:2024.2', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-volume:2024.2', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-volume:2024.2', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-volume:2024.2', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-backup:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-backup:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-backup:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}})\u001b[0m",
                                "",
                                "TASK [cinder : include_tasks] **************************************************",
                                "Tuesday 24 February 2026  01:08:22 +0000 (0:00:02.803)       0:00:37.986 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [cinder : include_tasks] **************************************************",
                                "Tuesday 24 February 2026  01:08:22 +0000 (0:00:00.580)       0:00:38.567 ****** ",
                                "\u001b[0;36mincluded: /ansible/roles/cinder/tasks/external_ceph.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [cinder : Ensuring cinder service ceph config subdirs exists] *************",
                                "Tuesday 24 February 2026  01:08:23 +0000 (0:00:00.923)       0:00:39.491 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=cinder-volume)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=cinder-volume)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=cinder-backup)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=cinder-volume)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=cinder-backup)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=cinder-backup)\u001b[0m",
                                "",
                                "TASK [cinder : Copying over multiple ceph.conf for cinder services] ************",
                                "Tuesday 24 February 2026  01:08:28 +0000 (0:00:04.425)       0:00:43.916 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=[{'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-api:2024.2', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8776'], 'timeout': '30'}, 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}}}, {'name': 'rbd-1', 'cluster': 'ceph', 'enabled': True}]) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=[{'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-scheduler:2024.2', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}}, {'name': 'rbd-1', 'cluster': 'ceph', 'enabled': True}]) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=[{'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-api:2024.2', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8776'], 'timeout': '30'}, 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}}}, {'name': 'rbd-1', 'cluster': 'ceph', 'enabled': True}]) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=[{'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-scheduler:2024.2', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}}, {'name': 'rbd-1', 'cluster': 'ceph', 'enabled': True}]) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=[{'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-api:2024.2', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8776'], 'timeout': '30'}, 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}}}, {'name': 'rbd-1', 'cluster': 'ceph', 'enabled': True}]) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=[{'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-scheduler:2024.2', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}}, {'name': 'rbd-1', 'cluster': 'ceph', 'enabled': True}]) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=[{'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-volume:2024.2', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}}, {'name': 'rbd-1', 'cluster': 'ceph', 'enabled': True}])\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=[{'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-volume:2024.2', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}}, {'name': 'rbd-1', 'cluster': 'ceph', 'enabled': True}])\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=[{'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-volume:2024.2', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}}, {'name': 'rbd-1', 'cluster': 'ceph', 'enabled': True}])\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=[{'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-backup:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}}, {'name': 'rbd-1', 'cluster': 'ceph', 'enabled': True}])\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=[{'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-backup:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}}, {'name': 'rbd-1', 'cluster': 'ceph', 'enabled': True}])\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=[{'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-backup:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}}, {'name': 'rbd-1', 'cluster': 'ceph', 'enabled': True}])\u001b[0m",
                                "",
                                "TASK [cinder : Copy over Ceph keyring files for cinder-volume] *****************",
                                "Tuesday 24 February 2026  01:08:33 +0000 (0:00:05.051)       0:00:48.967 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'name': 'rbd-1', 'cluster': 'ceph', 'enabled': True})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'name': 'rbd-1', 'cluster': 'ceph', 'enabled': True})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'name': 'rbd-1', 'cluster': 'ceph', 'enabled': True})\u001b[0m",
                                "",
                                "TASK [cinder : Copy over Ceph keyring files for cinder-backup] *****************",
                                "Tuesday 24 February 2026  01:08:35 +0000 (0:00:02.638)       0:00:51.605 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=ceph.client.cinder.keyring)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=ceph.client.cinder.keyring)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=ceph.client.cinder.keyring)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=ceph.client.cinder-backup.keyring)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=ceph.client.cinder-backup.keyring)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=ceph.client.cinder-backup.keyring)\u001b[0m",
                                "",
                                "TASK [cinder : Ensuring config directory has correct owner and permission] *****",
                                "Tuesday 24 February 2026  01:08:39 +0000 (0:00:03.572)       0:00:55.178 ****** ",
                                "\u001b[0;32mok: [testbed-node-0] => (item=cinder-volume)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item=cinder-volume)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0] => (item=cinder-backup)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item=cinder-volume)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item=cinder-backup)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item=cinder-backup)\u001b[0m",
                                "",
                                "TASK [cinder : Check if policies shall be overwritten] *************************",
                                "Tuesday 24 February 2026  01:08:40 +0000 (0:00:01.564)       0:00:56.743 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [cinder : Set cinder policy file] *****************************************",
                                "Tuesday 24 February 2026  01:08:41 +0000 (0:00:00.429)       0:00:57.173 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [cinder : include_tasks] **************************************************",
                                "Tuesday 24 February 2026  01:08:41 +0000 (0:00:00.375)       0:00:57.548 ****** ",
                                "\u001b[0;36mincluded: /ansible/roles/cinder/tasks/copy-certs.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [service-cert-copy : cinder | Copying over extra CA certificates] *********",
                                "Tuesday 24 February 2026  01:08:42 +0000 (0:00:01.041)       0:00:58.590 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-api:2024.2', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8776'], 'timeout': '30'}, 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-api:2024.2', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8776'], 'timeout': '30'}, 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-api:2024.2', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8776'], 'timeout': '30'}, 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-scheduler:2024.2', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-scheduler:2024.2', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-scheduler:2024.2', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-volume:2024.2', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-volume:2024.2', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-volume:2024.2', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-backup:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-backup:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-backup:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}})\u001b[0m",
                                "",
                                "TASK [service-cert-copy : cinder | Copying over backend internal TLS certificate] ***",
                                "Tuesday 24 February 2026  01:08:47 +0000 (0:00:04.906)       0:01:03.497 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-api:2024.2', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8776'], 'timeout': '30'}, 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-api:2024.2', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8776'], 'timeout': '30'}, 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-scheduler:2024.2', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-scheduler:2024.2', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-volume:2024.2', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-volume:2024.2', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-backup:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-backup:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-api:2024.2', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8776'], 'timeout': '30'}, 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-scheduler:2024.2', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-volume:2024.2', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-backup:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [service-cert-copy : cinder | Copying over backend internal TLS key] ******",
                                "Tuesday 24 February 2026  01:08:48 +0000 (0:00:01.140)       0:01:04.637 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-api:2024.2', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8776'], 'timeout': '30'}, 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-scheduler:2024.2', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-volume:2024.2', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-backup:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-api:2024.2', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8776'], 'timeout': '30'}, 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-scheduler:2024.2', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-volume:2024.2', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-backup:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-api:2024.2', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8776'], 'timeout': '30'}, 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-scheduler:2024.2', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-volume:2024.2', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-backup:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [cinder : Copying over config.json files for services] ********************",
                                "Tuesday 24 February 2026  01:08:50 +0000 (0:00:02.056)       0:01:06.694 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-api:2024.2', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8776'], 'timeout': '30'}, 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-api:2024.2', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8776'], 'timeout': '30'}, 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-api:2024.2', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8776'], 'timeout': '30'}, 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-scheduler:2024.2', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-scheduler:2024.2', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-scheduler:2024.2', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-volume:2024.2', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-volume:2024.2', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-volume:2024.2', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-backup:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-backup:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-backup:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}})\u001b[0m",
                                "",
                                "TASK [cinder : Copying over cinder-wsgi.conf] **********************************",
                                "Tuesday 24 February 2026  01:08:56 +0000 (0:00:05.870)       0:01:12.565 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=/ansible/roles/cinder/templates/cinder-wsgi.conf.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=/ansible/roles/cinder/templates/cinder-wsgi.conf.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=/ansible/roles/cinder/templates/cinder-wsgi.conf.j2)\u001b[0m",
                                "",
                                "TASK [cinder : Copying over cinder.conf] ***************************************",
                                "Tuesday 24 February 2026  01:08:59 +0000 (0:00:02.417)       0:01:14.982 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-api:2024.2', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8776'], 'timeout': '30'}, 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-api:2024.2', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8776'], 'timeout': '30'}, 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-api:2024.2', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8776'], 'timeout': '30'}, 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-scheduler:2024.2', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-scheduler:2024.2', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-scheduler:2024.2', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-volume:2024.2', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-volume:2024.2', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-volume:2024.2', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-backup:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-backup:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-backup:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}})\u001b[0m",
                                "",
                                "TASK [cinder : Generating 'hostnqn' file for cinder_volume] ********************",
                                "Tuesday 24 February 2026  01:09:16 +0000 (0:00:16.882)       0:01:31.865 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [cinder : Copying over existing policy file] ******************************",
                                "Tuesday 24 February 2026  01:09:18 +0000 (0:00:02.007)       0:01:33.872 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-api:2024.2', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8776'], 'timeout': '30'}, 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-scheduler:2024.2', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-volume:2024.2', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-backup:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-api:2024.2', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8776'], 'timeout': '30'}, 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-scheduler:2024.2', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-volume:2024.2', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-api:2024.2', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8776'], 'timeout': '30'}, 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-backup:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-scheduler:2024.2', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-volume:2024.2', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-backup:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [cinder : Copying over nfs_shares files for cinder_volume] ****************",
                                "Tuesday 24 February 2026  01:09:18 +0000 (0:00:00.725)       0:01:34.597 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [cinder : Check cinder containers] ****************************************",
                                "Tuesday 24 February 2026  01:09:19 +0000 (0:00:00.353)       0:01:34.950 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-api:2024.2', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8776'], 'timeout': '30'}, 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-api:2024.2', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8776'], 'timeout': '30'}, 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'cinder-api', 'value': {'container_name': 'cinder_api', 'group': 'cinder-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-api:2024.2', 'volumes': ['/etc/kolla/cinder-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8776'], 'timeout': '30'}, 'haproxy': {'cinder_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}, 'cinder_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8776', 'listen_port': '8776', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-scheduler:2024.2', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-scheduler:2024.2', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'cinder-scheduler', 'value': {'container_name': 'cinder_scheduler', 'group': 'cinder-scheduler', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-scheduler:2024.2', 'volumes': ['/etc/kolla/cinder-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-scheduler 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-volume:2024.2', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-volume:2024.2', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'cinder-volume', 'value': {'container_name': 'cinder_volume', 'group': 'cinder-volume', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-volume:2024.2', 'privileged': True, 'ipc_mode': 'host', 'tmpfs': [''], 'volumes': ['/etc/kolla/cinder-volume/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-volume 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-backup:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-backup:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'cinder-backup', 'value': {'container_name': 'cinder_backup', 'group': 'cinder-backup', 'enabled': True, 'image': 'registry.osism.tech/kolla/cinder-backup:2024.2', 'privileged': True, 'volumes': ['/etc/kolla/cinder-backup/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/dev/:/dev/', '/lib/modules:/lib/modules:ro', '/run:/run:shared', 'cinder:/var/lib/cinder', '', 'kolla_logs:/var/log/kolla/', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port cinder-backup 5672'], 'timeout': '30'}}})\u001b[0m",
                                "",
                                "TASK [cinder : include_tasks] **************************************************",
                                "Tuesday 24 February 2026  01:09:23 +0000 (0:00:04.005)       0:01:38.956 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [cinder : Creating Cinder database] ***************************************",
                                "Tuesday 24 February 2026  01:09:23 +0000 (0:00:00.454)       0:01:39.411 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [cinder : Creating Cinder database user and setting permissions] **********",
                                "Tuesday 24 February 2026  01:09:25 +0000 (0:00:02.302)       0:01:41.714 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [cinder : Running Cinder bootstrap container] *****************************",
                                "Tuesday 24 February 2026  01:09:28 +0000 (0:00:02.453)       0:01:44.167 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [cinder : Flush handlers] *************************************************",
                                "Tuesday 24 February 2026  01:09:50 +0000 (0:00:22.048)       0:02:06.216 ****** ",
                                "",
                                "TASK [cinder : Flush handlers] *************************************************",
                                "Tuesday 24 February 2026  01:09:50 +0000 (0:00:00.083)       0:02:06.299 ****** ",
                                "",
                                "TASK [cinder : Flush handlers] *************************************************",
                                "Tuesday 24 February 2026  01:09:50 +0000 (0:00:00.114)       0:02:06.414 ****** ",
                                "",
                                "RUNNING HANDLER [cinder : Restart cinder-api container] ************************",
                                "Tuesday 24 February 2026  01:09:50 +0000 (0:00:00.127)       0:02:06.541 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [cinder : Restart cinder-scheduler container] ******************",
                                "Tuesday 24 February 2026  01:10:16 +0000 (0:00:26.155)       0:02:32.697 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [cinder : Restart cinder-volume container] *********************",
                                "Tuesday 24 February 2026  01:10:27 +0000 (0:00:11.017)       0:02:43.714 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "",
                                "RUNNING HANDLER [cinder : Restart cinder-backup container] *********************",
                                "Tuesday 24 February 2026  01:10:46 +0000 (0:00:18.186)       0:03:01.901 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "",
                                "RUNNING HANDLER [cinder : Wait for cinder services to update service versions] ***",
                                "Tuesday 24 February 2026  01:10:57 +0000 (0:00:11.005)       0:03:12.907 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=30  \u001b[0m \u001b[0;33mchanged=22  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=9   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=21  \u001b[0m \u001b[0;33mchanged=15  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=7   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=21  \u001b[0m \u001b[0;33mchanged=15  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=7   \u001b[0m rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Tuesday 24 February 2026  01:10:57 +0000 (0:00:00.259)       0:03:13.166 ****** ",
                                "=============================================================================== ",
                                "cinder : Restart cinder-api container ---------------------------------- 26.16s",
                                "cinder : Running Cinder bootstrap container ---------------------------- 22.05s",
                                "cinder : Restart cinder-volume container ------------------------------- 18.19s",
                                "cinder : Copying over cinder.conf -------------------------------------- 16.88s",
                                "cinder : Restart cinder-scheduler container ---------------------------- 11.02s",
                                "cinder : Restart cinder-backup container ------------------------------- 11.01s",
                                "service-ks-register : cinder | Granting user roles ---------------------- 8.02s",
                                "service-ks-register : cinder | Creating endpoints ----------------------- 7.41s",
                                "cinder : Copying over config.json files for services -------------------- 5.87s",
                                "cinder : Copying over multiple ceph.conf for cinder services ------------ 5.05s",
                                "service-cert-copy : cinder | Copying over extra CA certificates --------- 4.91s",
                                "service-ks-register : cinder | Creating users --------------------------- 4.49s",
                                "cinder : Ensuring cinder service ceph config subdirs exists ------------- 4.42s",
                                "service-ks-register : cinder | Creating roles --------------------------- 4.39s",
                                "cinder : Check cinder containers ---------------------------------------- 4.01s",
                                "service-ks-register : cinder | Creating projects ------------------------ 3.96s",
                                "service-ks-register : cinder | Creating services ------------------------ 3.95s",
                                "cinder : Copy over Ceph keyring files for cinder-backup ----------------- 3.57s",
                                "cinder : Ensuring config directories exist ------------------------------ 2.80s",
                                "cinder : Copy over Ceph keyring files for cinder-volume ----------------- 2.64s",
                                "\u001b[32m2026-02-24 01:10:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask de6aa5ad-1dfd-4a58-bf00-4755d79ef102 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:10:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 918da279-e500-4c4b-bdd7-1bd26f3de1b1 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:10:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4baf3a0c-387b-4215-b490-5edd972623e9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:10:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 01:11:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask de6aa5ad-1dfd-4a58-bf00-4755d79ef102 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:11:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 918da279-e500-4c4b-bdd7-1bd26f3de1b1 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:11:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4baf3a0c-387b-4215-b490-5edd972623e9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:11:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 01:11:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask de6aa5ad-1dfd-4a58-bf00-4755d79ef102 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:11:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 918da279-e500-4c4b-bdd7-1bd26f3de1b1 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:11:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4baf3a0c-387b-4215-b490-5edd972623e9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:11:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 01:11:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask de6aa5ad-1dfd-4a58-bf00-4755d79ef102 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:11:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 918da279-e500-4c4b-bdd7-1bd26f3de1b1 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:11:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4baf3a0c-387b-4215-b490-5edd972623e9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:11:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 01:11:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask de6aa5ad-1dfd-4a58-bf00-4755d79ef102 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:11:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 918da279-e500-4c4b-bdd7-1bd26f3de1b1 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:11:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4baf3a0c-387b-4215-b490-5edd972623e9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:11:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 01:11:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask de6aa5ad-1dfd-4a58-bf00-4755d79ef102 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:11:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 918da279-e500-4c4b-bdd7-1bd26f3de1b1 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:11:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4baf3a0c-387b-4215-b490-5edd972623e9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:11:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 01:11:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask de6aa5ad-1dfd-4a58-bf00-4755d79ef102 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:11:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 918da279-e500-4c4b-bdd7-1bd26f3de1b1 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:11:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4baf3a0c-387b-4215-b490-5edd972623e9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:11:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 01:11:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask de6aa5ad-1dfd-4a58-bf00-4755d79ef102 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:11:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 918da279-e500-4c4b-bdd7-1bd26f3de1b1 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:11:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4baf3a0c-387b-4215-b490-5edd972623e9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:11:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 01:11:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask de6aa5ad-1dfd-4a58-bf00-4755d79ef102 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:11:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 918da279-e500-4c4b-bdd7-1bd26f3de1b1 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:11:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4baf3a0c-387b-4215-b490-5edd972623e9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:11:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 01:11:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask de6aa5ad-1dfd-4a58-bf00-4755d79ef102 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:11:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 918da279-e500-4c4b-bdd7-1bd26f3de1b1 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:11:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4baf3a0c-387b-4215-b490-5edd972623e9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:11:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 01:11:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask de6aa5ad-1dfd-4a58-bf00-4755d79ef102 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:11:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 918da279-e500-4c4b-bdd7-1bd26f3de1b1 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:13:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4baf3a0c-387b-4215-b490-5edd972623e9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:13:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 01:13:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask de6aa5ad-1dfd-4a58-bf00-4755d79ef102 is in state SUCCESS\u001b[0m",
                                "\u001b[32m2026-02-24 01:13:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 918da279-e500-4c4b-bdd7-1bd26f3de1b1 is in state SUCCESS\u001b[0m",
                                "",
                                "",
                                "PLAY [Group hosts based on configuration] **************************************",
                                "",
                                "TASK [Group hosts based on Kolla action] ***************************************",
                                "Tuesday 24 February 2026  01:09:51 +0000 (0:00:00.430)       0:00:00.430 ****** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [Group hosts based on enabled services] ***********************************",
                                "Tuesday 24 February 2026  01:09:52 +0000 (0:00:00.701)       0:00:01.132 ****** ",
                                "\u001b[0;32mok: [testbed-node-0] => (item=enable_nova_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item=enable_nova_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item=enable_nova_True)\u001b[0m",
                                "",
                                "PLAY [Wait for the Nova service] ***********************************************",
                                "",
                                "TASK [Waiting for Nova public port to be UP] ***********************************",
                                "Tuesday 24 February 2026  01:09:53 +0000 (0:00:01.455)       0:00:02.587 ****** ",
                                "",
                                "STILL ALIVE [task 'Waiting for Nova public port to be UP' is running] **********",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;32mtestbed-node-0\u001b[0m             : \u001b[0;32mok=3   \u001b[0m changed=0    unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "\u001b[0;32mtestbed-node-1\u001b[0m             : \u001b[0;32mok=3   \u001b[0m changed=0    unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "\u001b[0;32mtestbed-node-2\u001b[0m             : \u001b[0;32mok=3   \u001b[0m changed=0    unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Tuesday 24 February 2026  01:12:32 +0000 (0:02:38.788)       0:02:41.376 ****** ",
                                "=============================================================================== ",
                                "Waiting for Nova public port to be UP --------------------------------- 158.79s",
                                "Group hosts based on enabled services ----------------------------------- 1.45s",
                                "Group hosts based on Kolla action --------------------------------------- 0.70s",
                                "",
                                "",
                                "PLAY [Group hosts based on configuration] **************************************",
                                "",
                                "TASK [Group hosts based on Kolla action] ***************************************",
                                "Tuesday 24 February 2026  01:10:33 +0000 (0:00:00.292)       0:00:00.292 ****** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [Group hosts based on enabled services] ***********************************",
                                "Tuesday 24 February 2026  01:10:33 +0000 (0:00:00.324)       0:00:00.616 ****** ",
                                "\u001b[0;32mok: [testbed-node-0] => (item=enable_grafana_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item=enable_grafana_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item=enable_grafana_True)\u001b[0m",
                                "",
                                "PLAY [Apply role grafana] ******************************************************",
                                "",
                                "TASK [grafana : include_tasks] *************************************************",
                                "Tuesday 24 February 2026  01:10:34 +0000 (0:00:00.486)       0:00:01.103 ****** ",
                                "\u001b[0;36mincluded: /ansible/roles/grafana/tasks/deploy.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [grafana : Ensuring config directories exist] *****************************",
                                "Tuesday 24 February 2026  01:10:34 +0000 (0:00:00.612)       0:00:01.715 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'grafana', 'value': {'container_name': 'grafana', 'group': 'grafana', 'enabled': True, 'image': 'registry.osism.tech/kolla/grafana:2024.2', 'volumes': ['/etc/kolla/grafana/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'grafana_server': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000'}, 'grafana_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'grafana', 'value': {'container_name': 'grafana', 'group': 'grafana', 'enabled': True, 'image': 'registry.osism.tech/kolla/grafana:2024.2', 'volumes': ['/etc/kolla/grafana/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'grafana_server': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000'}, 'grafana_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'grafana', 'value': {'container_name': 'grafana', 'group': 'grafana', 'enabled': True, 'image': 'registry.osism.tech/kolla/grafana:2024.2', 'volumes': ['/etc/kolla/grafana/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'grafana_server': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000'}, 'grafana_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000'}}}})\u001b[0m",
                                "",
                                "TASK [grafana : Check if extra configuration file exists] **********************",
                                "Tuesday 24 February 2026  01:10:35 +0000 (0:00:00.778)       0:00:02.494 ****** ",
                                "\u001b[1;35m[WARNING]: Skipped '/operations/prometheus/grafana' path due to this access\u001b[0m",
                                "\u001b[1;35missue: '/operations/prometheus/grafana' is not a directory\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0 -> localhost]\u001b[0m",
                                "",
                                "TASK [grafana : include_tasks] *************************************************",
                                "Tuesday 24 February 2026  01:10:36 +0000 (0:00:00.998)       0:00:03.492 ****** ",
                                "\u001b[0;36mincluded: /ansible/roles/grafana/tasks/copy-certs.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [service-cert-copy : grafana | Copying over extra CA certificates] ********",
                                "Tuesday 24 February 2026  01:10:37 +0000 (0:00:00.919)       0:00:04.412 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'grafana', 'value': {'container_name': 'grafana', 'group': 'grafana', 'enabled': True, 'image': 'registry.osism.tech/kolla/grafana:2024.2', 'volumes': ['/etc/kolla/grafana/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'grafana_server': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000'}, 'grafana_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'grafana', 'value': {'container_name': 'grafana', 'group': 'grafana', 'enabled': True, 'image': 'registry.osism.tech/kolla/grafana:2024.2', 'volumes': ['/etc/kolla/grafana/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'grafana_server': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000'}, 'grafana_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'grafana', 'value': {'container_name': 'grafana', 'group': 'grafana', 'enabled': True, 'image': 'registry.osism.tech/kolla/grafana:2024.2', 'volumes': ['/etc/kolla/grafana/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'grafana_server': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000'}, 'grafana_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000'}}}})\u001b[0m",
                                "",
                                "TASK [service-cert-copy : grafana | Copying over backend internal TLS certificate] ***",
                                "Tuesday 24 February 2026  01:10:38 +0000 (0:00:01.346)       0:00:05.758 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'grafana', 'value': {'container_name': 'grafana', 'group': 'grafana', 'enabled': True, 'image': 'registry.osism.tech/kolla/grafana:2024.2', 'volumes': ['/etc/kolla/grafana/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'grafana_server': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000'}, 'grafana_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'grafana', 'value': {'container_name': 'grafana', 'group': 'grafana', 'enabled': True, 'image': 'registry.osism.tech/kolla/grafana:2024.2', 'volumes': ['/etc/kolla/grafana/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'grafana_server': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000'}, 'grafana_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'grafana', 'value': {'container_name': 'grafana', 'group': 'grafana', 'enabled': True, 'image': 'registry.osism.tech/kolla/grafana:2024.2', 'volumes': ['/etc/kolla/grafana/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'grafana_server': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000'}, 'grafana_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [service-cert-copy : grafana | Copying over backend internal TLS key] *****",
                                "Tuesday 24 February 2026  01:10:39 +0000 (0:00:00.459)       0:00:06.218 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'grafana', 'value': {'container_name': 'grafana', 'group': 'grafana', 'enabled': True, 'image': 'registry.osism.tech/kolla/grafana:2024.2', 'volumes': ['/etc/kolla/grafana/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'grafana_server': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000'}, 'grafana_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'grafana', 'value': {'container_name': 'grafana', 'group': 'grafana', 'enabled': True, 'image': 'registry.osism.tech/kolla/grafana:2024.2', 'volumes': ['/etc/kolla/grafana/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'grafana_server': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000'}, 'grafana_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'grafana', 'value': {'container_name': 'grafana', 'group': 'grafana', 'enabled': True, 'image': 'registry.osism.tech/kolla/grafana:2024.2', 'volumes': ['/etc/kolla/grafana/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'grafana_server': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000'}, 'grafana_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [grafana : Copying over config.json files] ********************************",
                                "Tuesday 24 February 2026  01:10:40 +0000 (0:00:00.967)       0:00:07.185 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'grafana', 'value': {'container_name': 'grafana', 'group': 'grafana', 'enabled': True, 'image': 'registry.osism.tech/kolla/grafana:2024.2', 'volumes': ['/etc/kolla/grafana/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'grafana_server': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000'}, 'grafana_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'grafana', 'value': {'container_name': 'grafana', 'group': 'grafana', 'enabled': True, 'image': 'registry.osism.tech/kolla/grafana:2024.2', 'volumes': ['/etc/kolla/grafana/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'grafana_server': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000'}, 'grafana_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'grafana', 'value': {'container_name': 'grafana', 'group': 'grafana', 'enabled': True, 'image': 'registry.osism.tech/kolla/grafana:2024.2', 'volumes': ['/etc/kolla/grafana/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'grafana_server': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000'}, 'grafana_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000'}}}})\u001b[0m",
                                "",
                                "TASK [grafana : Copying over grafana.ini] **************************************",
                                "Tuesday 24 February 2026  01:10:41 +0000 (0:00:01.228)       0:00:08.414 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'grafana', 'value': {'container_name': 'grafana', 'group': 'grafana', 'enabled': True, 'image': 'registry.osism.tech/kolla/grafana:2024.2', 'volumes': ['/etc/kolla/grafana/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'grafana_server': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000'}, 'grafana_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'grafana', 'value': {'container_name': 'grafana', 'group': 'grafana', 'enabled': True, 'image': 'registry.osism.tech/kolla/grafana:2024.2', 'volumes': ['/etc/kolla/grafana/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'grafana_server': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000'}, 'grafana_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'grafana', 'value': {'container_name': 'grafana', 'group': 'grafana', 'enabled': True, 'image': 'registry.osism.tech/kolla/grafana:2024.2', 'volumes': ['/etc/kolla/grafana/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'grafana_server': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000'}, 'grafana_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000'}}}})\u001b[0m",
                                "",
                                "TASK [grafana : Copying over extra configuration file] *************************",
                                "Tuesday 24 February 2026  01:10:42 +0000 (0:00:01.471)       0:00:09.886 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [grafana : Configuring Prometheus as data source for Grafana] *************",
                                "Tuesday 24 February 2026  01:10:43 +0000 (0:00:00.573)       0:00:10.460 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=/ansible/roles/grafana/templates/prometheus.yaml.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=/ansible/roles/grafana/templates/prometheus.yaml.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=/ansible/roles/grafana/templates/prometheus.yaml.j2)\u001b[0m",
                                "",
                                "TASK [grafana : Configuring dashboards provisioning] ***************************",
                                "Tuesday 24 February 2026  01:10:44 +0000 (0:00:01.188)       0:00:11.648 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=/opt/configuration/environments/kolla/files/overlays/grafana/provisioning.yaml)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=/opt/configuration/environments/kolla/files/overlays/grafana/provisioning.yaml)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=/opt/configuration/environments/kolla/files/overlays/grafana/provisioning.yaml)\u001b[0m",
                                "",
                                "TASK [grafana : Find custom grafana dashboards] ********************************",
                                "Tuesday 24 February 2026  01:10:45 +0000 (0:00:01.212)       0:00:12.861 ****** ",
                                "\u001b[0;32mok: [testbed-node-0 -> localhost]\u001b[0m",
                                "",
                                "TASK [grafana : Find templated grafana dashboards] *****************************",
                                "Tuesday 24 February 2026  01:10:46 +0000 (0:00:00.839)       0:00:13.700 ****** ",
                                "\u001b[1;35m[WARNING]: Skipped '/etc/kolla/grafana/dashboards' path due to this access\u001b[0m",
                                "\u001b[1;35missue: '/etc/kolla/grafana/dashboards' is not a directory\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [grafana : Prune templated Grafana dashboards] ****************************",
                                "Tuesday 24 February 2026  01:10:47 +0000 (0:00:00.792)       0:00:14.493 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [grafana : Copying over custom dashboards] ********************************",
                                "Tuesday 24 February 2026  01:10:48 +0000 (0:00:00.780)       0:00:15.274 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ceph/ceph-cluster-advanced.json', 'value': {'path': '/operations/grafana/dashboards/ceph/ceph-cluster-advanced.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 121701, 'inode': 1316890, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.147216, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ceph/ceph-cluster-advanced.json', 'value': {'path': '/operations/grafana/dashboards/ceph/ceph-cluster-advanced.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 121701, 'inode': 1316890, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.147216, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ceph/ceph-cluster-advanced.json', 'value': {'path': '/operations/grafana/dashboards/ceph/ceph-cluster-advanced.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 121701, 'inode': 1316890, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.147216, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ceph/cephfsdashboard.json', 'value': {'path': '/operations/grafana/dashboards/ceph/cephfsdashboard.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 143913, 'inode': 1316910, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.155399, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ceph/cephfsdashboard.json', 'value': {'path': '/operations/grafana/dashboards/ceph/cephfsdashboard.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 143913, 'inode': 1316910, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.155399, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ceph/cephfsdashboard.json', 'value': {'path': '/operations/grafana/dashboards/ceph/cephfsdashboard.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 143913, 'inode': 1316910, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.155399, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ceph/rbd-overview.json', 'value': {'path': '/operations/grafana/dashboards/ceph/rbd-overview.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 26019, 'inode': 1316976, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.180783, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ceph/rbd-overview.json', 'value': {'path': '/operations/grafana/dashboards/ceph/rbd-overview.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 26019, 'inode': 1316976, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.180783, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ceph/rbd-overview.json', 'value': {'path': '/operations/grafana/dashboards/ceph/rbd-overview.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 26019, 'inode': 1316976, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.180783, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ceph/ceph_pools.json', 'value': {'path': '/operations/grafana/dashboards/ceph/ceph_pools.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 25279, 'inode': 1316904, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.152629, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ceph/ceph_pools.json', 'value': {'path': '/operations/grafana/dashboards/ceph/ceph_pools.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 25279, 'inode': 1316904, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.152629, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ceph/ceph_pools.json', 'value': {'path': '/operations/grafana/dashboards/ceph/ceph_pools.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 25279, 'inode': 1316904, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.152629, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ceph/rgw-s3-analytics.json', 'value': {'path': '/operations/grafana/dashboards/ceph/rgw-s3-analytics.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 170293, 'inode': 1316977, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.1822166, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ceph/rgw-s3-analytics.json', 'value': {'path': '/operations/grafana/dashboards/ceph/rgw-s3-analytics.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 170293, 'inode': 1316977, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.1822166, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ceph/rgw-s3-analytics.json', 'value': {'path': '/operations/grafana/dashboards/ceph/rgw-s3-analytics.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 170293, 'inode': 1316977, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.1822166, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ceph/ceph-nvmeof-performance.json', 'value': {'path': '/operations/grafana/dashboards/ceph/ceph-nvmeof-performance.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 33297, 'inode': 1316894, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.1502383, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ceph/ceph-nvmeof-performance.json', 'value': {'path': '/operations/grafana/dashboards/ceph/ceph-nvmeof-performance.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 33297, 'inode': 1316894, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.1502383, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ceph/ceph-nvmeof-performance.json', 'value': {'path': '/operations/grafana/dashboards/ceph/ceph-nvmeof-performance.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 33297, 'inode': 1316894, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.1502383, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ceph/osd-device-details.json', 'value': {'path': '/operations/grafana/dashboards/ceph/osd-device-details.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 26346, 'inode': 1316926, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.1593084, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ceph/osd-device-details.json', 'value': {'path': '/operations/grafana/dashboards/ceph/osd-device-details.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 26346, 'inode': 1316926, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.1593084, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ceph/osd-device-details.json', 'value': {'path': '/operations/grafana/dashboards/ceph/osd-device-details.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 26346, 'inode': 1316926, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.1593084, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ceph/radosgw-overview.json', 'value': {'path': '/operations/grafana/dashboards/ceph/radosgw-overview.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 46110, 'inode': 1316969, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.1797712, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ceph/radosgw-overview.json', 'value': {'path': '/operations/grafana/dashboards/ceph/radosgw-overview.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 46110, 'inode': 1316969, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.1797712, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ceph/radosgw-overview.json', 'value': {'path': '/operations/grafana/dashboards/ceph/radosgw-overview.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 46110, 'inode': 1316969, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.1797712, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ceph/README.md', 'value': {'path': '/operations/grafana/dashboards/ceph/README.md', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 84, 'inode': 1316889, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.1467896, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ceph/README.md', 'value': {'path': '/operations/grafana/dashboards/ceph/README.md', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 84, 'inode': 1316889, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.1467896, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ceph/README.md', 'value': {'path': '/operations/grafana/dashboards/ceph/README.md', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 84, 'inode': 1316889, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.1467896, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ceph/ceph-cluster.json', 'value': {'path': '/operations/grafana/dashboards/ceph/ceph-cluster.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 34113, 'inode': 1316891, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.1491747, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ceph/ceph-cluster.json', 'value': {'path': '/operations/grafana/dashboards/ceph/ceph-cluster.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 34113, 'inode': 1316891, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.1491747, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ceph/ceph-cluster.json', 'value': {'path': '/operations/grafana/dashboards/ceph/ceph-cluster.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 34113, 'inode': 1316891, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.1491747, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ceph/cephfs-overview.json', 'value': {'path': '/operations/grafana/dashboards/ceph/cephfs-overview.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 9025, 'inode': 1316907, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.1528525, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ceph/cephfs-overview.json', 'value': {'path': '/operations/grafana/dashboards/ceph/cephfs-overview.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 9025, 'inode': 1316907, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.1528525, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ceph/cephfs-overview.json', 'value': {'path': '/operations/grafana/dashboards/ceph/cephfs-overview.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 9025, 'inode': 1316907, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.1528525, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ceph/pool-detail.json', 'value': {'path': '/operations/grafana/dashboards/ceph/pool-detail.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 19231, 'inode': 1316930, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.160852, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ceph/pool-detail.json', 'value': {'path': '/operations/grafana/dashboards/ceph/pool-detail.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 19231, 'inode': 1316930, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.160852, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ceph/pool-detail.json', 'value': {'path': '/operations/grafana/dashboards/ceph/pool-detail.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 19231, 'inode': 1316930, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.160852, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ceph/rbd-details.json', 'value': {'path': '/operations/grafana/dashboards/ceph/rbd-details.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 13320, 'inode': 1316975, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.1803944, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ceph/rbd-details.json', 'value': {'path': '/operations/grafana/dashboards/ceph/rbd-details.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 13320, 'inode': 1316975, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.1803944, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ceph/rbd-details.json', 'value': {'path': '/operations/grafana/dashboards/ceph/rbd-details.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 13320, 'inode': 1316975, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.1803944, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ceph/ceph_overview.json', 'value': {'path': '/operations/grafana/dashboards/ceph/ceph_overview.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 80386, 'inode': 1316900, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.1517806, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ceph/ceph_overview.json', 'value': {'path': '/operations/grafana/dashboards/ceph/ceph_overview.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 80386, 'inode': 1316900, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.1517806, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ceph/ceph_overview.json', 'value': {'path': '/operations/grafana/dashboards/ceph/ceph_overview.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 80386, 'inode': 1316900, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.1517806, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ceph/radosgw-detail.json', 'value': {'path': '/operations/grafana/dashboards/ceph/radosgw-detail.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 20042, 'inode': 1316935, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.1782167, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ceph/radosgw-detail.json', 'value': {'path': '/operations/grafana/dashboards/ceph/radosgw-detail.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 20042, 'inode': 1316935, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.1782167, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ceph/radosgw-detail.json', 'value': {'path': '/operations/grafana/dashboards/ceph/radosgw-detail.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 20042, 'inode': 1316935, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.1782167, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ceph/smb-overview.json', 'value': {'path': '/operations/grafana/dashboards/ceph/smb-overview.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 29877, 'inode': 1316980, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.1822166, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ceph/smb-overview.json', 'value': {'path': '/operations/grafana/dashboards/ceph/smb-overview.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 29877, 'inode': 1316980, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.1822166, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ceph/smb-overview.json', 'value': {'path': '/operations/grafana/dashboards/ceph/smb-overview.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 29877, 'inode': 1316980, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.1822166, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ceph/osds-overview.json', 'value': {'path': '/operations/grafana/dashboards/ceph/osds-overview.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 38375, 'inode': 1316927, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.160412, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ceph/osds-overview.json', 'value': {'path': '/operations/grafana/dashboards/ceph/osds-overview.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 38375, 'inode': 1316927, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.160412, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ceph/osds-overview.json', 'value': {'path': '/operations/grafana/dashboards/ceph/osds-overview.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 38375, 'inode': 1316927, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.160412, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ceph/multi-cluster-overview.json', 'value': {'path': '/operations/grafana/dashboards/ceph/multi-cluster-overview.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 63043, 'inode': 1316922, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.1593084, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ceph/multi-cluster-overview.json', 'value': {'path': '/operations/grafana/dashboards/ceph/multi-cluster-overview.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 63043, 'inode': 1316922, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.1593084, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ceph/multi-cluster-overview.json', 'value': {'path': '/operations/grafana/dashboards/ceph/multi-cluster-overview.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 63043, 'inode': 1316922, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.1593084, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ceph/hosts-overview.json', 'value': {'path': '/operations/grafana/dashboards/ceph/hosts-overview.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 27387, 'inode': 1316919, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.1575785, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ceph/hosts-overview.json', 'value': {'path': '/operations/grafana/dashboards/ceph/hosts-overview.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 27387, 'inode': 1316919, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.1575785, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ceph/hosts-overview.json', 'value': {'path': '/operations/grafana/dashboards/ceph/hosts-overview.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 27387, 'inode': 1316919, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.1575785, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ceph/pool-overview.json', 'value': {'path': '/operations/grafana/dashboards/ceph/pool-overview.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 49016, 'inode': 1316932, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.1612163, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ceph/pool-overview.json', 'value': {'path': '/operations/grafana/dashboards/ceph/pool-overview.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 49016, 'inode': 1316932, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.1612163, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ceph/pool-overview.json', 'value': {'path': '/operations/grafana/dashboards/ceph/pool-overview.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 49016, 'inode': 1316932, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.1612163, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ceph/host-details.json', 'value': {'path': '/operations/grafana/dashboards/ceph/host-details.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 43303, 'inode': 1316916, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.1567206, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ceph/host-details.json', 'value': {'path': '/operations/grafana/dashboards/ceph/host-details.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 43303, 'inode': 1316916, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.1567206, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ceph/host-details.json', 'value': {'path': '/operations/grafana/dashboards/ceph/host-details.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 43303, 'inode': 1316916, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.1567206, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ceph/radosgw-sync-overview.json', 'value': {'path': '/operations/grafana/dashboards/ceph/radosgw-sync-overview.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 16614, 'inode': 1316973, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.1800752, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ceph/radosgw-sync-overview.json', 'value': {'path': '/operations/grafana/dashboards/ceph/radosgw-sync-overview.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 16614, 'inode': 1316973, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.1800752, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ceph/radosgw-sync-overview.json', 'value': {'path': '/operations/grafana/dashboards/ceph/radosgw-sync-overview.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 16614, 'inode': 1316973, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.1800752, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'ceph/ceph-nvmeof.json', 'value': {'path': '/operations/grafana/dashboards/ceph/ceph-nvmeof.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 52667, 'inode': 1316898, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.1502383, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'ceph/ceph-nvmeof.json', 'value': {'path': '/operations/grafana/dashboards/ceph/ceph-nvmeof.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 52667, 'inode': 1316898, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.1502383, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'ceph/ceph-nvmeof.json', 'value': {'path': '/operations/grafana/dashboards/ceph/ceph-nvmeof.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 52667, 'inode': 1316898, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.1502383, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'openstack/openstack.json', 'value': {'path': '/operations/grafana/dashboards/openstack/openstack.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 57270, 'inode': 1317236, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.3583703, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'openstack/openstack.json', 'value': {'path': '/operations/grafana/dashboards/openstack/openstack.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 57270, 'inode': 1317236, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.3583703, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'openstack/openstack.json', 'value': {'path': '/operations/grafana/dashboards/openstack/openstack.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 57270, 'inode': 1317236, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.3583703, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'infrastructure/haproxy.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/haproxy.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 410814, 'inode': 1317001, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.200857, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'infrastructure/haproxy.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/haproxy.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 410814, 'inode': 1317001, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.200857, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'infrastructure/haproxy.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/haproxy.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 410814, 'inode': 1317001, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.200857, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'infrastructure/database.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/database.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 30898, 'inode': 1316990, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.1872168, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'infrastructure/database.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/database.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 30898, 'inode': 1316990, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.1872168, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'infrastructure/database.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/database.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 30898, 'inode': 1316990, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.1872168, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'infrastructure/node-rsrc-use.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/node-rsrc-use.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 15767, 'inode': 1317017, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.2022812, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'infrastructure/node-rsrc-use.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/node-rsrc-use.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 15767, 'inode': 1317017, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.2022812, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'infrastructure/node-rsrc-use.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/node-rsrc-use.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 15767, 'inode': 1317017, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.2022812, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'infrastructure/alertmanager-overview.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/alertmanager-overview.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 9645, 'inode': 1316984, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.1841078, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'infrastructure/alertmanager-overview.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/alertmanager-overview.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 9645, 'inode': 1316984, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.1841078, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'infrastructure/alertmanager-overview.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/alertmanager-overview.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 9645, 'inode': 1316984, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.1841078, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'infrastructure/opensearch.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/opensearch.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 65458, 'inode': 1317050, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.2272177, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'infrastructure/opensearch.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/opensearch.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 65458, 'inode': 1317050, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.2272177, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'infrastructure/opensearch.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/opensearch.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 65458, 'inode': 1317050, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.2272177, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'infrastructure/node_exporter_full.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/node_exporter_full.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 682774, 'inode': 1317019, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.2222176, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'infrastructure/node_exporter_full.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/node_exporter_full.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 682774, 'inode': 1317019, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.2222176, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'infrastructure/node_exporter_full.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/node_exporter_full.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 682774, 'inode': 1317019, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.2222176, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'infrastructure/prometheus-remote-write.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/prometheus-remote-write.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 22303, 'inode': 1317051, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.2282176, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'infrastructure/prometheus-remote-write.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/prometheus-remote-write.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 22303, 'inode': 1317051, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.2282176, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'infrastructure/prometheus-remote-write.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/prometheus-remote-write.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 22303, 'inode': 1317051, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.2282176, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'infrastructure/redfish.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/redfish.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 38087, 'inode': 1317055, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.2342176, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'infrastructure/redfish.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/redfish.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 38087, 'inode': 1317055, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.2342176, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'infrastructure/redfish.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/redfish.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 38087, 'inode': 1317055, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.2342176, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'infrastructure/nodes.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/nodes.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 21194, 'inode': 1317049, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.2242174, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'infrastructure/nodes.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/nodes.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 21194, 'inode': 1317049, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.2242174, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'infrastructure/nodes.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/nodes.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 21194, 'inode': 1317049, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.2242174, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'infrastructure/memcached.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/memcached.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 24243, 'inode': 1317013, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.2016113, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'infrastructure/memcached.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/memcached.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 24243, 'inode': 1317013, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.2016113, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'infrastructure/memcached.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/memcached.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 24243, 'inode': 1317013, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.2016113, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'infrastructure/fluentd.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/fluentd.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 82960, 'inode': 1316998, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.194241, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'infrastructure/fluentd.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/fluentd.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 82960, 'inode': 1316998, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.194241, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'infrastructure/fluentd.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/fluentd.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 82960, 'inode': 1316998, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.194241, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'infrastructure/libvirt.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/libvirt.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 29672, 'inode': 1317011, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.2011416, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'infrastructure/libvirt.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/libvirt.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 29672, 'inode': 1317011, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.2011416, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'infrastructure/libvirt.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/libvirt.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 29672, 'inode': 1317011, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.2011416, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'infrastructure/elasticsearch.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/elasticsearch.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 187864, 'inode': 1316991, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.1930285, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'infrastructure/elasticsearch.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/elasticsearch.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 187864, 'inode': 1316991, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.1930285, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'infrastructure/elasticsearch.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/elasticsearch.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 187864, 'inode': 1316991, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.1930285, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'infrastructure/node-cluster-rsrc-use.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/node-cluster-rsrc-use.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 15957, 'inode': 1317015, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.2019198, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'infrastructure/node-cluster-rsrc-use.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/node-cluster-rsrc-use.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 15957, 'inode': 1317015, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.2019198, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'infrastructure/node-cluster-rsrc-use.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/node-cluster-rsrc-use.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 15957, 'inode': 1317015, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.2019198, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'infrastructure/rabbitmq.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/rabbitmq.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 222049, 'inode': 1317054, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.2332177, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'infrastructure/rabbitmq.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/rabbitmq.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 222049, 'inode': 1317054, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.2332177, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'infrastructure/rabbitmq.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/rabbitmq.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 222049, 'inode': 1317054, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.2332177, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'infrastructure/prometheus_alertmanager.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/prometheus_alertmanager.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 115472, 'inode': 1317053, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.2312176, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'infrastructure/prometheus_alertmanager.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/prometheus_alertmanager.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 115472, 'inode': 1317053, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.2312176, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'infrastructure/prometheus_alertmanager.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/prometheus_alertmanager.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 115472, 'inode': 1317053, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.2312176, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'infrastructure/blackbox.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/blackbox.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 31128, 'inode': 1316985, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.1844537, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'infrastructure/blackbox.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/blackbox.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 31128, 'inode': 1316985, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.1844537, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'infrastructure/blackbox.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/blackbox.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 31128, 'inode': 1316985, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.1844537, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'infrastructure/cadvisor.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/cadvisor.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 53882, 'inode': 1316987, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.1867146, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'infrastructure/cadvisor.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/cadvisor.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 53882, 'inode': 1316987, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.1867146, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'infrastructure/cadvisor.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/cadvisor.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 53882, 'inode': 1316987, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.1867146, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'infrastructure/node_exporter_side_by_side.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/node_exporter_side_by_side.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 70691, 'inode': 1317048, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.2232175, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'infrastructure/node_exporter_side_by_side.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/node_exporter_side_by_side.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 70691, 'inode': 1317048, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.2232175, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'infrastructure/node_exporter_side_by_side.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/node_exporter_side_by_side.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 70691, 'inode': 1317048, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.2232175, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'infrastructure/prometheus.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/prometheus.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 21951, 'inode': 1317052, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.2292175, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'infrastructure/prometheus.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/prometheus.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 21951, 'inode': 1317052, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.2292175, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'infrastructure/prometheus.json', 'value': {'path': '/operations/grafana/dashboards/infrastructure/prometheus.json', 'mode': '0644', 'isdir': False, 'ischr': False, 'isblk': False, 'isreg': True, 'isfifo': False, 'islnk': False, 'issock': False, 'uid': 0, 'gid': 0, 'size': 21951, 'inode': 1317052, 'dev': 123, 'nlink': 1, 'atime': 1771891342.0, 'mtime': 1771891342.0, 'ctime': 1771892309.2292175, 'gr_name': 'root', 'pw_name': 'root', 'wusr': True, 'rusr': True, 'xusr': False, 'wgrp': False, 'rgrp': True, 'xgrp': False, 'woth': False, 'roth': True, 'xoth': False, 'isuid': False, 'isgid': False}})\u001b[0m",
                                "",
                                "TASK [grafana : Check grafana containers] **************************************",
                                "Tuesday 24 February 2026  01:11:29 +0000 (0:00:41.241)       0:00:56.515 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'grafana', 'value': {'container_name': 'grafana', 'group': 'grafana', 'enabled': True, 'image': 'registry.osism.tech/kolla/grafana:2024.2', 'volumes': ['/etc/kolla/grafana/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'grafana_server': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000'}, 'grafana_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'grafana', 'value': {'container_name': 'grafana', 'group': 'grafana', 'enabled': True, 'image': 'registry.osism.tech/kolla/grafana:2024.2', 'volumes': ['/etc/kolla/grafana/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'grafana_server': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000'}, 'grafana_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'grafana', 'value': {'container_name': 'grafana', 'group': 'grafana', 'enabled': True, 'image': 'registry.osism.tech/kolla/grafana:2024.2', 'volumes': ['/etc/kolla/grafana/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/'], 'dimensions': {}, 'haproxy': {'grafana_server': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '3000', 'listen_port': '3000'}, 'grafana_server_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '3000', 'listen_port': '3000'}}}})\u001b[0m",
                                "",
                                "TASK [grafana : Creating grafana database] *************************************",
                                "Tuesday 24 February 2026  01:11:30 +0000 (0:00:01.099)       0:00:57.614 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [grafana : Creating grafana database user and setting permissions] ********",
                                "Tuesday 24 February 2026  01:11:33 +0000 (0:00:02.638)       0:01:00.253 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [grafana : Flush handlers] ************************************************",
                                "Tuesday 24 February 2026  01:11:35 +0000 (0:00:02.735)       0:01:02.989 ****** ",
                                "",
                                "TASK [grafana : Flush handlers] ************************************************",
                                "Tuesday 24 February 2026  01:11:35 +0000 (0:00:00.066)       0:01:03.055 ****** ",
                                "",
                                "TASK [grafana : Flush handlers] ************************************************",
                                "Tuesday 24 February 2026  01:11:36 +0000 (0:00:00.257)       0:01:03.313 ****** ",
                                "",
                                "RUNNING HANDLER [grafana : Restart first grafana container] ********************",
                                "Tuesday 24 February 2026  01:11:36 +0000 (0:00:00.072)       0:01:03.385 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "RUNNING HANDLER [grafana : Waiting for grafana to start on first node] *********",
                                "Tuesday 24 February 2026  01:11:38 +0000 (0:00:02.118)       0:01:05.504 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[1;30mFAILED - RETRYING: [testbed-node-0]: Waiting for grafana to start on first node (12 retries left).\u001b[0m",
                                "\u001b[1;30mFAILED - RETRYING: [testbed-node-0]: Waiting for grafana to start on first node (11 retries left).\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "RUNNING HANDLER [grafana : Restart remaining grafana containers] ***************",
                                "Tuesday 24 February 2026  01:12:05 +0000 (0:00:27.152)       0:01:32.656 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [grafana : Wait for grafana application ready] ****************************",
                                "Tuesday 24 February 2026  01:12:30 +0000 (0:00:24.871)       0:01:57.527 ****** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [grafana : Remove old grafana docker volume] ******************************",
                                "Tuesday 24 February 2026  01:12:32 +0000 (0:00:02.381)       0:01:59.909 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [grafana : Enable grafana datasources] ************************************",
                                "Tuesday 24 February 2026  01:12:33 +0000 (0:00:00.600)       0:02:00.510 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'influxdb', 'value': {'enabled': False, 'data': {'isDefault': True, 'database': 'telegraf', 'name': 'telegraf', 'type': 'influxdb', 'url': 'https://api-int.testbed.osism.xyz:8086', 'access': 'proxy', 'basicAuth': False}}}) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'opensearch', 'value': {'enabled': True, 'data': {'name': 'opensearch', 'type': 'grafana-opensearch-datasource', 'access': 'proxy', 'url': 'https://api-int.testbed.osism.xyz:9200', 'jsonData': {'flavor': 'OpenSearch', 'database': 'flog-*', 'version': '2.11.1', 'timeField': '@timestamp', 'logLevelField': 'log_level'}}}})\u001b[0m",
                                "",
                                "TASK [grafana : Disable Getting Started panel] *********************************",
                                "Tuesday 24 February 2026  01:12:36 +0000 (0:00:02.886)       0:02:03.396 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=21  \u001b[0m \u001b[0;33mchanged=12  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=7   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=14  \u001b[0m \u001b[0;33mchanged=9   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=7   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=14  \u001b[0m \u001b[0;33mchanged=9   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=7   \u001b[0m rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Tuesday 24 February 2026  01:12:36 +0000 (0:00:00.295)       0:02:03.692 ****** ",
                                "=============================================================================== ",
                                "grafana : Copying over custom dashboards ------------------------------- 41.24s",
                                "grafana : Waiting for grafana to start on first node ------------------- 27.15s",
                                "grafana : Restart remaining grafana containers ------------------------- 24.87s",
                                "grafana : Enable grafana datasources ------------------------------------ 2.89s",
                                "grafana : Creating grafana database user and setting permissions -------- 2.74s",
                                "grafana : Creating grafana database ------------------------------------- 2.64s",
                                "grafana : Wait for grafana application ready ---------------------------- 2.38s",
                                "grafana : Restart first grafana container ------------------------------- 2.12s",
                                "grafana : Copying over grafana.ini -------------------------------------- 1.47s",
                                "service-cert-copy : grafana | Copying over extra CA certificates -------- 1.35s",
                                "grafana : Copying over config.json files -------------------------------- 1.23s",
                                "grafana : Configuring dashboards provisioning --------------------------- 1.21s",
                                "grafana : Configuring Prometheus as data source for Grafana ------------- 1.19s",
                                "grafana : Check grafana containers -------------------------------------- 1.10s",
                                "grafana : Check if extra configuration file exists ---------------------- 1.00s",
                                "service-cert-copy : grafana | Copying over backend internal TLS key ----- 0.97s",
                                "grafana : include_tasks ------------------------------------------------- 0.92s",
                                "grafana : Find custom grafana dashboards -------------------------------- 0.84s",
                                "grafana : Find templated grafana dashboards ----------------------------- 0.79s",
                                "grafana : Prune templated Grafana dashboards ---------------------------- 0.78s",
                                "\u001b[32m2026-02-24 01:13:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 79f56a03-a0d2-400b-aa03-9d9321f018b2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:13:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4baf3a0c-387b-4215-b490-5edd972623e9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:13:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 01:13:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 79f56a03-a0d2-400b-aa03-9d9321f018b2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:13:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4baf3a0c-387b-4215-b490-5edd972623e9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:13:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 01:13:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 79f56a03-a0d2-400b-aa03-9d9321f018b2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:13:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4baf3a0c-387b-4215-b490-5edd972623e9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:13:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 01:13:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 79f56a03-a0d2-400b-aa03-9d9321f018b2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:13:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4baf3a0c-387b-4215-b490-5edd972623e9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:13:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 01:13:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 79f56a03-a0d2-400b-aa03-9d9321f018b2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:13:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4baf3a0c-387b-4215-b490-5edd972623e9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:13:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 01:13:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 79f56a03-a0d2-400b-aa03-9d9321f018b2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:13:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4baf3a0c-387b-4215-b490-5edd972623e9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:13:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 01:13:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 79f56a03-a0d2-400b-aa03-9d9321f018b2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:13:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4baf3a0c-387b-4215-b490-5edd972623e9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:13:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 01:13:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 79f56a03-a0d2-400b-aa03-9d9321f018b2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:13:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4baf3a0c-387b-4215-b490-5edd972623e9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:13:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 01:13:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 79f56a03-a0d2-400b-aa03-9d9321f018b2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:13:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4baf3a0c-387b-4215-b490-5edd972623e9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:13:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 01:13:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 79f56a03-a0d2-400b-aa03-9d9321f018b2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:13:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4baf3a0c-387b-4215-b490-5edd972623e9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:13:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 01:14:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 79f56a03-a0d2-400b-aa03-9d9321f018b2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:14:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4baf3a0c-387b-4215-b490-5edd972623e9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:14:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 01:14:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 79f56a03-a0d2-400b-aa03-9d9321f018b2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:14:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4baf3a0c-387b-4215-b490-5edd972623e9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:14:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 01:14:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 79f56a03-a0d2-400b-aa03-9d9321f018b2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:14:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4baf3a0c-387b-4215-b490-5edd972623e9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:14:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 01:14:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 79f56a03-a0d2-400b-aa03-9d9321f018b2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:14:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4baf3a0c-387b-4215-b490-5edd972623e9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:14:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 01:14:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 79f56a03-a0d2-400b-aa03-9d9321f018b2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:14:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4baf3a0c-387b-4215-b490-5edd972623e9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:14:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 01:14:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 79f56a03-a0d2-400b-aa03-9d9321f018b2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:14:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4baf3a0c-387b-4215-b490-5edd972623e9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:14:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 01:14:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 79f56a03-a0d2-400b-aa03-9d9321f018b2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:14:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4baf3a0c-387b-4215-b490-5edd972623e9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:14:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 01:14:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 79f56a03-a0d2-400b-aa03-9d9321f018b2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:14:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4baf3a0c-387b-4215-b490-5edd972623e9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:14:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 01:14:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 79f56a03-a0d2-400b-aa03-9d9321f018b2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:14:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4baf3a0c-387b-4215-b490-5edd972623e9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:14:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 01:14:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 79f56a03-a0d2-400b-aa03-9d9321f018b2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:14:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4baf3a0c-387b-4215-b490-5edd972623e9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:14:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 01:14:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 79f56a03-a0d2-400b-aa03-9d9321f018b2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:14:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4baf3a0c-387b-4215-b490-5edd972623e9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:14:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 01:14:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 79f56a03-a0d2-400b-aa03-9d9321f018b2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:14:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4baf3a0c-387b-4215-b490-5edd972623e9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:14:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 01:14:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 79f56a03-a0d2-400b-aa03-9d9321f018b2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:14:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4baf3a0c-387b-4215-b490-5edd972623e9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:14:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 01:14:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 79f56a03-a0d2-400b-aa03-9d9321f018b2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:14:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4baf3a0c-387b-4215-b490-5edd972623e9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:14:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 01:14:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 79f56a03-a0d2-400b-aa03-9d9321f018b2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:14:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4baf3a0c-387b-4215-b490-5edd972623e9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:14:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 01:14:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 79f56a03-a0d2-400b-aa03-9d9321f018b2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:14:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4baf3a0c-387b-4215-b490-5edd972623e9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:14:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 01:14:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 79f56a03-a0d2-400b-aa03-9d9321f018b2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:14:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4baf3a0c-387b-4215-b490-5edd972623e9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:14:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 01:14:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 79f56a03-a0d2-400b-aa03-9d9321f018b2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:14:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4baf3a0c-387b-4215-b490-5edd972623e9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:14:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 01:14:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 79f56a03-a0d2-400b-aa03-9d9321f018b2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:14:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4baf3a0c-387b-4215-b490-5edd972623e9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:14:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 01:15:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 79f56a03-a0d2-400b-aa03-9d9321f018b2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:15:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4baf3a0c-387b-4215-b490-5edd972623e9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:15:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 01:15:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 79f56a03-a0d2-400b-aa03-9d9321f018b2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:15:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4baf3a0c-387b-4215-b490-5edd972623e9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:15:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 01:15:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 79f56a03-a0d2-400b-aa03-9d9321f018b2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:15:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4baf3a0c-387b-4215-b490-5edd972623e9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:15:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 01:15:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 79f56a03-a0d2-400b-aa03-9d9321f018b2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:15:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4baf3a0c-387b-4215-b490-5edd972623e9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:15:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 01:15:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 79f56a03-a0d2-400b-aa03-9d9321f018b2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:15:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4baf3a0c-387b-4215-b490-5edd972623e9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:15:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 01:15:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 79f56a03-a0d2-400b-aa03-9d9321f018b2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:15:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4baf3a0c-387b-4215-b490-5edd972623e9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:15:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 01:15:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 79f56a03-a0d2-400b-aa03-9d9321f018b2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:15:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4baf3a0c-387b-4215-b490-5edd972623e9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:15:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 01:15:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 79f56a03-a0d2-400b-aa03-9d9321f018b2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:15:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4baf3a0c-387b-4215-b490-5edd972623e9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:15:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 01:15:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 79f56a03-a0d2-400b-aa03-9d9321f018b2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:15:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4baf3a0c-387b-4215-b490-5edd972623e9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:15:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 01:15:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 79f56a03-a0d2-400b-aa03-9d9321f018b2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:15:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4baf3a0c-387b-4215-b490-5edd972623e9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:15:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 01:15:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 79f56a03-a0d2-400b-aa03-9d9321f018b2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:15:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4baf3a0c-387b-4215-b490-5edd972623e9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:15:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 01:15:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 79f56a03-a0d2-400b-aa03-9d9321f018b2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:15:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4baf3a0c-387b-4215-b490-5edd972623e9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:15:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 01:15:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 79f56a03-a0d2-400b-aa03-9d9321f018b2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:15:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4baf3a0c-387b-4215-b490-5edd972623e9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:15:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 01:15:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 79f56a03-a0d2-400b-aa03-9d9321f018b2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:15:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4baf3a0c-387b-4215-b490-5edd972623e9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:15:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 01:15:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 79f56a03-a0d2-400b-aa03-9d9321f018b2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:15:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4baf3a0c-387b-4215-b490-5edd972623e9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:15:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 01:15:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 79f56a03-a0d2-400b-aa03-9d9321f018b2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:15:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4baf3a0c-387b-4215-b490-5edd972623e9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:15:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 01:15:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 79f56a03-a0d2-400b-aa03-9d9321f018b2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:15:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4baf3a0c-387b-4215-b490-5edd972623e9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:15:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 01:15:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 79f56a03-a0d2-400b-aa03-9d9321f018b2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:15:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4baf3a0c-387b-4215-b490-5edd972623e9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:15:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 01:15:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 79f56a03-a0d2-400b-aa03-9d9321f018b2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:15:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4baf3a0c-387b-4215-b490-5edd972623e9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:15:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 01:15:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 79f56a03-a0d2-400b-aa03-9d9321f018b2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:15:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4baf3a0c-387b-4215-b490-5edd972623e9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:15:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 01:16:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 79f56a03-a0d2-400b-aa03-9d9321f018b2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:16:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4baf3a0c-387b-4215-b490-5edd972623e9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:16:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 01:16:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 79f56a03-a0d2-400b-aa03-9d9321f018b2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:16:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4baf3a0c-387b-4215-b490-5edd972623e9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:16:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 01:16:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 79f56a03-a0d2-400b-aa03-9d9321f018b2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:16:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4baf3a0c-387b-4215-b490-5edd972623e9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:16:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 01:16:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 79f56a03-a0d2-400b-aa03-9d9321f018b2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:16:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4baf3a0c-387b-4215-b490-5edd972623e9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:16:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 01:16:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 79f56a03-a0d2-400b-aa03-9d9321f018b2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:16:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4baf3a0c-387b-4215-b490-5edd972623e9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:16:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 01:16:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 79f56a03-a0d2-400b-aa03-9d9321f018b2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:16:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4baf3a0c-387b-4215-b490-5edd972623e9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:16:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 01:16:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 79f56a03-a0d2-400b-aa03-9d9321f018b2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:16:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4baf3a0c-387b-4215-b490-5edd972623e9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:16:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 01:16:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 79f56a03-a0d2-400b-aa03-9d9321f018b2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:16:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4baf3a0c-387b-4215-b490-5edd972623e9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:16:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 01:16:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 79f56a03-a0d2-400b-aa03-9d9321f018b2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:16:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4baf3a0c-387b-4215-b490-5edd972623e9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:16:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 01:16:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 79f56a03-a0d2-400b-aa03-9d9321f018b2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:16:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4baf3a0c-387b-4215-b490-5edd972623e9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:16:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 01:16:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 79f56a03-a0d2-400b-aa03-9d9321f018b2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:16:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4baf3a0c-387b-4215-b490-5edd972623e9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:16:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 01:16:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 79f56a03-a0d2-400b-aa03-9d9321f018b2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:16:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4baf3a0c-387b-4215-b490-5edd972623e9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:16:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 01:16:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 79f56a03-a0d2-400b-aa03-9d9321f018b2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:16:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4baf3a0c-387b-4215-b490-5edd972623e9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:16:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 01:16:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 79f56a03-a0d2-400b-aa03-9d9321f018b2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:16:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4baf3a0c-387b-4215-b490-5edd972623e9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:16:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 01:16:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 79f56a03-a0d2-400b-aa03-9d9321f018b2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:16:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4baf3a0c-387b-4215-b490-5edd972623e9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:16:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 01:16:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 79f56a03-a0d2-400b-aa03-9d9321f018b2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:16:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4baf3a0c-387b-4215-b490-5edd972623e9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:16:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 01:16:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 79f56a03-a0d2-400b-aa03-9d9321f018b2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:16:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4baf3a0c-387b-4215-b490-5edd972623e9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:16:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 01:16:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 79f56a03-a0d2-400b-aa03-9d9321f018b2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:16:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4baf3a0c-387b-4215-b490-5edd972623e9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:16:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 01:16:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 79f56a03-a0d2-400b-aa03-9d9321f018b2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:16:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4baf3a0c-387b-4215-b490-5edd972623e9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:16:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 01:16:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 79f56a03-a0d2-400b-aa03-9d9321f018b2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:16:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4baf3a0c-387b-4215-b490-5edd972623e9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:16:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 01:17:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 79f56a03-a0d2-400b-aa03-9d9321f018b2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:17:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4baf3a0c-387b-4215-b490-5edd972623e9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:17:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 01:17:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 79f56a03-a0d2-400b-aa03-9d9321f018b2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:17:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4baf3a0c-387b-4215-b490-5edd972623e9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:17:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 01:17:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 79f56a03-a0d2-400b-aa03-9d9321f018b2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:17:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4baf3a0c-387b-4215-b490-5edd972623e9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:17:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 01:17:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 79f56a03-a0d2-400b-aa03-9d9321f018b2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:17:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4baf3a0c-387b-4215-b490-5edd972623e9 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:17:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 01:17:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 79f56a03-a0d2-400b-aa03-9d9321f018b2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:17:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4baf3a0c-387b-4215-b490-5edd972623e9 is in state SUCCESS\u001b[0m",
                                "",
                                "",
                                "PLAY [Group hosts based on configuration] **************************************",
                                "",
                                "TASK [Group hosts based on OpenStack release] **********************************",
                                "Tuesday 24 February 2026  01:07:57 +0000 (0:00:00.344)       0:00:00.344 ****** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Group hosts based on Kolla action] ***************************************",
                                "Tuesday 24 February 2026  01:07:58 +0000 (0:00:01.277)       0:00:01.622 ****** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Group hosts based on enabled services] ***********************************",
                                "Tuesday 24 February 2026  01:07:59 +0000 (0:00:00.952)       0:00:02.574 ****** ",
                                "\u001b[0;33mchanged: [testbed-manager] => (item=enable_nova_True)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=enable_nova_True)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=enable_nova_True)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=enable_nova_True)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=enable_nova_True)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=enable_nova_True)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=enable_nova_True)\u001b[0m",
                                "",
                                "PLAY [Bootstrap nova API databases] ********************************************",
                                "",
                                "TASK [Bootstrap deploy] ********************************************************",
                                "Tuesday 24 February 2026  01:08:00 +0000 (0:00:00.931)       0:00:03.506 ****** ",
                                "\u001b[0;36mincluded: nova for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [nova : Creating Nova databases] ******************************************",
                                "Tuesday 24 February 2026  01:08:01 +0000 (0:00:01.042)       0:00:04.549 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=nova_cell0)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=nova_api)\u001b[0m",
                                "",
                                "TASK [nova : Creating Nova databases user and setting permissions] *************",
                                "Tuesday 24 February 2026  01:08:06 +0000 (0:00:04.784)       0:00:09.333 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=None)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=None)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [nova : Ensuring config directories exist] ********************************",
                                "Tuesday 24 February 2026  01:08:11 +0000 (0:00:04.789)       0:00:14.123 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [nova : Copying over config.json files for nova-api-bootstrap] ************",
                                "Tuesday 24 February 2026  01:08:12 +0000 (0:00:01.198)       0:00:15.322 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [nova : Copying over nova.conf for nova-api-bootstrap] ********************",
                                "Tuesday 24 February 2026  01:08:14 +0000 (0:00:01.801)       0:00:17.123 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [nova : include_tasks] ****************************************************",
                                "Tuesday 24 February 2026  01:08:18 +0000 (0:00:04.441)       0:00:21.564 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [nova : Running Nova API bootstrap container] *****************************",
                                "Tuesday 24 February 2026  01:08:19 +0000 (0:00:00.463)       0:00:22.027 ****** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [nova : Create cell0 mappings] ********************************************",
                                "Tuesday 24 February 2026  01:08:54 +0000 (0:00:35.166)       0:00:57.194 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [nova-cell : Get a list of existing cells] ********************************",
                                "Tuesday 24 February 2026  01:09:11 +0000 (0:00:17.691)       0:01:14.886 ****** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [nova-cell : Extract current cell settings from list] *********************",
                                "Tuesday 24 February 2026  01:09:26 +0000 (0:00:14.691)       0:01:29.577 ****** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [nova : Update cell0 mappings] ********************************************",
                                "Tuesday 24 February 2026  01:09:27 +0000 (0:00:00.979)       0:01:30.557 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [nova : include_tasks] ****************************************************",
                                "Tuesday 24 February 2026  01:09:27 +0000 (0:00:00.422)       0:01:30.980 ****** ",
                                "\u001b[0;36mincluded: /ansible/roles/nova/tasks/bootstrap_service.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [nova : Running Nova API bootstrap container] *****************************",
                                "Tuesday 24 February 2026  01:09:28 +0000 (0:00:00.543)       0:01:31.523 ****** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [Bootstrap upgrade] *******************************************************",
                                "Tuesday 24 February 2026  01:09:51 +0000 (0:00:23.084)       0:01:54.607 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "PLAY [Bootstrap nova cell databases] *******************************************",
                                "",
                                "TASK [Bootstrap deploy] ********************************************************",
                                "Tuesday 24 February 2026  01:09:52 +0000 (0:00:00.516)       0:01:55.123 ****** ",
                                "\u001b[0;36mincluded: nova-cell for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [nova-cell : Creating Nova cell database] *********************************",
                                "Tuesday 24 February 2026  01:09:53 +0000 (0:00:01.123)       0:01:56.247 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [nova-cell : Creating Nova cell database user and setting permissions] ****",
                                "Tuesday 24 February 2026  01:09:55 +0000 (0:00:02.466)       0:01:58.713 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [service-rabbitmq : nova | Ensure RabbitMQ vhosts exist] ******************",
                                "Tuesday 24 February 2026  01:09:58 +0000 (0:00:02.859)       0:02:01.573 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [service-rabbitmq : nova | Ensure RabbitMQ users exist] *******************",
                                "Tuesday 24 February 2026  01:09:59 +0000 (0:00:00.542)       0:02:02.115 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=None) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=None) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0] => (item=None)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0 -> {{ service_rabbitmq_delegate_host }}]\u001b[0m",
                                "",
                                "TASK [service-rabbitmq : nova | Ensure RabbitMQ vhosts exist] ******************",
                                "Tuesday 24 February 2026  01:10:07 +0000 (0:00:08.820)       0:02:10.935 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [service-rabbitmq : nova | Ensure RabbitMQ users exist] *******************",
                                "Tuesday 24 February 2026  01:10:08 +0000 (0:00:00.397)       0:02:11.333 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=None) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=None) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=None) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [nova-cell : Ensuring config directories exist] ***************************",
                                "Tuesday 24 February 2026  01:10:08 +0000 (0:00:00.612)       0:02:11.946 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [nova-cell : Copying over config.json files for nova-cell-bootstrap] ******",
                                "Tuesday 24 February 2026  01:10:09 +0000 (0:00:00.714)       0:02:12.660 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [nova-cell : Copying over nova.conf for nova-cell-bootstrap] **************",
                                "Tuesday 24 February 2026  01:10:10 +0000 (0:00:01.001)       0:02:13.662 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [nova-cell : Running Nova cell bootstrap container] ***********************",
                                "Tuesday 24 February 2026  01:10:13 +0000 (0:00:02.410)       0:02:16.073 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [nova-cell : Get a list of existing cells] ********************************",
                                "Tuesday 24 February 2026  01:10:35 +0000 (0:00:22.161)       0:02:38.234 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [nova-cell : Extract current cell settings from list] *********************",
                                "Tuesday 24 February 2026  01:10:48 +0000 (0:00:13.546)       0:02:51.780 ****** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [nova-cell : Create cell] *************************************************",
                                "Tuesday 24 February 2026  01:10:49 +0000 (0:00:00.963)       0:02:52.743 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [nova-cell : Update cell] *************************************************",
                                "Tuesday 24 February 2026  01:11:04 +0000 (0:00:14.398)       0:03:07.141 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [Bootstrap upgrade] *******************************************************",
                                "Tuesday 24 February 2026  01:11:05 +0000 (0:00:01.307)       0:03:08.449 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "PLAY [Apply role nova] *********************************************************",
                                "",
                                "TASK [nova : include_tasks] ****************************************************",
                                "Tuesday 24 February 2026  01:11:06 +0000 (0:00:00.691)       0:03:09.141 ****** ",
                                "\u001b[0;36mincluded: /ansible/roles/nova/tasks/deploy.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [service-ks-register : nova | Creating services] **************************",
                                "Tuesday 24 February 2026  01:11:06 +0000 (0:00:00.637)       0:03:09.778 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=nova_legacy (compute_legacy)) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=nova (compute))\u001b[0m",
                                "",
                                "TASK [service-ks-register : nova | Creating endpoints] *************************",
                                "Tuesday 24 February 2026  01:11:10 +0000 (0:00:03.678)       0:03:13.457 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=nova_legacy -> https://api-int.testbed.osism.xyz:8774/v2/%(tenant_id)s -> internal) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=nova_legacy -> https://api.testbed.osism.xyz:8774/v2/%(tenant_id)s -> public) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=nova -> https://api-int.testbed.osism.xyz:8774/v2.1 -> internal)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=nova -> https://api.testbed.osism.xyz:8774/v2.1 -> public)\u001b[0m",
                                "",
                                "TASK [service-ks-register : nova | Creating projects] **************************",
                                "Tuesday 24 February 2026  01:11:17 +0000 (0:00:07.286)       0:03:20.743 ****** ",
                                "\u001b[0;32mok: [testbed-node-0] => (item=service)\u001b[0m",
                                "",
                                "TASK [service-ks-register : nova | Creating users] *****************************",
                                "Tuesday 24 February 2026  01:11:21 +0000 (0:00:03.290)       0:03:24.034 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=nova -> service)\u001b[0m",
                                "\u001b[1;35m[WARNING]: Module did not set no_log for update_password\u001b[0m",
                                "",
                                "TASK [service-ks-register : nova | Creating roles] *****************************",
                                "Tuesday 24 February 2026  01:11:24 +0000 (0:00:03.880)       0:03:27.915 ****** ",
                                "\u001b[0;32mok: [testbed-node-0] => (item=admin)\u001b[0m",
                                "",
                                "TASK [service-ks-register : nova | Granting user roles] ************************",
                                "Tuesday 24 February 2026  01:11:28 +0000 (0:00:03.797)       0:03:31.712 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=nova -> service -> admin)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=nova -> service -> service)\u001b[0m",
                                "",
                                "TASK [nova : Ensuring config directories exist] ********************************",
                                "Tuesday 24 February 2026  01:11:37 +0000 (0:00:08.533)       0:03:40.246 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/nova-api:2024.2', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8774 '], 'timeout': '30'}, 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/nova-api:2024.2', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8774 '], 'timeout': '30'}, 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/nova-api:2024.2', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8774 '], 'timeout': '30'}, 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/nova-scheduler:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/nova-scheduler:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/nova-scheduler:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}})\u001b[0m",
                                "",
                                "TASK [nova : Check if policies shall be overwritten] ***************************",
                                "Tuesday 24 February 2026  01:11:38 +0000 (0:00:01.522)       0:03:41.768 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [nova : Set nova policy file] *********************************************",
                                "Tuesday 24 February 2026  01:11:38 +0000 (0:00:00.135)       0:03:41.904 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [nova : Check for vendordata file] ****************************************",
                                "Tuesday 24 February 2026  01:11:39 +0000 (0:00:00.338)       0:03:42.242 ****** ",
                                "\u001b[0;32mok: [testbed-node-0 -> localhost]\u001b[0m",
                                "",
                                "TASK [nova : Set vendordata file path] *****************************************",
                                "Tuesday 24 February 2026  01:11:40 +0000 (0:00:01.181)       0:03:43.424 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [nova : include_tasks] ****************************************************",
                                "Tuesday 24 February 2026  01:11:40 +0000 (0:00:00.352)       0:03:43.776 ****** ",
                                "\u001b[0;36mincluded: /ansible/roles/nova/tasks/copy-certs.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [service-cert-copy : nova | Copying over extra CA certificates] ***********",
                                "Tuesday 24 February 2026  01:11:41 +0000 (0:00:00.613)       0:03:44.390 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/nova-api:2024.2', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8774 '], 'timeout': '30'}, 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/nova-api:2024.2', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8774 '], 'timeout': '30'}, 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/nova-api:2024.2', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8774 '], 'timeout': '30'}, 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/nova-scheduler:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/nova-scheduler:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/nova-scheduler:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}})\u001b[0m",
                                "",
                                "TASK [service-cert-copy : nova | Copying over backend internal TLS certificate] ***",
                                "Tuesday 24 February 2026  01:11:44 +0000 (0:00:03.219)       0:03:47.609 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/nova-api:2024.2', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8774 '], 'timeout': '30'}, 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/nova-scheduler:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/nova-api:2024.2', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8774 '], 'timeout': '30'}, 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/nova-scheduler:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/nova-api:2024.2', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8774 '], 'timeout': '30'}, 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/nova-scheduler:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [service-cert-copy : nova | Copying over backend internal TLS key] ********",
                                "Tuesday 24 February 2026  01:11:45 +0000 (0:00:00.685)       0:03:48.295 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/nova-api:2024.2', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8774 '], 'timeout': '30'}, 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/nova-scheduler:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/nova-api:2024.2', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8774 '], 'timeout': '30'}, 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/nova-scheduler:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/nova-api:2024.2', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8774 '], 'timeout': '30'}, 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/nova-scheduler:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [nova : Copying over config.json files for services] **********************",
                                "Tuesday 24 February 2026  01:11:46 +0000 (0:00:00.966)       0:03:49.261 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/nova-api:2024.2', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8774 '], 'timeout': '30'}, 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/nova-api:2024.2', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8774 '], 'timeout': '30'}, 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/nova-api:2024.2', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8774 '], 'timeout': '30'}, 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/nova-scheduler:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/nova-scheduler:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/nova-scheduler:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}})\u001b[0m",
                                "",
                                "TASK [nova : Copying over nova.conf] *******************************************",
                                "Tuesday 24 February 2026  01:11:49 +0000 (0:00:02.883)       0:03:52.145 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/nova-api:2024.2', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8774 '], 'timeout': '30'}, 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/nova-api:2024.2', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8774 '], 'timeout': '30'}, 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/nova-api:2024.2', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8774 '], 'timeout': '30'}, 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/nova-scheduler:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/nova-scheduler:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/nova-scheduler:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}})\u001b[0m",
                                "",
                                "TASK [nova : Copying over existing policy file] ********************************",
                                "Tuesday 24 February 2026  01:11:55 +0000 (0:00:06.354)       0:03:58.499 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/nova-api:2024.2', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8774 '], 'timeout': '30'}, 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/nova-scheduler:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/nova-api:2024.2', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8774 '], 'timeout': '30'}, 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/nova-scheduler:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/nova-api:2024.2', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8774 '], 'timeout': '30'}, 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/nova-scheduler:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [nova : Copying over nova-api-wsgi.conf] **********************************",
                                "Tuesday 24 February 2026  01:11:56 +0000 (0:00:00.635)       0:03:59.134 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [nova : Copying over vendordata file] *************************************",
                                "Tuesday 24 February 2026  01:11:57 +0000 (0:00:01.659)       0:04:00.794 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [nova : Check nova containers] ********************************************",
                                "Tuesday 24 February 2026  01:11:58 +0000 (0:00:00.327)       0:04:01.121 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/nova-api:2024.2', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:8774 '], 'timeout': '30'}, 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/nova-api:2024.2', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:8774 '], 'timeout': '30'}, 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'nova-api', 'value': {'container_name': 'nova_api', 'group': 'nova-api', 'image': 'registry.osism.tech/kolla/nova-api:2024.2', 'enabled': True, 'privileged': True, 'volumes': ['/etc/kolla/nova-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:8774 '], 'timeout': '30'}, 'haproxy': {'nova_api': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_api_external': {'enabled': True, 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8774', 'listen_port': '8774', 'tls_backend': 'no'}, 'nova_metadata': {'enabled': True, 'mode': 'http', 'external': False, 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}, 'nova_metadata_external': {'enabled': 'no', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '8775', 'listen_port': '8775', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/nova-scheduler:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/nova-scheduler:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'nova-scheduler', 'value': {'container_name': 'nova_scheduler', 'group': 'nova-scheduler', 'image': 'registry.osism.tech/kolla/nova-scheduler:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-scheduler/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-scheduler 5672'], 'timeout': '30'}}})\u001b[0m",
                                "",
                                "TASK [nova : Flush handlers] ***************************************************",
                                "Tuesday 24 February 2026  01:12:00 +0000 (0:00:01.987)       0:04:03.108 ****** ",
                                "",
                                "TASK [nova : Flush handlers] ***************************************************",
                                "Tuesday 24 February 2026  01:12:00 +0000 (0:00:00.134)       0:04:03.243 ****** ",
                                "",
                                "TASK [nova : Flush handlers] ***************************************************",
                                "Tuesday 24 February 2026  01:12:00 +0000 (0:00:00.123)       0:04:03.366 ****** ",
                                "",
                                "RUNNING HANDLER [nova : Restart nova-scheduler container] **********************",
                                "Tuesday 24 February 2026  01:12:00 +0000 (0:00:00.124)       0:04:03.491 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [nova : Restart nova-api container] ****************************",
                                "Tuesday 24 February 2026  01:12:23 +0000 (0:00:22.540)       0:04:26.032 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "",
                                "PLAY [Apply role nova-cell] ****************************************************",
                                "",
                                "TASK [nova-cell : include_tasks] ***********************************************",
                                "Tuesday 24 February 2026  01:12:29 +0000 (0:00:06.843)       0:04:32.875 ****** ",
                                "\u001b[0;36mincluded: /ansible/roles/nova-cell/tasks/deploy.yml for testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [nova-cell : include_tasks] ***********************************************",
                                "Tuesday 24 February 2026  01:12:31 +0000 (0:00:01.427)       0:04:34.303 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [Load and persist br_netfilter module] ************************************",
                                "Tuesday 24 February 2026  01:12:31 +0000 (0:00:00.686)       0:04:34.989 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mincluded: module-load for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m",
                                "",
                                "TASK [module-load : Load modules] **********************************************",
                                "Tuesday 24 February 2026  01:12:33 +0000 (0:00:01.138)       0:04:36.128 ****** ",
                                "\u001b[0;32mok: [testbed-node-3] => (item=br_netfilter)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item=br_netfilter)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item=br_netfilter)\u001b[0m",
                                "",
                                "TASK [module-load : Persist modules via modules-load.d] ************************",
                                "Tuesday 24 February 2026  01:12:33 +0000 (0:00:00.751)       0:04:36.879 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=br_netfilter)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=br_netfilter)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=br_netfilter)\u001b[0m",
                                "",
                                "TASK [module-load : Drop module persistence] ***********************************",
                                "Tuesday 24 February 2026  01:12:35 +0000 (0:00:01.655)       0:04:38.534 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=br_netfilter) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=br_netfilter) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=br_netfilter) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [nova-cell : Enable bridge-nf-call sysctl variables] **********************",
                                "Tuesday 24 February 2026  01:12:36 +0000 (0:00:00.628)       0:04:39.163 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=net.bridge.bridge-nf-call-iptables) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=net.bridge.bridge-nf-call-ip6tables) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=net.bridge.bridge-nf-call-iptables) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=net.bridge.bridge-nf-call-ip6tables) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=net.bridge.bridge-nf-call-iptables)\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=net.bridge.bridge-nf-call-iptables)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=net.bridge.bridge-nf-call-iptables)\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=net.bridge.bridge-nf-call-iptables) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=net.bridge.bridge-nf-call-ip6tables) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=net.bridge.bridge-nf-call-ip6tables)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=net.bridge.bridge-nf-call-ip6tables)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=net.bridge.bridge-nf-call-ip6tables)\u001b[0m",
                                "",
                                "TASK [nova-cell : Install udev kolla kvm rules] ********************************",
                                "Tuesday 24 February 2026  01:12:37 +0000 (0:00:01.306)       0:04:40.470 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [nova-cell : Mask qemu-kvm service] ***************************************",
                                "Tuesday 24 February 2026  01:12:38 +0000 (0:00:01.329)       0:04:41.799 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [nova-cell : Ensuring config directories exist] ***************************",
                                "Tuesday 24 February 2026  01:12:40 +0000 (0:00:02.076)       0:04:43.876 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'nova-libvirt', 'value': {'container_name': 'nova_libvirt', 'group': 'compute', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-libvirt:2024.2', 'pid_mode': 'host', 'cgroupns_mode': 'host', 'privileged': True, 'volumes': ['/etc/kolla/nova-libvirt/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', '', '/sys/fs/cgroup:/sys/fs/cgroup', 'kolla_logs:/var/log/kolla/', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', 'nova_libvirt_qemu:/etc/libvirt/qemu', ''], 'dimensions': {'ulimits': {'memlock': {'soft': 67108864, 'hard': 67108864}}}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'virsh version --daemon'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'nova-libvirt', 'value': {'container_name': 'nova_libvirt', 'group': 'compute', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-libvirt:2024.2', 'pid_mode': 'host', 'cgroupns_mode': 'host', 'privileged': True, 'volumes': ['/etc/kolla/nova-libvirt/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', '', '/sys/fs/cgroup:/sys/fs/cgroup', 'kolla_logs:/var/log/kolla/', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', 'nova_libvirt_qemu:/etc/libvirt/qemu', ''], 'dimensions': {'ulimits': {'memlock': {'soft': 67108864, 'hard': 67108864}}}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'virsh version --daemon'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'nova-ssh', 'value': {'container_name': 'nova_ssh', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-ssh:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla', 'nova_compute:/var/lib/nova', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8022'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'nova-libvirt', 'value': {'container_name': 'nova_libvirt', 'group': 'compute', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-libvirt:2024.2', 'pid_mode': 'host', 'cgroupns_mode': 'host', 'privileged': True, 'volumes': ['/etc/kolla/nova-libvirt/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', '', '/sys/fs/cgroup:/sys/fs/cgroup', 'kolla_logs:/var/log/kolla/', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', 'nova_libvirt_qemu:/etc/libvirt/qemu', ''], 'dimensions': {'ulimits': {'memlock': {'soft': 67108864, 'hard': 67108864}}}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'virsh version --daemon'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'nova-novncproxy', 'value': {'container_name': 'nova_novncproxy', 'group': 'nova-novncproxy', 'image': 'registry.osism.tech/kolla/nova-novncproxy:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-novncproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:6080/vnc_lite.html'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'nova-ssh', 'value': {'container_name': 'nova_ssh', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-ssh:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla', 'nova_compute:/var/lib/nova', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8022'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'nova-ssh', 'value': {'container_name': 'nova_ssh', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-ssh:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla', 'nova_compute:/var/lib/nova', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8022'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'nova-compute', 'value': {'container_name': 'nova_compute', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-compute:2024.2', 'environment': {'LIBGUESTFS_BACKEND': 'direct'}, 'privileged': True, 'enabled': True, 'ipc_mode': 'host', 'volumes': ['/etc/kolla/nova-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', 'kolla_logs:/var/log/kolla/', '', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-compute 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'nova-novncproxy', 'value': {'container_name': 'nova_novncproxy', 'group': 'nova-novncproxy', 'image': 'registry.osism.tech/kolla/nova-novncproxy:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-novncproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:6080/vnc_lite.html'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'nova-novncproxy', 'value': {'container_name': 'nova_novncproxy', 'group': 'nova-novncproxy', 'image': 'registry.osism.tech/kolla/nova-novncproxy:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-novncproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:6080/vnc_lite.html'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'nova-compute', 'value': {'container_name': 'nova_compute', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-compute:2024.2', 'environment': {'LIBGUESTFS_BACKEND': 'direct'}, 'privileged': True, 'enabled': True, 'ipc_mode': 'host', 'volumes': ['/etc/kolla/nova-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', 'kolla_logs:/var/log/kolla/', '', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-compute 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'nova-conductor', 'value': {'container_name': 'nova_conductor', 'group': 'nova-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-conductor:2024.2', 'volumes': ['/etc/kolla/nova-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'nova-compute', 'value': {'container_name': 'nova_compute', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-compute:2024.2', 'environment': {'LIBGUESTFS_BACKEND': 'direct'}, 'privileged': True, 'enabled': True, 'ipc_mode': 'host', 'volumes': ['/etc/kolla/nova-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', 'kolla_logs:/var/log/kolla/', '', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-compute 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'nova-conductor', 'value': {'container_name': 'nova_conductor', 'group': 'nova-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-conductor:2024.2', 'volumes': ['/etc/kolla/nova-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'nova-conductor', 'value': {'container_name': 'nova_conductor', 'group': 'nova-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-conductor:2024.2', 'volumes': ['/etc/kolla/nova-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}})\u001b[0m",
                                "",
                                "TASK [nova-cell : include_tasks] ***********************************************",
                                "Tuesday 24 February 2026  01:12:43 +0000 (0:00:02.268)       0:04:46.144 ****** ",
                                "\u001b[0;36mincluded: /ansible/roles/nova-cell/tasks/copy-certs.yml for testbed-node-3, testbed-node-4, testbed-node-5, testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [service-cert-copy : nova | Copying over extra CA certificates] ***********",
                                "Tuesday 24 February 2026  01:12:44 +0000 (0:00:01.361)       0:04:47.505 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'nova-libvirt', 'value': {'container_name': 'nova_libvirt', 'group': 'compute', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-libvirt:2024.2', 'pid_mode': 'host', 'cgroupns_mode': 'host', 'privileged': True, 'volumes': ['/etc/kolla/nova-libvirt/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', '', '/sys/fs/cgroup:/sys/fs/cgroup', 'kolla_logs:/var/log/kolla/', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', 'nova_libvirt_qemu:/etc/libvirt/qemu', ''], 'dimensions': {'ulimits': {'memlock': {'soft': 67108864, 'hard': 67108864}}}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'virsh version --daemon'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'nova-libvirt', 'value': {'container_name': 'nova_libvirt', 'group': 'compute', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-libvirt:2024.2', 'pid_mode': 'host', 'cgroupns_mode': 'host', 'privileged': True, 'volumes': ['/etc/kolla/nova-libvirt/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', '', '/sys/fs/cgroup:/sys/fs/cgroup', 'kolla_logs:/var/log/kolla/', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', 'nova_libvirt_qemu:/etc/libvirt/qemu', ''], 'dimensions': {'ulimits': {'memlock': {'soft': 67108864, 'hard': 67108864}}}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'virsh version --daemon'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'nova-libvirt', 'value': {'container_name': 'nova_libvirt', 'group': 'compute', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-libvirt:2024.2', 'pid_mode': 'host', 'cgroupns_mode': 'host', 'privileged': True, 'volumes': ['/etc/kolla/nova-libvirt/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', '', '/sys/fs/cgroup:/sys/fs/cgroup', 'kolla_logs:/var/log/kolla/', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', 'nova_libvirt_qemu:/etc/libvirt/qemu', ''], 'dimensions': {'ulimits': {'memlock': {'soft': 67108864, 'hard': 67108864}}}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'virsh version --daemon'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'nova-novncproxy', 'value': {'container_name': 'nova_novncproxy', 'group': 'nova-novncproxy', 'image': 'registry.osism.tech/kolla/nova-novncproxy:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-novncproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:6080/vnc_lite.html'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'nova-ssh', 'value': {'container_name': 'nova_ssh', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-ssh:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla', 'nova_compute:/var/lib/nova', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8022'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'nova-novncproxy', 'value': {'container_name': 'nova_novncproxy', 'group': 'nova-novncproxy', 'image': 'registry.osism.tech/kolla/nova-novncproxy:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-novncproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:6080/vnc_lite.html'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'nova-novncproxy', 'value': {'container_name': 'nova_novncproxy', 'group': 'nova-novncproxy', 'image': 'registry.osism.tech/kolla/nova-novncproxy:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-novncproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:6080/vnc_lite.html'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'nova-ssh', 'value': {'container_name': 'nova_ssh', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-ssh:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla', 'nova_compute:/var/lib/nova', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8022'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'nova-ssh', 'value': {'container_name': 'nova_ssh', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-ssh:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla', 'nova_compute:/var/lib/nova', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8022'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'nova-conductor', 'value': {'container_name': 'nova_conductor', 'group': 'nova-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-conductor:2024.2', 'volumes': ['/etc/kolla/nova-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'nova-compute', 'value': {'container_name': 'nova_compute', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-compute:2024.2', 'environment': {'LIBGUESTFS_BACKEND': 'direct'}, 'privileged': True, 'enabled': True, 'ipc_mode': 'host', 'volumes': ['/etc/kolla/nova-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', 'kolla_logs:/var/log/kolla/', '', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-compute 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'nova-conductor', 'value': {'container_name': 'nova_conductor', 'group': 'nova-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-conductor:2024.2', 'volumes': ['/etc/kolla/nova-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'nova-conductor', 'value': {'container_name': 'nova_conductor', 'group': 'nova-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-conductor:2024.2', 'volumes': ['/etc/kolla/nova-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'nova-compute', 'value': {'container_name': 'nova_compute', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-compute:2024.2', 'environment': {'LIBGUESTFS_BACKEND': 'direct'}, 'privileged': True, 'enabled': True, 'ipc_mode': 'host', 'volumes': ['/etc/kolla/nova-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', 'kolla_logs:/var/log/kolla/', '', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-compute 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'nova-compute', 'value': {'container_name': 'nova_compute', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-compute:2024.2', 'environment': {'LIBGUESTFS_BACKEND': 'direct'}, 'privileged': True, 'enabled': True, 'ipc_mode': 'host', 'volumes': ['/etc/kolla/nova-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', 'kolla_logs:/var/log/kolla/', '', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-compute 5672'], 'timeout': '30'}}})\u001b[0m",
                                "",
                                "TASK [service-cert-copy : nova | Copying over backend internal TLS certificate] ***",
                                "Tuesday 24 February 2026  01:12:48 +0000 (0:00:03.878)       0:04:51.383 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'nova-libvirt', 'value': {'container_name': 'nova_libvirt', 'group': 'compute', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-libvirt:2024.2', 'pid_mode': 'host', 'cgroupns_mode': 'host', 'privileged': True, 'volumes': ['/etc/kolla/nova-libvirt/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', '', '/sys/fs/cgroup:/sys/fs/cgroup', 'kolla_logs:/var/log/kolla/', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', 'nova_libvirt_qemu:/etc/libvirt/qemu', ''], 'dimensions': {'ulimits': {'memlock': {'soft': 67108864, 'hard': 67108864}}}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'virsh version --daemon'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'nova-ssh', 'value': {'container_name': 'nova_ssh', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-ssh:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla', 'nova_compute:/var/lib/nova', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8022'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'nova-compute', 'value': {'container_name': 'nova_compute', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-compute:2024.2', 'environment': {'LIBGUESTFS_BACKEND': 'direct'}, 'privileged': True, 'enabled': True, 'ipc_mode': 'host', 'volumes': ['/etc/kolla/nova-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', 'kolla_logs:/var/log/kolla/', '', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-compute 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'nova-libvirt', 'value': {'container_name': 'nova_libvirt', 'group': 'compute', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-libvirt:2024.2', 'pid_mode': 'host', 'cgroupns_mode': 'host', 'privileged': True, 'volumes': ['/etc/kolla/nova-libvirt/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', '', '/sys/fs/cgroup:/sys/fs/cgroup', 'kolla_logs:/var/log/kolla/', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', 'nova_libvirt_qemu:/etc/libvirt/qemu', ''], 'dimensions': {'ulimits': {'memlock': {'soft': 67108864, 'hard': 67108864}}}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'virsh version --daemon'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'nova-ssh', 'value': {'container_name': 'nova_ssh', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-ssh:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla', 'nova_compute:/var/lib/nova', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8022'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'nova-compute', 'value': {'container_name': 'nova_compute', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-compute:2024.2', 'environment': {'LIBGUESTFS_BACKEND': 'direct'}, 'privileged': True, 'enabled': True, 'ipc_mode': 'host', 'volumes': ['/etc/kolla/nova-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', 'kolla_logs:/var/log/kolla/', '', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-compute 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'nova-libvirt', 'value': {'container_name': 'nova_libvirt', 'group': 'compute', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-libvirt:2024.2', 'pid_mode': 'host', 'cgroupns_mode': 'host', 'privileged': True, 'volumes': ['/etc/kolla/nova-libvirt/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', '', '/sys/fs/cgroup:/sys/fs/cgroup', 'kolla_logs:/var/log/kolla/', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', 'nova_libvirt_qemu:/etc/libvirt/qemu', ''], 'dimensions': {'ulimits': {'memlock': {'soft': 67108864, 'hard': 67108864}}}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'virsh version --daemon'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'nova-ssh', 'value': {'container_name': 'nova_ssh', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-ssh:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla', 'nova_compute:/var/lib/nova', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8022'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'nova-compute', 'value': {'container_name': 'nova_compute', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-compute:2024.2', 'environment': {'LIBGUESTFS_BACKEND': 'direct'}, 'privileged': True, 'enabled': True, 'ipc_mode': 'host', 'volumes': ['/etc/kolla/nova-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', 'kolla_logs:/var/log/kolla/', '', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-compute 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova-novncproxy', 'value': {'container_name': 'nova_novncproxy', 'group': 'nova-novncproxy', 'image': 'registry.osism.tech/kolla/nova-novncproxy:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-novncproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:6080/vnc_lite.html'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova-conductor', 'value': {'container_name': 'nova_conductor', 'group': 'nova-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-conductor:2024.2', 'volumes': ['/etc/kolla/nova-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova-novncproxy', 'value': {'container_name': 'nova_novncproxy', 'group': 'nova-novncproxy', 'image': 'registry.osism.tech/kolla/nova-novncproxy:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-novncproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:6080/vnc_lite.html'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova-conductor', 'value': {'container_name': 'nova_conductor', 'group': 'nova-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-conductor:2024.2', 'volumes': ['/etc/kolla/nova-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova-novncproxy', 'value': {'container_name': 'nova_novncproxy', 'group': 'nova-novncproxy', 'image': 'registry.osism.tech/kolla/nova-novncproxy:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-novncproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:6080/vnc_lite.html'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova-conductor', 'value': {'container_name': 'nova_conductor', 'group': 'nova-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-conductor:2024.2', 'volumes': ['/etc/kolla/nova-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [service-cert-copy : nova | Copying over backend internal TLS key] ********",
                                "Tuesday 24 February 2026  01:12:50 +0000 (0:00:01.752)       0:04:53.136 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'nova-libvirt', 'value': {'container_name': 'nova_libvirt', 'group': 'compute', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-libvirt:2024.2', 'pid_mode': 'host', 'cgroupns_mode': 'host', 'privileged': True, 'volumes': ['/etc/kolla/nova-libvirt/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', '', '/sys/fs/cgroup:/sys/fs/cgroup', 'kolla_logs:/var/log/kolla/', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', 'nova_libvirt_qemu:/etc/libvirt/qemu', ''], 'dimensions': {'ulimits': {'memlock': {'soft': 67108864, 'hard': 67108864}}}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'virsh version --daemon'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'nova-ssh', 'value': {'container_name': 'nova_ssh', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-ssh:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla', 'nova_compute:/var/lib/nova', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8022'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'nova-compute', 'value': {'container_name': 'nova_compute', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-compute:2024.2', 'environment': {'LIBGUESTFS_BACKEND': 'direct'}, 'privileged': True, 'enabled': True, 'ipc_mode': 'host', 'volumes': ['/etc/kolla/nova-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', 'kolla_logs:/var/log/kolla/', '', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-compute 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'nova-libvirt', 'value': {'container_name': 'nova_libvirt', 'group': 'compute', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-libvirt:2024.2', 'pid_mode': 'host', 'cgroupns_mode': 'host', 'privileged': True, 'volumes': ['/etc/kolla/nova-libvirt/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', '', '/sys/fs/cgroup:/sys/fs/cgroup', 'kolla_logs:/var/log/kolla/', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', 'nova_libvirt_qemu:/etc/libvirt/qemu', ''], 'dimensions': {'ulimits': {'memlock': {'soft': 67108864, 'hard': 67108864}}}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'virsh version --daemon'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'nova-ssh', 'value': {'container_name': 'nova_ssh', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-ssh:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla', 'nova_compute:/var/lib/nova', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8022'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'nova-compute', 'value': {'container_name': 'nova_compute', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-compute:2024.2', 'environment': {'LIBGUESTFS_BACKEND': 'direct'}, 'privileged': True, 'enabled': True, 'ipc_mode': 'host', 'volumes': ['/etc/kolla/nova-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', 'kolla_logs:/var/log/kolla/', '', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-compute 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'nova-libvirt', 'value': {'container_name': 'nova_libvirt', 'group': 'compute', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-libvirt:2024.2', 'pid_mode': 'host', 'cgroupns_mode': 'host', 'privileged': True, 'volumes': ['/etc/kolla/nova-libvirt/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', '', '/sys/fs/cgroup:/sys/fs/cgroup', 'kolla_logs:/var/log/kolla/', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', 'nova_libvirt_qemu:/etc/libvirt/qemu', ''], 'dimensions': {'ulimits': {'memlock': {'soft': 67108864, 'hard': 67108864}}}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'virsh version --daemon'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'nova-ssh', 'value': {'container_name': 'nova_ssh', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-ssh:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla', 'nova_compute:/var/lib/nova', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8022'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova-novncproxy', 'value': {'container_name': 'nova_novncproxy', 'group': 'nova-novncproxy', 'image': 'registry.osism.tech/kolla/nova-novncproxy:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-novncproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:6080/vnc_lite.html'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova-conductor', 'value': {'container_name': 'nova_conductor', 'group': 'nova-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-conductor:2024.2', 'volumes': ['/etc/kolla/nova-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'nova-compute', 'value': {'container_name': 'nova_compute', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-compute:2024.2', 'environment': {'LIBGUESTFS_BACKEND': 'direct'}, 'privileged': True, 'enabled': True, 'ipc_mode': 'host', 'volumes': ['/etc/kolla/nova-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', 'kolla_logs:/var/log/kolla/', '', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-compute 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova-novncproxy', 'value': {'container_name': 'nova_novncproxy', 'group': 'nova-novncproxy', 'image': 'registry.osism.tech/kolla/nova-novncproxy:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-novncproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:6080/vnc_lite.html'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova-conductor', 'value': {'container_name': 'nova_conductor', 'group': 'nova-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-conductor:2024.2', 'volumes': ['/etc/kolla/nova-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova-novncproxy', 'value': {'container_name': 'nova_novncproxy', 'group': 'nova-novncproxy', 'image': 'registry.osism.tech/kolla/nova-novncproxy:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-novncproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:6080/vnc_lite.html'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova-conductor', 'value': {'container_name': 'nova_conductor', 'group': 'nova-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-conductor:2024.2', 'volumes': ['/etc/kolla/nova-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [nova-cell : include_tasks] ***********************************************",
                                "Tuesday 24 February 2026  01:12:52 +0000 (0:00:02.331)       0:04:55.467 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/roles/nova-cell/tasks/external_ceph.yml for testbed-node-3, testbed-node-4, testbed-node-5\u001b[0m",
                                "",
                                "TASK [nova-cell : Check nova keyring file] *************************************",
                                "Tuesday 24 February 2026  01:12:53 +0000 (0:00:01.241)       0:04:56.709 ****** ",
                                "\u001b[0;32mok: [testbed-node-3 -> localhost]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4 -> localhost]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5 -> localhost]\u001b[0m",
                                "",
                                "TASK [nova-cell : Check cinder keyring file] ***********************************",
                                "Tuesday 24 February 2026  01:12:54 +0000 (0:00:01.242)       0:04:57.951 ****** ",
                                "\u001b[0;32mok: [testbed-node-3 -> localhost]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4 -> localhost]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5 -> localhost]\u001b[0m",
                                "",
                                "TASK [nova-cell : Extract nova key from file] **********************************",
                                "Tuesday 24 February 2026  01:12:56 +0000 (0:00:01.241)       0:04:59.192 ****** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [nova-cell : Extract cinder key from file] ********************************",
                                "Tuesday 24 February 2026  01:12:56 +0000 (0:00:00.555)       0:04:59.748 ****** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [nova-cell : Copy over ceph nova keyring file] ****************************",
                                "Tuesday 24 February 2026  01:12:57 +0000 (0:00:00.897)       0:05:00.645 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=nova-compute)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=nova-compute)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=nova-compute)\u001b[0m",
                                "",
                                "TASK [nova-cell : Copy over ceph cinder keyring file] **************************",
                                "Tuesday 24 February 2026  01:12:58 +0000 (0:00:01.291)       0:05:01.937 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=nova-compute)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=nova-compute)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=nova-compute)\u001b[0m",
                                "",
                                "TASK [nova-cell : Copy over ceph.conf] *****************************************",
                                "Tuesday 24 February 2026  01:13:00 +0000 (0:00:01.244)       0:05:03.182 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=nova-compute)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=nova-compute)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=nova-compute)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=nova-libvirt)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=nova-libvirt)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=nova-libvirt)\u001b[0m",
                                "",
                                "TASK [nova-cell : Ensure /etc/ceph directory exists (host libvirt)] ************",
                                "Tuesday 24 February 2026  01:13:04 +0000 (0:00:04.219)       0:05:07.401 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [nova-cell : Copy over ceph.conf (host libvirt)] **************************",
                                "Tuesday 24 February 2026  01:13:05 +0000 (0:00:00.669)       0:05:08.071 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [nova-cell : Ensuring libvirt secrets directory exists] *******************",
                                "Tuesday 24 February 2026  01:13:05 +0000 (0:00:00.372)       0:05:08.443 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [nova-cell : Pushing nova secret xml for libvirt] *************************",
                                "Tuesday 24 February 2026  01:13:06 +0000 (0:00:01.365)       0:05:09.809 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'uuid': '5a2bf0bf-e1ab-4a6a-bc32-404bb6ba91fd', 'name': 'client.nova secret', 'enabled': True})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'uuid': '5a2bf0bf-e1ab-4a6a-bc32-404bb6ba91fd', 'name': 'client.nova secret', 'enabled': True})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'uuid': '5a2bf0bf-e1ab-4a6a-bc32-404bb6ba91fd', 'name': 'client.nova secret', 'enabled': True})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'uuid': '63dd366f-e403-41f2-beff-dad9980a1637', 'name': 'client.cinder secret', 'enabled': 'yes'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'uuid': '63dd366f-e403-41f2-beff-dad9980a1637', 'name': 'client.cinder secret', 'enabled': 'yes'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'uuid': '63dd366f-e403-41f2-beff-dad9980a1637', 'name': 'client.cinder secret', 'enabled': 'yes'})\u001b[0m",
                                "",
                                "TASK [nova-cell : Pushing secrets key for libvirt] *****************************",
                                "Tuesday 24 February 2026  01:13:10 +0000 (0:00:03.628)       0:05:13.438 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=None)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=None)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=None)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item=None)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item=None)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item=None)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [nova-cell : Include tasks from qemu_wrapper.yml] *************************",
                                "Tuesday 24 February 2026  01:13:14 +0000 (0:00:03.975)       0:05:17.413 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/roles/nova-cell/tasks/qemu_wrapper.yml for testbed-node-3, testbed-node-5, testbed-node-4\u001b[0m",
                                "",
                                "TASK [nova-cell : Check qemu wrapper file] *************************************",
                                "Tuesday 24 February 2026  01:13:16 +0000 (0:00:01.687)       0:05:19.100 ****** ",
                                "\u001b[0;32mok: [testbed-node-3 -> localhost]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4 -> localhost]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5 -> localhost]\u001b[0m",
                                "",
                                "TASK [nova-cell : Copy qemu wrapper] *******************************************",
                                "Tuesday 24 February 2026  01:13:17 +0000 (0:00:01.379)       0:05:20.479 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [nova-cell : Check if policies shall be overwritten] **********************",
                                "Tuesday 24 February 2026  01:13:17 +0000 (0:00:00.340)       0:05:20.820 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [nova-cell : Set nova policy file] ****************************************",
                                "Tuesday 24 February 2026  01:13:17 +0000 (0:00:00.155)       0:05:20.976 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [nova-cell : Check for vendordata file] ***********************************",
                                "Tuesday 24 February 2026  01:13:18 +0000 (0:00:00.618)       0:05:21.594 ****** ",
                                "\u001b[0;32mok: [testbed-node-3 -> localhost]\u001b[0m",
                                "",
                                "TASK [nova-cell : Set vendordata file path] ************************************",
                                "Tuesday 24 February 2026  01:13:19 +0000 (0:00:01.121)       0:05:22.716 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [nova-cell : Copying over config.json files for services] *****************",
                                "Tuesday 24 February 2026  01:13:20 +0000 (0:00:00.714)       0:05:23.430 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'nova-libvirt', 'value': {'container_name': 'nova_libvirt', 'group': 'compute', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-libvirt:2024.2', 'pid_mode': 'host', 'cgroupns_mode': 'host', 'privileged': True, 'volumes': ['/etc/kolla/nova-libvirt/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', '', '/sys/fs/cgroup:/sys/fs/cgroup', 'kolla_logs:/var/log/kolla/', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', 'nova_libvirt_qemu:/etc/libvirt/qemu', ''], 'dimensions': {'ulimits': {'memlock': {'soft': 67108864, 'hard': 67108864}}}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'virsh version --daemon'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'nova-libvirt', 'value': {'container_name': 'nova_libvirt', 'group': 'compute', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-libvirt:2024.2', 'pid_mode': 'host', 'cgroupns_mode': 'host', 'privileged': True, 'volumes': ['/etc/kolla/nova-libvirt/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', '', '/sys/fs/cgroup:/sys/fs/cgroup', 'kolla_logs:/var/log/kolla/', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', 'nova_libvirt_qemu:/etc/libvirt/qemu', ''], 'dimensions': {'ulimits': {'memlock': {'soft': 67108864, 'hard': 67108864}}}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'virsh version --daemon'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'nova-libvirt', 'value': {'container_name': 'nova_libvirt', 'group': 'compute', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-libvirt:2024.2', 'pid_mode': 'host', 'cgroupns_mode': 'host', 'privileged': True, 'volumes': ['/etc/kolla/nova-libvirt/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', '', '/sys/fs/cgroup:/sys/fs/cgroup', 'kolla_logs:/var/log/kolla/', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', 'nova_libvirt_qemu:/etc/libvirt/qemu', ''], 'dimensions': {'ulimits': {'memlock': {'soft': 67108864, 'hard': 67108864}}}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'virsh version --daemon'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'nova-novncproxy', 'value': {'container_name': 'nova_novncproxy', 'group': 'nova-novncproxy', 'image': 'registry.osism.tech/kolla/nova-novncproxy:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-novncproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:6080/vnc_lite.html'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'nova-novncproxy', 'value': {'container_name': 'nova_novncproxy', 'group': 'nova-novncproxy', 'image': 'registry.osism.tech/kolla/nova-novncproxy:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-novncproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:6080/vnc_lite.html'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'nova-ssh', 'value': {'container_name': 'nova_ssh', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-ssh:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla', 'nova_compute:/var/lib/nova', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8022'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'nova-novncproxy', 'value': {'container_name': 'nova_novncproxy', 'group': 'nova-novncproxy', 'image': 'registry.osism.tech/kolla/nova-novncproxy:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-novncproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:6080/vnc_lite.html'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'nova-ssh', 'value': {'container_name': 'nova_ssh', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-ssh:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla', 'nova_compute:/var/lib/nova', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8022'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'nova-ssh', 'value': {'container_name': 'nova_ssh', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-ssh:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla', 'nova_compute:/var/lib/nova', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8022'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'nova-conductor', 'value': {'container_name': 'nova_conductor', 'group': 'nova-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-conductor:2024.2', 'volumes': ['/etc/kolla/nova-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'nova-conductor', 'value': {'container_name': 'nova_conductor', 'group': 'nova-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-conductor:2024.2', 'volumes': ['/etc/kolla/nova-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'nova-conductor', 'value': {'container_name': 'nova_conductor', 'group': 'nova-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-conductor:2024.2', 'volumes': ['/etc/kolla/nova-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'nova-compute', 'value': {'container_name': 'nova_compute', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-compute:2024.2', 'environment': {'LIBGUESTFS_BACKEND': 'direct'}, 'privileged': True, 'enabled': True, 'ipc_mode': 'host', 'volumes': ['/etc/kolla/nova-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', 'kolla_logs:/var/log/kolla/', '', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-compute 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'nova-compute', 'value': {'container_name': 'nova_compute', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-compute:2024.2', 'environment': {'LIBGUESTFS_BACKEND': 'direct'}, 'privileged': True, 'enabled': True, 'ipc_mode': 'host', 'volumes': ['/etc/kolla/nova-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', 'kolla_logs:/var/log/kolla/', '', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-compute 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'nova-compute', 'value': {'container_name': 'nova_compute', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-compute:2024.2', 'environment': {'LIBGUESTFS_BACKEND': 'direct'}, 'privileged': True, 'enabled': True, 'ipc_mode': 'host', 'volumes': ['/etc/kolla/nova-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', 'kolla_logs:/var/log/kolla/', '', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-compute 5672'], 'timeout': '30'}}})\u001b[0m",
                                "",
                                "TASK [nova-cell : Copying over nova.conf] **************************************",
                                "Tuesday 24 February 2026  01:13:24 +0000 (0:00:04.272)       0:05:27.702 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'nova-libvirt', 'value': {'container_name': 'nova_libvirt', 'group': 'compute', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-libvirt:2024.2', 'pid_mode': 'host', 'cgroupns_mode': 'host', 'privileged': True, 'volumes': ['/etc/kolla/nova-libvirt/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', '', '/sys/fs/cgroup:/sys/fs/cgroup', 'kolla_logs:/var/log/kolla/', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', 'nova_libvirt_qemu:/etc/libvirt/qemu', ''], 'dimensions': {'ulimits': {'memlock': {'soft': 67108864, 'hard': 67108864}}}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'virsh version --daemon'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'nova-ssh', 'value': {'container_name': 'nova_ssh', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-ssh:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla', 'nova_compute:/var/lib/nova', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8022'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'nova-libvirt', 'value': {'container_name': 'nova_libvirt', 'group': 'compute', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-libvirt:2024.2', 'pid_mode': 'host', 'cgroupns_mode': 'host', 'privileged': True, 'volumes': ['/etc/kolla/nova-libvirt/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', '', '/sys/fs/cgroup:/sys/fs/cgroup', 'kolla_logs:/var/log/kolla/', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', 'nova_libvirt_qemu:/etc/libvirt/qemu', ''], 'dimensions': {'ulimits': {'memlock': {'soft': 67108864, 'hard': 67108864}}}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'virsh version --daemon'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'nova-ssh', 'value': {'container_name': 'nova_ssh', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-ssh:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla', 'nova_compute:/var/lib/nova', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8022'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'nova-libvirt', 'value': {'container_name': 'nova_libvirt', 'group': 'compute', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-libvirt:2024.2', 'pid_mode': 'host', 'cgroupns_mode': 'host', 'privileged': True, 'volumes': ['/etc/kolla/nova-libvirt/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', '', '/sys/fs/cgroup:/sys/fs/cgroup', 'kolla_logs:/var/log/kolla/', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', 'nova_libvirt_qemu:/etc/libvirt/qemu', ''], 'dimensions': {'ulimits': {'memlock': {'soft': 67108864, 'hard': 67108864}}}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'virsh version --daemon'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'nova-ssh', 'value': {'container_name': 'nova_ssh', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-ssh:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla', 'nova_compute:/var/lib/nova', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8022'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'nova-compute', 'value': {'container_name': 'nova_compute', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-compute:2024.2', 'environment': {'LIBGUESTFS_BACKEND': 'direct'}, 'privileged': True, 'enabled': True, 'ipc_mode': 'host', 'volumes': ['/etc/kolla/nova-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', 'kolla_logs:/var/log/kolla/', '', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-compute 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'nova-novncproxy', 'value': {'container_name': 'nova_novncproxy', 'group': 'nova-novncproxy', 'image': 'registry.osism.tech/kolla/nova-novncproxy:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-novncproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:6080/vnc_lite.html'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'nova-compute', 'value': {'container_name': 'nova_compute', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-compute:2024.2', 'environment': {'LIBGUESTFS_BACKEND': 'direct'}, 'privileged': True, 'enabled': True, 'ipc_mode': 'host', 'volumes': ['/etc/kolla/nova-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', 'kolla_logs:/var/log/kolla/', '', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-compute 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'nova-compute', 'value': {'container_name': 'nova_compute', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-compute:2024.2', 'environment': {'LIBGUESTFS_BACKEND': 'direct'}, 'privileged': True, 'enabled': True, 'ipc_mode': 'host', 'volumes': ['/etc/kolla/nova-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', 'kolla_logs:/var/log/kolla/', '', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-compute 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'nova-novncproxy', 'value': {'container_name': 'nova_novncproxy', 'group': 'nova-novncproxy', 'image': 'registry.osism.tech/kolla/nova-novncproxy:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-novncproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:6080/vnc_lite.html'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'nova-novncproxy', 'value': {'container_name': 'nova_novncproxy', 'group': 'nova-novncproxy', 'image': 'registry.osism.tech/kolla/nova-novncproxy:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-novncproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:6080/vnc_lite.html'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'nova-conductor', 'value': {'container_name': 'nova_conductor', 'group': 'nova-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-conductor:2024.2', 'volumes': ['/etc/kolla/nova-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'nova-conductor', 'value': {'container_name': 'nova_conductor', 'group': 'nova-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-conductor:2024.2', 'volumes': ['/etc/kolla/nova-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'nova-conductor', 'value': {'container_name': 'nova_conductor', 'group': 'nova-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-conductor:2024.2', 'volumes': ['/etc/kolla/nova-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}})\u001b[0m",
                                "",
                                "TASK [nova-cell : Copying over Nova compute provider config] *******************",
                                "Tuesday 24 February 2026  01:13:31 +0000 (0:00:07.196)       0:05:34.899 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [nova-cell : Copying over libvirt configuration] **************************",
                                "Tuesday 24 February 2026  01:13:33 +0000 (0:00:01.789)       0:05:36.688 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'src': 'qemu.conf.j2', 'dest': 'qemu.conf'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'src': 'qemu.conf.j2', 'dest': 'qemu.conf'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'src': 'qemu.conf.j2', 'dest': 'qemu.conf'}) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'src': 'qemu.conf.j2', 'dest': 'qemu.conf'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'src': 'qemu.conf.j2', 'dest': 'qemu.conf'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'src': 'qemu.conf.j2', 'dest': 'qemu.conf'})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'src': 'libvirtd.conf.j2', 'dest': 'libvirtd.conf'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'src': 'libvirtd.conf.j2', 'dest': 'libvirtd.conf'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'src': 'libvirtd.conf.j2', 'dest': 'libvirtd.conf'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'src': 'libvirtd.conf.j2', 'dest': 'libvirtd.conf'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'src': 'libvirtd.conf.j2', 'dest': 'libvirtd.conf'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'src': 'libvirtd.conf.j2', 'dest': 'libvirtd.conf'})\u001b[0m",
                                "",
                                "TASK [nova-cell : Copying over libvirt TLS keys] *******************************",
                                "Tuesday 24 February 2026  01:13:37 +0000 (0:00:04.058)       0:05:40.747 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [nova-cell : Copying over libvirt SASL configuration] *********************",
                                "Tuesday 24 February 2026  01:13:38 +0000 (0:00:00.674)       0:05:41.421 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'src': 'auth.conf.j2', 'dest': 'auth.conf', 'service': 'nova-compute'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'src': 'auth.conf.j2', 'dest': 'auth.conf', 'service': 'nova-compute'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'src': 'auth.conf.j2', 'dest': 'auth.conf', 'service': 'nova-compute'}) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'src': 'auth.conf.j2', 'dest': 'auth.conf', 'service': 'nova-compute'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'src': 'auth.conf.j2', 'dest': 'auth.conf', 'service': 'nova-compute'})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'src': 'auth.conf.j2', 'dest': 'auth.conf', 'service': 'nova-libvirt'}) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'src': 'auth.conf.j2', 'dest': 'auth.conf', 'service': 'nova-compute'})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'src': 'auth.conf.j2', 'dest': 'auth.conf', 'service': 'nova-libvirt'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'src': 'auth.conf.j2', 'dest': 'auth.conf', 'service': 'nova-libvirt'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'src': 'sasl.conf.j2', 'dest': 'sasl.conf', 'service': 'nova-libvirt'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'src': 'sasl.conf.j2', 'dest': 'sasl.conf', 'service': 'nova-libvirt'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'src': 'sasl.conf.j2', 'dest': 'sasl.conf', 'service': 'nova-libvirt'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'src': 'auth.conf.j2', 'dest': 'auth.conf', 'service': 'nova-libvirt'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'src': 'auth.conf.j2', 'dest': 'auth.conf', 'service': 'nova-libvirt'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'src': 'auth.conf.j2', 'dest': 'auth.conf', 'service': 'nova-libvirt'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'src': 'sasl.conf.j2', 'dest': 'sasl.conf', 'service': 'nova-libvirt'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'src': 'sasl.conf.j2', 'dest': 'sasl.conf', 'service': 'nova-libvirt'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'src': 'sasl.conf.j2', 'dest': 'sasl.conf', 'service': 'nova-libvirt'})\u001b[0m",
                                "",
                                "TASK [nova-cell : Copying files for nova-ssh] **********************************",
                                "Tuesday 24 February 2026  01:13:44 +0000 (0:00:05.897)       0:05:47.319 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'src': 'sshd_config.j2', 'dest': 'sshd_config'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'src': 'sshd_config.j2', 'dest': 'sshd_config'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'src': 'sshd_config.j2', 'dest': 'sshd_config'}) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'src': 'sshd_config.j2', 'dest': 'sshd_config'})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'src': 'id_rsa', 'dest': 'id_rsa'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'src': 'id_rsa', 'dest': 'id_rsa'}) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'src': 'sshd_config.j2', 'dest': 'sshd_config'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'src': 'sshd_config.j2', 'dest': 'sshd_config'})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'src': 'id_rsa.pub', 'dest': 'id_rsa.pub'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'src': 'id_rsa.pub', 'dest': 'id_rsa.pub'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'src': 'id_rsa', 'dest': 'id_rsa'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'src': 'ssh_config.j2', 'dest': 'ssh_config'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'src': 'ssh_config.j2', 'dest': 'ssh_config'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'src': 'id_rsa.pub', 'dest': 'id_rsa.pub'}) \u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'src': 'id_rsa', 'dest': 'id_rsa'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'src': 'id_rsa', 'dest': 'id_rsa'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'src': 'id_rsa', 'dest': 'id_rsa'})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'src': 'ssh_config.j2', 'dest': 'ssh_config'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'src': 'id_rsa.pub', 'dest': 'id_rsa.pub'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'src': 'id_rsa.pub', 'dest': 'id_rsa.pub'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'src': 'id_rsa.pub', 'dest': 'id_rsa.pub'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'src': 'ssh_config.j2', 'dest': 'ssh_config'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'src': 'ssh_config.j2', 'dest': 'ssh_config'})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'src': 'ssh_config.j2', 'dest': 'ssh_config'})\u001b[0m",
                                "",
                                "TASK [nova-cell : Copying VMware vCenter CA file] ******************************",
                                "Tuesday 24 February 2026  01:13:52 +0000 (0:00:07.804)       0:05:55.123 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [nova-cell : Copying 'release' file for nova_compute] *********************",
                                "Tuesday 24 February 2026  01:13:53 +0000 (0:00:00.924)       0:05:56.048 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [nova-cell : Generating 'hostnqn' file for nova_compute] ******************",
                                "Tuesday 24 February 2026  01:13:53 +0000 (0:00:00.635)       0:05:56.683 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "",
                                "TASK [nova-cell : Copying over existing policy file] ***************************",
                                "Tuesday 24 February 2026  01:13:55 +0000 (0:00:02.193)       0:05:58.877 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'nova-libvirt', 'value': {'container_name': 'nova_libvirt', 'group': 'compute', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-libvirt:2024.2', 'pid_mode': 'host', 'cgroupns_mode': 'host', 'privileged': True, 'volumes': ['/etc/kolla/nova-libvirt/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', '', '/sys/fs/cgroup:/sys/fs/cgroup', 'kolla_logs:/var/log/kolla/', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', 'nova_libvirt_qemu:/etc/libvirt/qemu', ''], 'dimensions': {'ulimits': {'memlock': {'soft': 67108864, 'hard': 67108864}}}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'virsh version --daemon'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'nova-ssh', 'value': {'container_name': 'nova_ssh', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-ssh:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla', 'nova_compute:/var/lib/nova', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8022'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'key': 'nova-compute', 'value': {'container_name': 'nova_compute', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-compute:2024.2', 'environment': {'LIBGUESTFS_BACKEND': 'direct'}, 'privileged': True, 'enabled': True, 'ipc_mode': 'host', 'volumes': ['/etc/kolla/nova-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', 'kolla_logs:/var/log/kolla/', '', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-compute 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'nova-libvirt', 'value': {'container_name': 'nova_libvirt', 'group': 'compute', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-libvirt:2024.2', 'pid_mode': 'host', 'cgroupns_mode': 'host', 'privileged': True, 'volumes': ['/etc/kolla/nova-libvirt/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', '', '/sys/fs/cgroup:/sys/fs/cgroup', 'kolla_logs:/var/log/kolla/', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', 'nova_libvirt_qemu:/etc/libvirt/qemu', ''], 'dimensions': {'ulimits': {'memlock': {'soft': 67108864, 'hard': 67108864}}}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'virsh version --daemon'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'nova-ssh', 'value': {'container_name': 'nova_ssh', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-ssh:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla', 'nova_compute:/var/lib/nova', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8022'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'key': 'nova-compute', 'value': {'container_name': 'nova_compute', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-compute:2024.2', 'environment': {'LIBGUESTFS_BACKEND': 'direct'}, 'privileged': True, 'enabled': True, 'ipc_mode': 'host', 'volumes': ['/etc/kolla/nova-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', 'kolla_logs:/var/log/kolla/', '', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-compute 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'nova-libvirt', 'value': {'container_name': 'nova_libvirt', 'group': 'compute', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-libvirt:2024.2', 'pid_mode': 'host', 'cgroupns_mode': 'host', 'privileged': True, 'volumes': ['/etc/kolla/nova-libvirt/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', '', '/sys/fs/cgroup:/sys/fs/cgroup', 'kolla_logs:/var/log/kolla/', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', 'nova_libvirt_qemu:/etc/libvirt/qemu', ''], 'dimensions': {'ulimits': {'memlock': {'soft': 67108864, 'hard': 67108864}}}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'virsh version --daemon'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'nova-ssh', 'value': {'container_name': 'nova_ssh', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-ssh:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla', 'nova_compute:/var/lib/nova', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8022'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'key': 'nova-compute', 'value': {'container_name': 'nova_compute', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-compute:2024.2', 'environment': {'LIBGUESTFS_BACKEND': 'direct'}, 'privileged': True, 'enabled': True, 'ipc_mode': 'host', 'volumes': ['/etc/kolla/nova-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', 'kolla_logs:/var/log/kolla/', '', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-compute 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova-novncproxy', 'value': {'container_name': 'nova_novncproxy', 'group': 'nova-novncproxy', 'image': 'registry.osism.tech/kolla/nova-novncproxy:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-novncproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:6080/vnc_lite.html'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'nova-conductor', 'value': {'container_name': 'nova_conductor', 'group': 'nova-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-conductor:2024.2', 'volumes': ['/etc/kolla/nova-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova-novncproxy', 'value': {'container_name': 'nova_novncproxy', 'group': 'nova-novncproxy', 'image': 'registry.osism.tech/kolla/nova-novncproxy:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-novncproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:6080/vnc_lite.html'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'nova-conductor', 'value': {'container_name': 'nova_conductor', 'group': 'nova-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-conductor:2024.2', 'volumes': ['/etc/kolla/nova-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova-novncproxy', 'value': {'container_name': 'nova_novncproxy', 'group': 'nova-novncproxy', 'image': 'registry.osism.tech/kolla/nova-novncproxy:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-novncproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:6080/vnc_lite.html'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'nova-conductor', 'value': {'container_name': 'nova_conductor', 'group': 'nova-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-conductor:2024.2', 'volumes': ['/etc/kolla/nova-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [nova-cell : Copying over vendordata file to containers] ******************",
                                "Tuesday 24 February 2026  01:13:57 +0000 (0:00:01.834)       0:06:00.711 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=nova-compute) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=nova-compute-ironic) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=nova-compute) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=nova-compute-ironic) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=nova-compute) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=nova-compute-ironic) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=nova-compute) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=nova-compute-ironic) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=nova-compute) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=nova-compute-ironic) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=nova-compute) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=nova-compute-ironic) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [nova-cell : Check nova-cell containers] **********************************",
                                "Tuesday 24 February 2026  01:13:58 +0000 (0:00:01.001)       0:06:01.713 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'nova-libvirt', 'value': {'container_name': 'nova_libvirt', 'group': 'compute', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-libvirt:2024.2', 'pid_mode': 'host', 'cgroupns_mode': 'host', 'privileged': True, 'volumes': ['/etc/kolla/nova-libvirt/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', '', '/sys/fs/cgroup:/sys/fs/cgroup', 'kolla_logs:/var/log/kolla/', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', 'nova_libvirt_qemu:/etc/libvirt/qemu', ''], 'dimensions': {'ulimits': {'memlock': {'soft': 67108864, 'hard': 67108864}}}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'virsh version --daemon'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'nova-libvirt', 'value': {'container_name': 'nova_libvirt', 'group': 'compute', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-libvirt:2024.2', 'pid_mode': 'host', 'cgroupns_mode': 'host', 'privileged': True, 'volumes': ['/etc/kolla/nova-libvirt/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', '', '/sys/fs/cgroup:/sys/fs/cgroup', 'kolla_logs:/var/log/kolla/', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', 'nova_libvirt_qemu:/etc/libvirt/qemu', ''], 'dimensions': {'ulimits': {'memlock': {'soft': 67108864, 'hard': 67108864}}}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'virsh version --daemon'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'nova-libvirt', 'value': {'container_name': 'nova_libvirt', 'group': 'compute', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-libvirt:2024.2', 'pid_mode': 'host', 'cgroupns_mode': 'host', 'privileged': True, 'volumes': ['/etc/kolla/nova-libvirt/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', '', '/sys/fs/cgroup:/sys/fs/cgroup', 'kolla_logs:/var/log/kolla/', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', 'nova_libvirt_qemu:/etc/libvirt/qemu', ''], 'dimensions': {'ulimits': {'memlock': {'soft': 67108864, 'hard': 67108864}}}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'virsh version --daemon'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'nova-novncproxy', 'value': {'container_name': 'nova_novncproxy', 'group': 'nova-novncproxy', 'image': 'registry.osism.tech/kolla/nova-novncproxy:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-novncproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:6080/vnc_lite.html'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'nova-novncproxy', 'value': {'container_name': 'nova_novncproxy', 'group': 'nova-novncproxy', 'image': 'registry.osism.tech/kolla/nova-novncproxy:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-novncproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:6080/vnc_lite.html'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'nova-ssh', 'value': {'container_name': 'nova_ssh', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-ssh:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla', 'nova_compute:/var/lib/nova', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8022'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'nova-ssh', 'value': {'container_name': 'nova_ssh', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-ssh:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla', 'nova_compute:/var/lib/nova', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8022'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'nova-ssh', 'value': {'container_name': 'nova_ssh', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-ssh:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-ssh/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla', 'nova_compute:/var/lib/nova', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_listen sshd 8022'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4] => (item={'key': 'nova-compute', 'value': {'container_name': 'nova_compute', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-compute:2024.2', 'environment': {'LIBGUESTFS_BACKEND': 'direct'}, 'privileged': True, 'enabled': True, 'ipc_mode': 'host', 'volumes': ['/etc/kolla/nova-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', 'kolla_logs:/var/log/kolla/', '', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-compute 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'nova-novncproxy', 'value': {'container_name': 'nova_novncproxy', 'group': 'nova-novncproxy', 'image': 'registry.osism.tech/kolla/nova-novncproxy:2024.2', 'enabled': True, 'volumes': ['/etc/kolla/nova-novncproxy/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:6080/vnc_lite.html'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'nova-conductor', 'value': {'container_name': 'nova_conductor', 'group': 'nova-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-conductor:2024.2', 'volumes': ['/etc/kolla/nova-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'nova-conductor', 'value': {'container_name': 'nova_conductor', 'group': 'nova-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-conductor:2024.2', 'volumes': ['/etc/kolla/nova-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3] => (item={'key': 'nova-compute', 'value': {'container_name': 'nova_compute', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-compute:2024.2', 'environment': {'LIBGUESTFS_BACKEND': 'direct'}, 'privileged': True, 'enabled': True, 'ipc_mode': 'host', 'volumes': ['/etc/kolla/nova-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', 'kolla_logs:/var/log/kolla/', '', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-compute 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5] => (item={'key': 'nova-compute', 'value': {'container_name': 'nova_compute', 'group': 'compute', 'image': 'registry.osism.tech/kolla/nova-compute:2024.2', 'environment': {'LIBGUESTFS_BACKEND': 'direct'}, 'privileged': True, 'enabled': True, 'ipc_mode': 'host', 'volumes': ['/etc/kolla/nova-compute/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', '/lib/modules:/lib/modules:ro', '/run:/run:shared', '/dev:/dev', 'kolla_logs:/var/log/kolla/', '', 'libvirtd:/var/lib/libvirt', 'nova_compute:/var/lib/nova/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-compute 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'nova-conductor', 'value': {'container_name': 'nova_conductor', 'group': 'nova-conductor', 'enabled': True, 'image': 'registry.osism.tech/kolla/nova-conductor:2024.2', 'volumes': ['/etc/kolla/nova-conductor/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port nova-conductor 5672'], 'timeout': '30'}}})\u001b[0m",
                                "",
                                "TASK [nova-cell : include_tasks] ***********************************************",
                                "Tuesday 24 February 2026  01:14:01 +0000 (0:00:03.157)       0:06:04.871 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [nova-cell : Flush handlers] **********************************************",
                                "Tuesday 24 February 2026  01:14:02 +0000 (0:00:00.891)       0:06:05.762 ****** ",
                                "",
                                "TASK [nova-cell : Flush handlers] **********************************************",
                                "Tuesday 24 February 2026  01:14:02 +0000 (0:00:00.154)       0:06:05.917 ****** ",
                                "",
                                "TASK [nova-cell : Flush handlers] **********************************************",
                                "Tuesday 24 February 2026  01:14:03 +0000 (0:00:00.141)       0:06:06.058 ****** ",
                                "",
                                "TASK [nova-cell : Flush handlers] **********************************************",
                                "Tuesday 24 February 2026  01:14:03 +0000 (0:00:00.140)       0:06:06.199 ****** ",
                                "",
                                "TASK [nova-cell : Flush handlers] **********************************************",
                                "Tuesday 24 February 2026  01:14:03 +0000 (0:00:00.133)       0:06:06.332 ****** ",
                                "",
                                "TASK [nova-cell : Flush handlers] **********************************************",
                                "Tuesday 24 February 2026  01:14:03 +0000 (0:00:00.345)       0:06:06.678 ****** ",
                                "",
                                "RUNNING HANDLER [nova-cell : Restart nova-conductor container] *****************",
                                "Tuesday 24 February 2026  01:14:03 +0000 (0:00:00.154)       0:06:06.832 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [nova-cell : Restart nova-novncproxy container] ****************",
                                "Tuesday 24 February 2026  01:14:11 +0000 (0:00:07.882)       0:06:14.715 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [nova-cell : Restart nova-ssh container] ***********************",
                                "Tuesday 24 February 2026  01:14:30 +0000 (0:00:18.894)       0:06:33.610 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "",
                                "RUNNING HANDLER [nova-cell : Restart nova-libvirt container] *******************",
                                "Tuesday 24 February 2026  01:14:53 +0000 (0:00:23.141)       0:06:56.751 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "",
                                "RUNNING HANDLER [nova-cell : Checking libvirt container is ready] **************",
                                "Tuesday 24 February 2026  01:15:26 +0000 (0:00:32.606)       0:07:29.357 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "RUNNING HANDLER [nova-cell : Create libvirt SASL user] *************************",
                                "Tuesday 24 February 2026  01:15:27 +0000 (0:00:00.821)       0:07:30.178 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "RUNNING HANDLER [nova-cell : Restart nova-compute container] *******************",
                                "Tuesday 24 February 2026  01:15:27 +0000 (0:00:00.807)       0:07:30.986 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "",
                                "RUNNING HANDLER [nova-cell : Wait for nova-compute services to update service versions] ***",
                                "Tuesday 24 February 2026  01:15:53 +0000 (0:00:25.952)       0:07:56.938 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [nova-cell : Waiting for nova-compute services to register themselves] ****",
                                "Tuesday 24 February 2026  01:15:54 +0000 (0:00:00.184)       0:07:57.123 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[1;30mFAILED - RETRYING: [testbed-node-3 -> testbed-node-0]: Waiting for nova-compute services to register themselves (20 retries left).\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-0(192.168.16.10)]\u001b[0m",
                                "",
                                "TASK [nova-cell : Fail if nova-compute service failed to register] *************",
                                "Tuesday 24 February 2026  01:16:17 +0000 (0:00:23.609)       0:08:20.732 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [nova-cell : Include discover_computes.yml] *******************************",
                                "Tuesday 24 February 2026  01:16:30 +0000 (0:00:12.676)       0:08:33.408 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mincluded: /ansible/roles/nova-cell/tasks/discover_computes.yml for testbed-node-3\u001b[0m",
                                "",
                                "TASK [nova-cell : Get a list of existing cells] ********************************",
                                "Tuesday 24 February 2026  01:16:34 +0000 (0:00:04.225)       0:08:37.634 ****** ",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-0(192.168.16.10)]\u001b[0m",
                                "",
                                "TASK [nova-cell : Extract current cell settings from list] *********************",
                                "Tuesday 24 February 2026  01:16:49 +0000 (0:00:14.715)       0:08:52.349 ****** ",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-0(192.168.16.10)]\u001b[0m",
                                "",
                                "TASK [nova-cell : Fail if cell settings not found] *****************************",
                                "Tuesday 24 February 2026  01:16:51 +0000 (0:00:01.656)       0:08:54.006 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [nova-cell : Discover nova hosts] *****************************************",
                                "Tuesday 24 February 2026  01:16:52 +0000 (0:00:01.577)       0:08:55.583 ****** ",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-node-0(192.168.16.10)]\u001b[0m",
                                "",
                                "TASK [nova-cell : Remove old nova_libvirt_secrets container volume] ************",
                                "Tuesday 24 February 2026  01:17:05 +0000 (0:00:12.692)       0:09:08.275 ****** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "PLAY [Refresh nova scheduler cell cache] ***************************************",
                                "",
                                "TASK [nova : Refresh cell cache in nova scheduler] *****************************",
                                "Tuesday 24 February 2026  01:17:07 +0000 (0:00:02.057)       0:09:10.333 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "PLAY [Reload global Nova super conductor services] *****************************",
                                "",
                                "TASK [nova : Reload nova super conductor services to remove RPC version pin] ***",
                                "Tuesday 24 February 2026  01:17:08 +0000 (0:00:01.342)       0:09:11.675 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "PLAY [Reload Nova cell services] ***********************************************",
                                "",
                                "TASK [nova-cell : Reload nova cell services to remove RPC version cap] *********",
                                "Tuesday 24 February 2026  01:17:09 +0000 (0:00:00.689)       0:09:12.365 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=nova-conductor) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=nova-compute) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=nova-compute-ironic) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=nova-novncproxy) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=nova-serialproxy) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item=nova-spicehtml5proxy) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=nova-conductor) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=nova-compute) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=nova-compute-ironic) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=nova-novncproxy) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=nova-serialproxy) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item=nova-spicehtml5proxy) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=nova-conductor) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=nova-compute) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=nova-compute-ironic) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=nova-novncproxy) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=nova-serialproxy) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item=nova-spicehtml5proxy) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=nova-conductor) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=nova-compute) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=nova-compute-ironic) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=nova-novncproxy) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=nova-serialproxy) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=nova-spicehtml5proxy) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=nova-conductor) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=nova-compute) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=nova-compute-ironic) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=nova-conductor) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=nova-compute) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=nova-novncproxy) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=nova-compute-ironic) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=nova-serialproxy) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=nova-novncproxy) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=nova-spicehtml5proxy) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=nova-serialproxy) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=nova-spicehtml5proxy) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "PLAY [Reload global Nova API services] *****************************************",
                                "",
                                "TASK [nova : Reload nova API services to remove RPC version pin] ***************",
                                "Tuesday 24 February 2026  01:17:11 +0000 (0:00:01.956)       0:09:14.321 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=nova-scheduler) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item=nova-api) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=nova-scheduler) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item=nova-api) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=nova-scheduler) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item=nova-api) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "PLAY [Run Nova API online data migrations] *************************************",
                                "",
                                "TASK [nova : Run Nova API online database migrations] **************************",
                                "Tuesday 24 February 2026  01:17:12 +0000 (0:00:00.916)       0:09:15.238 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "PLAY [Run Nova cell online data migrations] ************************************",
                                "",
                                "TASK [nova-cell : Run Nova cell online database migrations] ********************",
                                "Tuesday 24 February 2026  01:17:13 +0000 (0:00:01.028)       0:09:16.267 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;33mtestbed-manager\u001b[0m            : \u001b[0;32mok=3   \u001b[0m \u001b[0;33mchanged=3   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=54  \u001b[0m \u001b[0;33mchanged=35  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=45  \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=27  \u001b[0m \u001b[0;33mchanged=19  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=52  \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=27  \u001b[0m \u001b[0;33mchanged=19  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=52  \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-3\u001b[0m             : \u001b[0;32mok=45  \u001b[0m \u001b[0;33mchanged=27  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=21  \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-4\u001b[0m             : \u001b[0;32mok=39  \u001b[0m \u001b[0;33mchanged=27  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=20  \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-5\u001b[0m             : \u001b[0;32mok=39  \u001b[0m \u001b[0;33mchanged=27  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=20  \u001b[0m rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Tuesday 24 February 2026  01:17:14 +0000 (0:00:00.757)       0:09:17.024 ****** ",
                                "=============================================================================== ",
                                "nova : Running Nova API bootstrap container ---------------------------- 35.17s",
                                "nova-cell : Restart nova-libvirt container ----------------------------- 32.61s",
                                "nova-cell : Restart nova-compute container ----------------------------- 25.95s",
                                "nova-cell : Waiting for nova-compute services to register themselves --- 23.61s",
                                "nova-cell : Restart nova-ssh container --------------------------------- 23.14s",
                                "nova : Running Nova API bootstrap container ---------------------------- 23.08s",
                                "nova : Restart nova-scheduler container -------------------------------- 22.54s",
                                "nova-cell : Running Nova cell bootstrap container ---------------------- 22.16s",
                                "nova-cell : Restart nova-novncproxy container -------------------------- 18.89s",
                                "nova : Create cell0 mappings ------------------------------------------- 17.69s",
                                "nova-cell : Get a list of existing cells ------------------------------- 14.72s",
                                "nova-cell : Get a list of existing cells ------------------------------- 14.69s",
                                "nova-cell : Create cell ------------------------------------------------ 14.40s",
                                "nova-cell : Get a list of existing cells ------------------------------- 13.55s",
                                "nova-cell : Discover nova hosts ---------------------------------------- 12.69s",
                                "nova-cell : Fail if nova-compute service failed to register ------------ 12.68s",
                                "service-rabbitmq : nova | Ensure RabbitMQ users exist ------------------- 8.82s",
                                "service-ks-register : nova | Granting user roles ------------------------ 8.53s",
                                "nova-cell : Restart nova-conductor container ---------------------------- 7.88s",
                                "nova-cell : Copying files for nova-ssh ---------------------------------- 7.80s",
                                "\u001b[32m2026-02-24 01:17:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 01:17:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 79f56a03-a0d2-400b-aa03-9d9321f018b2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:17:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 01:17:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 79f56a03-a0d2-400b-aa03-9d9321f018b2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:17:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 01:17:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 79f56a03-a0d2-400b-aa03-9d9321f018b2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:17:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 01:17:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 79f56a03-a0d2-400b-aa03-9d9321f018b2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:17:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 01:17:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 79f56a03-a0d2-400b-aa03-9d9321f018b2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:17:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 01:17:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 79f56a03-a0d2-400b-aa03-9d9321f018b2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:17:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 01:17:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 79f56a03-a0d2-400b-aa03-9d9321f018b2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:17:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 01:17:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 79f56a03-a0d2-400b-aa03-9d9321f018b2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:17:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "\u001b[32m2026-02-24 01:17:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 79f56a03-a0d2-400b-aa03-9d9321f018b2 is in state STARTED\u001b[0m",
                                "\u001b[32m2026-02-24 01:17:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until the next check\u001b[0m",
                                "",
                                "",
                                "PLAY [Group hosts based on configuration] **************************************",
                                "",
                                "TASK [Group hosts based on Kolla action] ***************************************",
                                "Tuesday 24 February 2026  01:12:37 +0000 (0:00:00.286)       0:00:00.286 ****** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [Group hosts based on enabled services] ***********************************",
                                "Tuesday 24 February 2026  01:12:38 +0000 (0:00:00.385)       0:00:00.672 ****** ",
                                "\u001b[0;32mok: [testbed-node-0] => (item=enable_octavia_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item=enable_octavia_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item=enable_octavia_True)\u001b[0m",
                                "",
                                "PLAY [Apply role octavia] ******************************************************",
                                "",
                                "TASK [octavia : include_tasks] *************************************************",
                                "Tuesday 24 February 2026  01:12:38 +0000 (0:00:00.502)       0:00:01.174 ****** ",
                                "\u001b[0;36mincluded: /ansible/roles/octavia/tasks/deploy.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [service-ks-register : octavia | Creating services] ***********************",
                                "Tuesday 24 February 2026  01:12:39 +0000 (0:00:00.591)       0:00:01.766 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=octavia (load-balancer))\u001b[0m",
                                "",
                                "TASK [service-ks-register : octavia | Creating endpoints] **********************",
                                "Tuesday 24 February 2026  01:12:43 +0000 (0:00:03.714)       0:00:05.480 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=octavia -> https://api-int.testbed.osism.xyz:9876 -> internal)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=octavia -> https://api.testbed.osism.xyz:9876 -> public)\u001b[0m",
                                "",
                                "TASK [service-ks-register : octavia | Creating projects] ***********************",
                                "Tuesday 24 February 2026  01:12:49 +0000 (0:00:06.856)       0:00:12.337 ****** ",
                                "\u001b[0;32mok: [testbed-node-0] => (item=service)\u001b[0m",
                                "",
                                "TASK [service-ks-register : octavia | Creating users] **************************",
                                "Tuesday 24 February 2026  01:12:53 +0000 (0:00:03.144)       0:00:15.482 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=octavia -> service)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=octavia -> service)\u001b[0m",
                                "\u001b[1;35m[WARNING]: Module did not set no_log for update_password\u001b[0m",
                                "",
                                "TASK [service-ks-register : octavia | Creating roles] **************************",
                                "Tuesday 24 February 2026  01:13:02 +0000 (0:00:08.997)       0:00:24.479 ****** ",
                                "\u001b[0;32mok: [testbed-node-0] => (item=admin)\u001b[0m",
                                "",
                                "TASK [service-ks-register : octavia | Granting user roles] *********************",
                                "Tuesday 24 February 2026  01:13:06 +0000 (0:00:04.127)       0:00:28.607 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=octavia -> service -> admin)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0] => (item=octavia -> service -> admin)\u001b[0m",
                                "",
                                "TASK [octavia : Adding octavia related roles] **********************************",
                                "Tuesday 24 February 2026  01:13:13 +0000 (0:00:07.061)       0:00:35.668 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=load-balancer_observer)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=load-balancer_global_observer)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=load-balancer_member)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=load-balancer_admin)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=load-balancer_quota_admin)\u001b[0m",
                                "",
                                "TASK [octavia : include_tasks] *************************************************",
                                "Tuesday 24 February 2026  01:13:30 +0000 (0:00:17.369)       0:00:53.038 ****** ",
                                "\u001b[0;36mincluded: /ansible/roles/octavia/tasks/prepare.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [octavia : Create amphora flavor] *****************************************",
                                "Tuesday 24 February 2026  01:13:31 +0000 (0:00:00.715)       0:00:53.753 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [octavia : Create nova keypair for amphora] *******************************",
                                "Tuesday 24 February 2026  01:13:37 +0000 (0:00:05.860)       0:00:59.614 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [octavia : Get service project id] ****************************************",
                                "Tuesday 24 February 2026  01:13:42 +0000 (0:00:05.274)       0:01:04.888 ****** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [octavia : Create security groups for octavia] ****************************",
                                "Tuesday 24 February 2026  01:13:46 +0000 (0:00:03.939)       0:01:08.827 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=lb-mgmt-sec-grp)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=lb-health-mgr-sec-grp)\u001b[0m",
                                "",
                                "TASK [octavia : Add rules for security groups] *********************************",
                                "Tuesday 24 February 2026  01:13:57 +0000 (0:00:11.020)       0:01:19.848 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=[{'name': 'lb-mgmt-sec-grp', 'enabled': True}, {'protocol': 'icmp'}])\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=[{'name': 'lb-mgmt-sec-grp', 'enabled': True}, {'protocol': 'tcp', 'src_port': 22, 'dst_port': 22}])\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=[{'name': 'lb-mgmt-sec-grp', 'enabled': True}, {'protocol': 'tcp', 'src_port': '9443', 'dst_port': '9443'}])\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=[{'name': 'lb-health-mgr-sec-grp', 'enabled': True}, {'protocol': 'udp', 'src_port': '5555', 'dst_port': '5555'}])\u001b[0m",
                                "",
                                "TASK [octavia : Create loadbalancer management network] ************************",
                                "Tuesday 24 February 2026  01:14:12 +0000 (0:00:15.234)       0:01:35.083 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [octavia : Create loadbalancer management subnet] *************************",
                                "Tuesday 24 February 2026  01:14:16 +0000 (0:00:04.275)       0:01:39.359 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [octavia : Create loadbalancer management router for IPv6] ****************",
                                "Tuesday 24 February 2026  01:14:23 +0000 (0:00:06.702)       0:01:46.062 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [octavia : Update loadbalancer management subnet] *************************",
                                "Tuesday 24 February 2026  01:14:23 +0000 (0:00:00.253)       0:01:46.315 ****** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [octavia : include_tasks] *************************************************",
                                "Tuesday 24 February 2026  01:14:27 +0000 (0:00:03.925)       0:01:50.240 ****** ",
                                "\u001b[0;36mincluded: /ansible/roles/octavia/tasks/hm-interface.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [octavia : Create ports for Octavia health-manager nodes] *****************",
                                "Tuesday 24 February 2026  01:14:29 +0000 (0:00:01.228)       0:01:51.468 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [octavia : Update Octavia health manager port host_id] ********************",
                                "Tuesday 24 February 2026  01:14:35 +0000 (0:00:06.029)       0:01:57.498 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [octavia : Add Octavia port to openvswitch br-int] ************************",
                                "Tuesday 24 February 2026  01:14:40 +0000 (0:00:05.274)       0:02:02.772 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [octavia : Install isc-dhcp-client package] *******************************",
                                "Tuesday 24 February 2026  01:14:41 +0000 (0:00:00.823)       0:02:03.595 ****** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "",
                                "TASK [octavia : Create octavia dhclient conf] **********************************",
                                "Tuesday 24 February 2026  01:14:44 +0000 (0:00:02.781)       0:02:06.377 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [octavia : Create octavia-interface service] ******************************",
                                "Tuesday 24 February 2026  01:14:45 +0000 (0:00:01.414)       0:02:07.791 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [octavia : Restart octavia-interface.service if required] *****************",
                                "Tuesday 24 February 2026  01:14:46 +0000 (0:00:01.272)       0:02:09.064 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [octavia : Enable and start octavia-interface.service] ********************",
                                "Tuesday 24 February 2026  01:14:48 +0000 (0:00:02.238)       0:02:11.303 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [octavia : Wait for interface ohm0 ip appear] *****************************",
                                "Tuesday 24 February 2026  01:14:51 +0000 (0:00:02.075)       0:02:13.378 ****** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [octavia : Gather facts] **************************************************",
                                "Tuesday 24 February 2026  01:14:51 +0000 (0:00:00.676)       0:02:14.055 ****** ",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [octavia : include_tasks] *************************************************",
                                "Tuesday 24 February 2026  01:14:55 +0000 (0:00:03.875)       0:02:17.930 ****** ",
                                "\u001b[0;36mincluded: /ansible/roles/octavia/tasks/get_resources_info.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [octavia : Get amphora flavor info] ***************************************",
                                "Tuesday 24 February 2026  01:14:56 +0000 (0:00:00.832)       0:02:18.763 ****** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [octavia : Get service project id] ****************************************",
                                "Tuesday 24 February 2026  01:15:00 +0000 (0:00:03.884)       0:02:22.647 ****** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [octavia : Get security groups for octavia] *******************************",
                                "Tuesday 24 February 2026  01:15:04 +0000 (0:00:04.068)       0:02:26.716 ****** ",
                                "\u001b[0;32mok: [testbed-node-0] => (item=lb-mgmt-sec-grp)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0] => (item=lb-health-mgr-sec-grp)\u001b[0m",
                                "",
                                "TASK [octavia : Get loadbalancer management network] ***************************",
                                "Tuesday 24 February 2026  01:15:11 +0000 (0:00:07.173)       0:02:33.890 ****** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [octavia : Set octavia resources facts] ***********************************",
                                "Tuesday 24 February 2026  01:15:15 +0000 (0:00:03.806)       0:02:37.696 ****** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [octavia : Ensuring config directories exist] *****************************",
                                "Tuesday 24 February 2026  01:15:15 +0000 (0:00:00.374)       0:02:38.071 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-api:2024.2', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-api:2024.2', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-api:2024.2', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-driver-agent:2024.2', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-driver-agent:2024.2', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-driver-agent:2024.2', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-health-manager:2024.2', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-health-manager:2024.2', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-health-manager:2024.2', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-housekeeping:2024.2', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-housekeeping:2024.2', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-housekeeping:2024.2', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-worker:2024.2', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-worker:2024.2', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-worker:2024.2', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}})\u001b[0m",
                                "",
                                "TASK [octavia : Check if policies shall be overwritten] ************************",
                                "Tuesday 24 February 2026  01:15:18 +0000 (0:00:02.697)       0:02:40.768 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [octavia : Set octavia policy file] ***************************************",
                                "Tuesday 24 February 2026  01:15:18 +0000 (0:00:00.147)       0:02:40.916 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [octavia : Copying over existing policy file] *****************************",
                                "Tuesday 24 February 2026  01:15:19 +0000 (0:00:00.637)       0:02:41.553 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-api:2024.2', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-driver-agent:2024.2', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-health-manager:2024.2', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-housekeeping:2024.2', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-worker:2024.2', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-api:2024.2', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-driver-agent:2024.2', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-health-manager:2024.2', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-housekeeping:2024.2', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-worker:2024.2', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-api:2024.2', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-driver-agent:2024.2', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-health-manager:2024.2', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-housekeeping:2024.2', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-worker:2024.2', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [octavia : include_tasks] *************************************************",
                                "Tuesday 24 February 2026  01:15:19 +0000 (0:00:00.795)       0:02:42.348 ****** ",
                                "\u001b[0;36mincluded: /ansible/roles/octavia/tasks/copy-certs.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [service-cert-copy : octavia | Copying over extra CA certificates] ********",
                                "Tuesday 24 February 2026  01:15:20 +0000 (0:00:00.647)       0:02:42.996 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-api:2024.2', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-api:2024.2', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9876'], 'timeout': '30'}, 'hapr\u001b[32m2026-02-24 01:17:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 79f56a03-a0d2-400b-aa03-9d9321f018b2 is in state SUCCESS\u001b[0m",
                                "oxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-api:2024.2', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-driver-agent:2024.2', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-driver-agent:2024.2', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-driver-agent:2024.2', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-health-manager:2024.2', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-health-manager:2024.2', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-health-manager:2024.2', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-housekeeping:2024.2', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-housekeeping:2024.2', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-housekeeping:2024.2', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-worker:2024.2', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-worker:2024.2', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-worker:2024.2', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}})\u001b[0m",
                                "",
                                "TASK [service-cert-copy : octavia | Copying over backend internal TLS certificate] ***",
                                "Tuesday 24 February 2026  01:15:26 +0000 (0:00:06.004)       0:02:49.001 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-api:2024.2', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-driver-agent:2024.2', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-health-manager:2024.2', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-housekeeping:2024.2', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-worker:2024.2', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-api:2024.2', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-driver-agent:2024.2', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-health-manager:2024.2', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-housekeeping:2024.2', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-worker:2024.2', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-api:2024.2', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-driver-agent:2024.2', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-health-manager:2024.2', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-housekeeping:2024.2', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-worker:2024.2', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [service-cert-copy : octavia | Copying over backend internal TLS key] *****",
                                "Tuesday 24 February 2026  01:15:27 +0000 (0:00:00.760)       0:02:49.762 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-api:2024.2', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-driver-agent:2024.2', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-health-manager:2024.2', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-housekeeping:2024.2', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-worker:2024.2', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-api:2024.2', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-driver-agent:2024.2', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-health-manager:2024.2', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-housekeeping:2024.2', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-worker:2024.2', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-api:2024.2', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-driver-agent:2024.2', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-health-manager:2024.2', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-housekeeping:2024.2', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-worker:2024.2', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [octavia : Copying over config.json files for services] *******************",
                                "Tuesday 24 February 2026  01:15:28 +0000 (0:00:01.303)       0:02:51.066 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-api:2024.2', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-api:2024.2', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-api:2024.2', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-driver-agent:2024.2', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-driver-agent:2024.2', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-driver-agent:2024.2', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[32m2026-02-24 01:17:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until refresh of running tasks\u001b[0m",
                                " => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-health-manager:2024.2', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-health-manager:2024.2', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-health-manager:2024.2', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-housekeeping:2024.2', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-housekeeping:2024.2', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-housekeeping:2024.2', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-worker:2024.2', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-worker:2024.2', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-worker:2024.2', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}})\u001b[0m",
                                "",
                                "TASK [octavia : Copying over octavia-wsgi.conf] ********************************",
                                "Tuesday 24 February 2026  01:15:35 +0000 (0:00:06.622)       0:02:57.689 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=/ansible/roles/octavia/templates/octavia-wsgi.conf.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=/ansible/roles/octavia/templates/octavia-wsgi.conf.j2)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=/ansible/roles/octavia/templates/octavia-wsgi.conf.j2)\u001b[0m",
                                "",
                                "TASK [octavia : Copying over octavia.conf] *************************************",
                                "Tuesday 24 February 2026  01:15:37 +0000 (0:00:02.575)       0:03:00.264 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-api:2024.2', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-api:2024.2', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-api:2024.2', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-driver-agent:2024.2', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-driver-agent:2024.2', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-driver-agent:2024.2', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-health-manager:2024.2', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-health-manager:2024.2', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-health-manager:2024.2', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-housekeeping:2024.2', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-housekeeping:2024.2', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-housekeeping:2024.2', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-worker:2024.2', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-worker:2024.2', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-worker:2024.2', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}})\u001b[0m",
                                "",
                                "TASK [octavia : Copying over Octavia SSH key] **********************************",
                                "Tuesday 24 February 2026  01:16:00 +0000 (0:00:22.424)       0:03:22.689 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [octavia : Copying certificate files for octavia-worker] ******************",
                                "Tuesday 24 February 2026  01:16:02 +0000 (0:00:01.718)       0:03:24.407 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=client.cert-and-key.pem)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=client.cert-and-key.pem)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=client.cert-and-key.pem)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=client_ca.cert.pem)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=client_ca.cert.pem)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=client_ca.cert.pem)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=server_ca.cert.pem)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=server_ca.cert.pem)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=server_ca.cert.pem)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=server_ca.key.pem)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=server_ca.key.pem)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=server_ca.key.pem)\u001b[0m",
                                "",
                                "TASK [octavia : Copying certificate files for octavia-housekeeping] ************",
                                "Tuesday 24 February 2026  01:16:07 +0000 (0:00:05.867)       0:03:30.275 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=client.cert-and-key.pem)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=client.cert-and-key.pem)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=client.cert-and-key.pem)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=client_ca.cert.pem)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=client_ca.cert.pem)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=client_ca.cert.pem)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=server_ca.cert.pem)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=server_ca.cert.pem)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=server_ca.cert.pem)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=server_ca.key.pem)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=server_ca.key.pem)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=server_ca.key.pem)\u001b[0m",
                                "",
                                "TASK [octavia : Copying certificate files for octavia-health-manager] **********",
                                "Tuesday 24 February 2026  01:16:14 +0000 (0:00:06.159)       0:03:36.434 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=client.cert-and-key.pem)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=client.cert-and-key.pem)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=client.cert-and-key.pem)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=client_ca.cert.pem)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=client_ca.cert.pem)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=client_ca.cert.pem)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=server_ca.cert.pem)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=server_ca.cert.pem)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=server_ca.cert.pem)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item=server_ca.key.pem)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item=server_ca.key.pem)\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item=server_ca.key.pem)\u001b[0m",
                                "",
                                "TASK [octavia : Check octavia containers] **************************************",
                                "Tuesday 24 February 2026  01:16:20 +0000 (0:00:06.732)       0:03:43.167 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-api:2024.2', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.11:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-api:2024.2', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.10:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'octavia-api', 'value': {'container_name': 'octavia_api', 'group': 'octavia-api', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-api:2024.2', 'volumes': ['/etc/kolla/octavia-api/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_curl http://192.168.16.12:9876'], 'timeout': '30'}, 'haproxy': {'octavia_api': {'enabled': 'yes', 'mode': 'http', 'external': False, 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}, 'octavia_api_external': {'enabled': 'yes', 'mode': 'http', 'external': True, 'external_fqdn': 'api.testbed.osism.xyz', 'port': '9876', 'listen_port': '9876', 'tls_backend': 'no'}}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-driver-agent:2024.2', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-driver-agent:2024.2', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'octavia-driver-agent', 'value': {'container_name': 'octavia_driver_agent', 'group': 'octavia-driver-agent', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-driver-agent:2024.2', 'volumes': ['/etc/kolla/octavia-driver-agent/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', '', 'octavia_driver_agent:/var/run/octavia/'], 'dimensions': {}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-health-manager:2024.2', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-health-manager:2024.2', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'octavia-health-manager', 'value': {'container_name': 'octavia_health_manager', 'group': 'octavia-health-manager', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-health-manager:2024.2', 'volumes': ['/etc/kolla/octavia-health-manager/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-health-manager 3306'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-housekeeping:2024.2', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-housekeeping:2024.2', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'octavia-housekeeping', 'value': {'container_name': 'octavia_housekeeping', 'group': 'octavia-housekeeping', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-housekeeping:2024.2', 'volumes': ['/etc/kolla/octavia-housekeeping/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-housekeeping 3306'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-worker:2024.2', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-worker:2024.2', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2] => (item={'key': 'octavia-worker', 'value': {'container_name': 'octavia_worker', 'group': 'octavia-worker', 'enabled': True, 'image': 'registry.osism.tech/kolla/octavia-worker:2024.2', 'volumes': ['/etc/kolla/octavia-worker/:/var/lib/kolla/config_files/:ro', '/etc/localtime:/etc/localtime:ro', '/etc/timezone:/etc/timezone:ro', 'kolla_logs:/var/log/kolla/', '', ''], 'dimensions': {}, 'healthcheck': {'interval': '30', 'retries': '3', 'start_period': '5', 'test': ['CMD-SHELL', 'healthcheck_port octavia-worker 5672'], 'timeout': '30'}}})\u001b[0m",
                                "",
                                "TASK [octavia : include_tasks] *************************************************",
                                "Tuesday 24 February 2026  01:16:26 +0000 (0:00:05.296)       0:03:48.463 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [octavia : Creating Octavia database] *************************************",
                                "Tuesday 24 February 2026  01:16:26 +0000 (0:00:00.770)       0:03:49.233 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [octavia : Creating Octavia persistence database] *************************",
                                "Tuesday 24 February 2026  01:16:29 +0000 (0:00:02.658)       0:03:51.892 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [octavia : Creating Octavia database user and setting permissions] ********",
                                "Tuesday 24 February 2026  01:16:32 +0000 (0:00:02.602)       0:03:54.494 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [octavia : Creating Octavia persistence database user and setting permissions] ***",
                                "Tuesday 24 February 2026  01:16:34 +0000 (0:00:02.733)       0:03:57.228 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [octavia : Running Octavia bootstrap container] ***************************",
                                "Tuesday 24 February 2026  01:16:37 +0000 (0:00:03.109)       0:04:00.337 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [octavia : Flush handlers] ************************************************",
                                "Tuesday 24 February 2026  01:16:58 +0000 (0:00:21.009)       0:04:21.347 ****** ",
                                "",
                                "TASK [octavia : Flush handlers] ************************************************",
                                "Tuesday 24 February 2026  01:16:59 +0000 (0:00:00.093)       0:04:21.441 ****** ",
                                "",
                                "TASK [octavia : Flush handlers] ************************************************",
                                "Tuesday 24 February 2026  01:16:59 +0000 (0:00:00.069)       0:04:21.510 ****** ",
                                "",
                                "RUNNING HANDLER [octavia : Restart octavia-api container] **********************",
                                "Tuesday 24 February 2026  01:16:59 +0000 (0:00:00.109)       0:04:21.619 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "RUNNING HANDLER [octavia : Restart octavia-driver-agent container] *************",
                                "Tuesday 24 February 2026  01:17:09 +0000 (0:00:09.924)       0:04:31.544 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "",
                                "RUNNING HANDLER [octavia : Restart octavia-health-manager container] ***********",
                                "Tuesday 24 February 2026  01:17:15 +0000 (0:00:06.526)       0:04:38.071 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "",
                                "RUNNING HANDLER [octavia : Restart octavia-housekeeping container] *************",
                                "Tuesday 24 February 2026  01:17:26 +0000 (0:00:11.115)       0:04:49.186 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "",
                                "RUNNING HANDLER [octavia : Restart octavia-worker container] *******************",
                                "Tuesday 24 February 2026  01:17:37 +0000 (0:00:10.444)       0:04:59.630 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-1]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-2]\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=57  \u001b[0m \u001b[0;33mchanged=38  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=7   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-1\u001b[0m             : \u001b[0;32mok=33  \u001b[0m \u001b[0;33mchanged=22  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=5   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-2\u001b[0m             : \u001b[0;32mok=33  \u001b[0m \u001b[0;33mchanged=22  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=5   \u001b[0m rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Tuesday 24 February 2026  01:17:43 +0000 (0:00:06.392)       0:05:06.023 ****** ",
                                "=============================================================================== ",
                                "octavia : Copying over octavia.conf ------------------------------------ 22.42s",
                                "octavia : Running Octavia bootstrap container -------------------------- 21.01s",
                                "octavia : Adding octavia related roles --------------------------------- 17.37s",
                                "octavia : Add rules for security groups -------------------------------- 15.23s",
                                "octavia : Restart octavia-health-manager container --------------------- 11.12s",
                                "octavia : Create security groups for octavia --------------------------- 11.02s",
                                "octavia : Restart octavia-housekeeping container ----------------------- 10.44s",
                                "octavia : Restart octavia-api container --------------------------------- 9.92s",
                                "service-ks-register : octavia | Creating users -------------------------- 9.00s",
                                "octavia : Get security groups for octavia ------------------------------- 7.17s",
                                "service-ks-register : octavia | Granting user roles --------------------- 7.06s",
                                "service-ks-register : octavia | Creating endpoints ---------------------- 6.86s",
                                "octavia : Copying certificate files for octavia-health-manager ---------- 6.73s",
                                "octavia : Create loadbalancer management subnet ------------------------- 6.70s",
                                "octavia : Copying over config.json files for services ------------------- 6.62s",
                                "octavia : Restart octavia-driver-agent container ------------------------ 6.53s",
                                "octavia : Restart octavia-worker container ------------------------------ 6.39s",
                                "octavia : Copying certificate files for octavia-housekeeping ------------ 6.16s",
                                "octavia : Create ports for Octavia health-manager nodes ----------------- 6.03s",
                                "service-cert-copy : octavia | Copying over extra CA certificates -------- 6.00s",
                                "\u001b[32m2026-02-24 01:17:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until refresh of running tasks\u001b[0m",
                                "\u001b[32m2026-02-24 01:17:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until refresh of running tasks\u001b[0m",
                                "\u001b[32m2026-02-24 01:17:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until refresh of running tasks\u001b[0m",
                                "\u001b[32m2026-02-24 01:17:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until refresh of running tasks\u001b[0m",
                                "\u001b[32m2026-02-24 01:18:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until refresh of running tasks\u001b[0m",
                                "\u001b[32m2026-02-24 01:18:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until refresh of running tasks\u001b[0m",
                                "\u001b[32m2026-02-24 01:18:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until refresh of running tasks\u001b[0m",
                                "\u001b[32m2026-02-24 01:18:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until refresh of running tasks\u001b[0m",
                                "\u001b[32m2026-02-24 01:18:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until refresh of running tasks\u001b[0m",
                                "\u001b[32m2026-02-24 01:18:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until refresh of running tasks\u001b[0m",
                                "\u001b[32m2026-02-24 01:18:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until refresh of running tasks\u001b[0m",
                                "\u001b[32m2026-02-24 01:18:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until refresh of running tasks\u001b[0m",
                                "\u001b[32m2026-02-24 01:18:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until refresh of running tasks\u001b[0m",
                                "\u001b[32m2026-02-24 01:18:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until refresh of running tasks\u001b[0m",
                                "\u001b[32m2026-02-24 01:18:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until refresh of running tasks\u001b[0m",
                                "\u001b[32m2026-02-24 01:18:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until refresh of running tasks\u001b[0m",
                                "\u001b[32m2026-02-24 01:18:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until refresh of running tasks\u001b[0m",
                                "\u001b[32m2026-02-24 01:18:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until refresh of running tasks\u001b[0m",
                                "\u001b[32m2026-02-24 01:18:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWait 1 second(s) until refresh of running tasks\u001b[0m",
                                "",
                                "",
                                "--> DEPLOY IN A NUTSHELL -- END -- Tue Feb 24 01:18:47 UTC 2026"
                            ],
                            "zuul_log_id": "0242ac1f-6606-1b2b-a63a-00000000001c-1-orchestrator"
                        }
                    },
                    "task": {
                        "duration": {
                            "end": "2026-02-24T01:18:47.710803Z",
                            "start": "2026-02-24T00:42:06.858515Z"
                        },
                        "id": "0242ac1f-6606-1b2b-a63a-00000000001c",
                        "name": "Deploy in a nutshell"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "ansible.builtin.command",
                            "changed": true,
                            "cmd": [
                                "ssh",
                                "-i",
                                "/home/zuul-testbed02/src/github.com/osism/testbed/terraform/.id_rsa.ci",
                                "dragon@81.163.192.195",
                                "/opt/configuration/scripts/bootstrap.sh"
                            ],
                            "delta": "0:03:25.689805",
                            "end": "2026-02-24 01:22:13.928872",
                            "invocation": {
                                "module_args": {
                                    "_raw_params": "ssh -i /home/zuul-testbed02/src/github.com/osism/testbed/terraform/.id_rsa.ci dragon@81.163.192.195 /opt/configuration/scripts/bootstrap.sh",
                                    "_uses_shell": false,
                                    "argv": null,
                                    "chdir": null,
                                    "creates": null,
                                    "executable": null,
                                    "expand_argument_vars": true,
                                    "removes": null,
                                    "stdin": null,
                                    "stdin_add_newline": true,
                                    "strip_empty_ends": true,
                                    "zuul_ansible_split_streams": false,
                                    "zuul_log_id": "0242ac1f-6606-1b2b-a63a-00000000001d-1-orchestrator",
                                    "zuul_output_max_bytes": 1073741824
                                }
                            },
                            "msg": "",
                            "rc": 0,
                            "start": "2026-02-24 01:18:48.239067",
                            "stderr": "",
                            "stderr_lines": [],
                            "stdout": "\n# BOOTSTRAP\n\n+ set -e\n+ echo\n+ echo '# BOOTSTRAP'\n+ echo\n+ sh -c /opt/configuration/scripts/bootstrap-services.sh\n+ set -e\n+ sh -c /opt/configuration/scripts/bootstrap/300-openstack.sh\n\u001b[32m2026-02-24 01:18:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task 1622eb3d-c11a-4807-80f7-a8df04effc93 (flavor-manager) has been started and output is visible here.\u001b[0m\n\u001b[32m2026-02-24 01:18:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mFlavor SCS-1L-1 created\u001b[0m\n\u001b[32m2026-02-24 01:18:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mFlavor SCS-1L-1-5 created\u001b[0m\n\u001b[32m2026-02-24 01:18:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mFlavor SCS-1V-2 created\u001b[0m\n\u001b[32m2026-02-24 01:18:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mFlavor SCS-1V-2-5 created\u001b[0m\n\u001b[32m2026-02-24 01:18:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mFlavor SCS-1V-4 created\u001b[0m\n\u001b[32m2026-02-24 01:18:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mFlavor SCS-1V-4-10 created\u001b[0m\n\u001b[32m2026-02-24 01:18:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mFlavor SCS-1V-8 created\u001b[0m\n\u001b[32m2026-02-24 01:18:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mFlavor SCS-1V-8-20 created\u001b[0m\n\u001b[32m2026-02-24 01:18:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mFlavor SCS-2V-4 created\u001b[0m\n\u001b[32m2026-02-24 01:18:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mFlavor SCS-2V-4-10 created\u001b[0m\n\u001b[32m2026-02-24 01:18:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mFlavor SCS-2V-8 created\u001b[0m\n\u001b[32m2026-02-24 01:18:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mFlavor SCS-2V-8-20 created\u001b[0m\n\u001b[32m2026-02-24 01:18:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mFlavor SCS-2V-16 created\u001b[0m\n\u001b[32m2026-02-24 01:18:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mFlavor SCS-2V-16-50 created\u001b[0m\n\u001b[32m2026-02-24 01:19:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mFlavor SCS-4V-8 created\u001b[0m\n\u001b[32m2026-02-24 01:19:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mFlavor SCS-4V-8-20 created\u001b[0m\n\u001b[32m2026-02-24 01:19:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mFlavor SCS-4V-16 created\u001b[0m\n\u001b[32m2026-02-24 01:19:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mFlavor SCS-4V-16-50 created\u001b[0m\n\u001b[32m2026-02-24 01:19:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mFlavor SCS-4V-32 created\u001b[0m\n\u001b[32m2026-02-24 01:19:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mFlavor SCS-4V-32-100 created\u001b[0m\n\u001b[32m2026-02-24 01:19:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mFlavor SCS-8V-16 created\u001b[0m\n\u001b[32m2026-02-24 01:19:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mFlavor SCS-8V-16-50 created\u001b[0m\n\u001b[32m2026-02-24 01:19:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mFlavor SCS-8V-32 created\u001b[0m\n\u001b[32m2026-02-24 01:19:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mFlavor SCS-8V-32-100 created\u001b[0m\n\u001b[32m2026-02-24 01:19:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mFlavor SCS-16V-32 created\u001b[0m\n\u001b[32m2026-02-24 01:19:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mFlavor SCS-16V-32-100 created\u001b[0m\n\u001b[32m2026-02-24 01:19:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mFlavor SCS-2V-4-20s created\u001b[0m\n\u001b[32m2026-02-24 01:19:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mFlavor SCS-4V-8-50s created\u001b[0m\n\u001b[32m2026-02-24 01:19:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mFlavor SCS-8V-32-100s created\u001b[0m\n\u001b[32m2026-02-24 01:19:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTrying to run play bootstrap-basic in environment openstack\u001b[0m\n\u001b[32m2026-02-24 01:19:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mPrepare task for execution of bootstrap-basic.\u001b[0m\n\u001b[32m2026-02-24 01:19:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 82896383-66d5-4d0c-8894-a5a03fb45b9b (bootstrap-basic) was prepared for execution.\u001b[0m\n\u001b[32m2026-02-24 01:19:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task 82896383-66d5-4d0c-8894-a5a03fb45b9b (bootstrap-basic) has been started and output is visible here.\u001b[0m\n\nPLAY [Bootstrap basic OpenStack services] **************************************\n\nTASK [Gathering Facts] *********************************************************\nTuesday 24 February 2026  01:19:20 +0000 (0:00:00.088)       0:00:00.088 ****** \n\u001b[0;32mok: [localhost]\u001b[0m\n\nTASK [Get volume type LUKS] ****************************************************\nTuesday 24 February 2026  01:19:23 +0000 (0:00:02.430)       0:00:02.519 ****** \n\u001b[0;32mok: [localhost]\u001b[0m\n\nTASK [Create volume type LUKS] *************************************************\nTuesday 24 February 2026  01:19:32 +0000 (0:00:09.307)       0:00:11.827 ****** \n\u001b[0;33mchanged: [localhost]\u001b[0m\n\nTASK [Create public network] ***************************************************\nTuesday 24 February 2026  01:19:41 +0000 (0:00:09.179)       0:00:21.007 ****** \n\u001b[0;33mchanged: [localhost]\u001b[0m\n\nTASK [Set public network to default] *******************************************\nTuesday 24 February 2026  01:19:47 +0000 (0:00:05.756)       0:00:26.764 ****** \n\u001b[0;33mchanged: [localhost]\u001b[0m\n\nTASK [Create public subnet] ****************************************************\nTuesday 24 February 2026  01:19:55 +0000 (0:00:07.786)       0:00:34.550 ****** \n\u001b[0;33mchanged: [localhost]\u001b[0m\n\nTASK [Create default IPv4 subnet pool] *****************************************\nTuesday 24 February 2026  01:20:00 +0000 (0:00:05.176)       0:00:39.726 ****** \n\u001b[0;33mchanged: [localhost]\u001b[0m\n\nTASK [Create manager role] *****************************************************\nTuesday 24 February 2026  01:20:04 +0000 (0:00:04.270)       0:00:43.997 ****** \n\u001b[0;32mok: [localhost]\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;33mlocalhost\u001b[0m                  : \u001b[0;32mok=8   \u001b[0m \u001b[0;33mchanged=5   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nTuesday 24 February 2026  01:20:08 +0000 (0:00:03.997)       0:00:47.995 ****** \n=============================================================================== \nGet volume type LUKS ---------------------------------------------------- 9.31s\nCreate volume type LUKS ------------------------------------------------- 9.18s\nSet public network to default ------------------------------------------- 7.79s\nCreate public network --------------------------------------------------- 5.76s\nCreate public subnet ---------------------------------------------------- 5.18s\nCreate default IPv4 subnet pool ----------------------------------------- 4.27s\nCreate manager role ----------------------------------------------------- 4.00s\nGathering Facts --------------------------------------------------------- 2.43s\n\u001b[32m2026-02-24 01:20:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task 88a6f34e-11d1-4ee3-b4dd-23abeb2c3966 (image-manager) has been started and output is visible here.\u001b[0m\n\u001b[32m2026-02-24 01:20:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mProcessing image 'Cirros 0.6.2'\u001b[0m\n\u001b[32m2026-02-24 01:20:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTested URL https://github.com/cirros-dev/cirros/releases/download/0.6.2/cirros-0.6.2-x86_64-disk.img: 302\u001b[0m\n\u001b[32m2026-02-24 01:20:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mImporting image Cirros 0.6.2\u001b[0m\n\u001b[32m2026-02-24 01:20:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mImporting from URL https://github.com/cirros-dev/cirros/releases/download/0.6.2/cirros-0.6.2-x86_64-disk.img\u001b[0m\n\u001b[32m2026-02-24 01:20:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWaiting for image to leave queued state...\u001b[0m\n\u001b[32m2026-02-24 01:20:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWaiting for import to complete...\u001b[0m\n\u001b[32m2026-02-24 01:20:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mImport of 'Cirros 0.6.2' successfully completed, reloading images\u001b[0m\n\u001b[32m2026-02-24 01:20:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mChecking parameters of 'Cirros 0.6.2'\u001b[0m\n\u001b[32m2026-02-24 01:20:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting internal_version = 0.6.2\u001b[0m\n\u001b[32m2026-02-24 01:20:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting image_original_user = cirros\u001b[0m\n\u001b[32m2026-02-24 01:20:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mAdding tag os:cirros\u001b[0m\n\u001b[32m2026-02-24 01:20:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property architecture: x86_64\u001b[0m\n\u001b[32m2026-02-24 01:20:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property hw_disk_bus: scsi\u001b[0m\n\u001b[32m2026-02-24 01:20:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property hw_rng_model: virtio\u001b[0m\n\u001b[32m2026-02-24 01:20:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property hw_scsi_model: virtio-scsi\u001b[0m\n\u001b[32m2026-02-24 01:20:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property hw_watchdog_action: reset\u001b[0m\n\u001b[32m2026-02-24 01:20:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property hypervisor_type: qemu\u001b[0m\n\u001b[32m2026-02-24 01:20:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property os_distro: cirros\u001b[0m\n\u001b[32m2026-02-24 01:20:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property os_purpose: minimal\u001b[0m\n\u001b[32m2026-02-24 01:20:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property replace_frequency: never\u001b[0m\n\u001b[32m2026-02-24 01:20:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property uuid_validity: none\u001b[0m\n\u001b[32m2026-02-24 01:20:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property provided_until: none\u001b[0m\n\u001b[32m2026-02-24 01:20:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property image_description: Cirros\u001b[0m\n\u001b[32m2026-02-24 01:20:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property image_name: Cirros\u001b[0m\n\u001b[32m2026-02-24 01:20:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property internal_version: 0.6.2\u001b[0m\n\u001b[32m2026-02-24 01:20:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property image_original_user: cirros\u001b[0m\n\u001b[32m2026-02-24 01:20:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property os_version: 0.6.2\u001b[0m\n\u001b[32m2026-02-24 01:20:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property image_source: https://github.com/cirros-dev/cirros/releases/download/0.6.2/cirros-0.6.2-x86_64-disk.img\u001b[0m\n\u001b[32m2026-02-24 01:20:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property image_build_date: 2023-05-30\u001b[0m\n\u001b[32m2026-02-24 01:20:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mChecking status of 'Cirros 0.6.2'\u001b[0m\n\u001b[32m2026-02-24 01:20:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mChecking visibility of 'Cirros 0.6.2'\u001b[0m\n\u001b[32m2026-02-24 01:20:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting visibility of 'Cirros 0.6.2' to 'public'\u001b[0m\n\u001b[32m2026-02-24 01:20:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mProcessing image 'Cirros 0.6.3'\u001b[0m\n\u001b[32m2026-02-24 01:20:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTested URL https://github.com/cirros-dev/cirros/releases/download/0.6.3/cirros-0.6.3-x86_64-disk.img: 302\u001b[0m\n\u001b[32m2026-02-24 01:20:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mImporting image Cirros 0.6.3\u001b[0m\n\u001b[32m2026-02-24 01:20:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mImporting from URL https://github.com/cirros-dev/cirros/releases/download/0.6.3/cirros-0.6.3-x86_64-disk.img\u001b[0m\n\u001b[32m2026-02-24 01:20:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWaiting for image to leave queued state...\u001b[0m\n\u001b[32m2026-02-24 01:20:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWaiting for import to complete...\u001b[0m\n\u001b[32m2026-02-24 01:20:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mImport of 'Cirros 0.6.3' successfully completed, reloading images\u001b[0m\n\u001b[32m2026-02-24 01:20:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mChecking parameters of 'Cirros 0.6.3'\u001b[0m\n\u001b[32m2026-02-24 01:20:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting internal_version = 0.6.3\u001b[0m\n\u001b[32m2026-02-24 01:20:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting image_original_user = cirros\u001b[0m\n\u001b[32m2026-02-24 01:20:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mAdding tag os:cirros\u001b[0m\n\u001b[32m2026-02-24 01:20:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property architecture: x86_64\u001b[0m\n\u001b[32m2026-02-24 01:20:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property hw_disk_bus: scsi\u001b[0m\n\u001b[32m2026-02-24 01:20:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property hw_rng_model: virtio\u001b[0m\n\u001b[32m2026-02-24 01:20:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property hw_scsi_model: virtio-scsi\u001b[0m\n\u001b[32m2026-02-24 01:20:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property hw_watchdog_action: reset\u001b[0m\n\u001b[32m2026-02-24 01:20:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property hypervisor_type: qemu\u001b[0m\n\u001b[32m2026-02-24 01:20:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property os_distro: cirros\u001b[0m\n\u001b[32m2026-02-24 01:20:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property os_purpose: minimal\u001b[0m\n\u001b[32m2026-02-24 01:20:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property replace_frequency: never\u001b[0m\n\u001b[32m2026-02-24 01:20:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property uuid_validity: none\u001b[0m\n\u001b[32m2026-02-24 01:20:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property provided_until: none\u001b[0m\n\u001b[32m2026-02-24 01:20:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property image_description: Cirros\u001b[0m\n\u001b[32m2026-02-24 01:20:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property image_name: Cirros\u001b[0m\n\u001b[32m2026-02-24 01:20:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property internal_version: 0.6.3\u001b[0m\n\u001b[32m2026-02-24 01:20:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property image_original_user: cirros\u001b[0m\n\u001b[32m2026-02-24 01:20:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property os_version: 0.6.3\u001b[0m\n\u001b[32m2026-02-24 01:20:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property image_source: https://github.com/cirros-dev/cirros/releases/download/0.6.3/cirros-0.6.3-x86_64-disk.img\u001b[0m\n\u001b[32m2026-02-24 01:20:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property image_build_date: 2024-09-26\u001b[0m\n\u001b[32m2026-02-24 01:20:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mChecking status of 'Cirros 0.6.3'\u001b[0m\n\u001b[32m2026-02-24 01:20:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mChecking visibility of 'Cirros 0.6.3'\u001b[0m\n\u001b[32m2026-02-24 01:20:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting visibility of 'Cirros 0.6.3' to 'public'\u001b[0m\n+ sh -c /opt/configuration/scripts/bootstrap/301-openstack-octavia-amhpora-image.sh\n\u001b[32m2026-02-24 01:21:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mdate: 2026-02-23\u001b[0m\n\u001b[32m2026-02-24 01:21:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mimage: octavia-amphora-haproxy-2024.2.20260223.qcow2\u001b[0m\n\u001b[32m2026-02-24 01:21:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1murl: https://nbg1.your-objectstorage.com/osism/openstack-octavia-amphora-image/octavia-amphora-haproxy-2024.2.20260223.qcow2\u001b[0m\n\u001b[32m2026-02-24 01:21:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mchecksum_url: https://nbg1.your-objectstorage.com/osism/openstack-octavia-amphora-image/octavia-amphora-haproxy-2024.2.20260223.qcow2.CHECKSUM\u001b[0m\n\u001b[32m2026-02-24 01:21:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mchecksum: 8fa1c871cfb41c51427d243d56beff3796ad151e7381526e195fa2edff653330\u001b[0m\n\u001b[32m2026-02-24 01:21:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task 7553e2cd-a545-433c-af4d-91b53ff89f84 (image-manager) has been started and output is visible here.\u001b[0m\n\u001b[32m2026-02-24 01:21:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mProcessing image 'OpenStack Octavia Amphora 2026-02-23'\u001b[0m\n\u001b[32m2026-02-24 01:21:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTested URL https://nbg1.your-objectstorage.com/osism/openstack-octavia-amphora-image/octavia-amphora-haproxy-2024.2.20260223.qcow2: 200\u001b[0m\n\u001b[32m2026-02-24 01:21:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mImporting image OpenStack Octavia Amphora 2026-02-23\u001b[0m\n\u001b[32m2026-02-24 01:21:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mImporting from URL https://nbg1.your-objectstorage.com/osism/openstack-octavia-amphora-image/octavia-amphora-haproxy-2024.2.20260223.qcow2\u001b[0m\n\u001b[32m2026-02-24 01:21:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWaiting for image to leave queued state...\u001b[0m\n\u001b[32m2026-02-24 01:21:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWaiting for import to complete...\u001b[0m\n\u001b[32m2026-02-24 01:21:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWaiting for import to complete...\u001b[0m\n\u001b[32m2026-02-24 01:21:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWaiting for import to complete...\u001b[0m\n\u001b[32m2026-02-24 01:21:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWaiting for import to complete...\u001b[0m\n\u001b[32m2026-02-24 01:21:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWaiting for import to complete...\u001b[0m\n\u001b[32m2026-02-24 01:21:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWaiting for import to complete...\u001b[0m\n\u001b[32m2026-02-24 01:22:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mImport of 'OpenStack Octavia Amphora 2026-02-23' successfully completed, reloading images\u001b[0m\n\u001b[32m2026-02-24 01:22:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mChecking parameters of 'OpenStack Octavia Amphora 2026-02-23'\u001b[0m\n\u001b[32m2026-02-24 01:22:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting internal_version = 2026-02-23\u001b[0m\n\u001b[32m2026-02-24 01:22:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting image_original_user = ubuntu\u001b[0m\n\u001b[32m2026-02-24 01:22:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mAdding tag amphora\u001b[0m\n\u001b[32m2026-02-24 01:22:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mAdding tag os:ubuntu\u001b[0m\n\u001b[32m2026-02-24 01:22:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property architecture: x86_64\u001b[0m\n\u001b[32m2026-02-24 01:22:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property hw_disk_bus: scsi\u001b[0m\n\u001b[32m2026-02-24 01:22:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property hw_rng_model: virtio\u001b[0m\n\u001b[32m2026-02-24 01:22:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property hw_scsi_model: virtio-scsi\u001b[0m\n\u001b[32m2026-02-24 01:22:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property hw_watchdog_action: reset\u001b[0m\n\u001b[32m2026-02-24 01:22:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property hypervisor_type: qemu\u001b[0m\n\u001b[32m2026-02-24 01:22:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property os_distro: ubuntu\u001b[0m\n\u001b[32m2026-02-24 01:22:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property replace_frequency: quarterly\u001b[0m\n\u001b[32m2026-02-24 01:22:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property uuid_validity: last-1\u001b[0m\n\u001b[32m2026-02-24 01:22:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property provided_until: none\u001b[0m\n\u001b[32m2026-02-24 01:22:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property os_purpose: network\u001b[0m\n\u001b[32m2026-02-24 01:22:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property image_description: OpenStack Octavia Amphora\u001b[0m\n\u001b[32m2026-02-24 01:22:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property image_name: OpenStack Octavia Amphora\u001b[0m\n\u001b[32m2026-02-24 01:22:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property internal_version: 2026-02-23\u001b[0m\n\u001b[32m2026-02-24 01:22:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property image_original_user: ubuntu\u001b[0m\n\u001b[32m2026-02-24 01:22:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property os_version: 2026-02-23\u001b[0m\n\u001b[32m2026-02-24 01:22:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property image_source: https://nbg1.your-objectstorage.com/osism/openstack-octavia-amphora-image/octavia-amphora-haproxy-2024.2.20260223.qcow2\u001b[0m\n\u001b[32m2026-02-24 01:22:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property image_build_date: 2026-02-23\u001b[0m\n\u001b[32m2026-02-24 01:22:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mChecking status of 'OpenStack Octavia Amphora 2026-02-23'\u001b[0m\n\u001b[32m2026-02-24 01:22:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mChecking visibility of 'OpenStack Octavia Amphora 2026-02-23'\u001b[0m\n\u001b[32m2026-02-24 01:22:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mProcessing image 'Cirros 0.6.3' (removal candidate)\u001b[0m\n\u001b[32m2026-02-24 01:22:13\u001b[0m | \u001b[33m\u001b[1mWARNING \u001b[0m | \u001b[33m\u001b[1mNo image definition found for 'Cirros 0.6.3', image will be ignored\u001b[0m\n\u001b[32m2026-02-24 01:22:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mProcessing image 'Cirros 0.6.2' (removal candidate)\u001b[0m\n\u001b[32m2026-02-24 01:22:13\u001b[0m | \u001b[33m\u001b[1mWARNING \u001b[0m | \u001b[33m\u001b[1mNo image definition found for 'Cirros 0.6.2', image will be ignored\u001b[0m",
                            "stdout_lines": [
                                "",
                                "# BOOTSTRAP",
                                "",
                                "+ set -e",
                                "+ echo",
                                "+ echo '# BOOTSTRAP'",
                                "+ echo",
                                "+ sh -c /opt/configuration/scripts/bootstrap-services.sh",
                                "+ set -e",
                                "+ sh -c /opt/configuration/scripts/bootstrap/300-openstack.sh",
                                "\u001b[32m2026-02-24 01:18:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task 1622eb3d-c11a-4807-80f7-a8df04effc93 (flavor-manager) has been started and output is visible here.\u001b[0m",
                                "\u001b[32m2026-02-24 01:18:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mFlavor SCS-1L-1 created\u001b[0m",
                                "\u001b[32m2026-02-24 01:18:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mFlavor SCS-1L-1-5 created\u001b[0m",
                                "\u001b[32m2026-02-24 01:18:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mFlavor SCS-1V-2 created\u001b[0m",
                                "\u001b[32m2026-02-24 01:18:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mFlavor SCS-1V-2-5 created\u001b[0m",
                                "\u001b[32m2026-02-24 01:18:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mFlavor SCS-1V-4 created\u001b[0m",
                                "\u001b[32m2026-02-24 01:18:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mFlavor SCS-1V-4-10 created\u001b[0m",
                                "\u001b[32m2026-02-24 01:18:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mFlavor SCS-1V-8 created\u001b[0m",
                                "\u001b[32m2026-02-24 01:18:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mFlavor SCS-1V-8-20 created\u001b[0m",
                                "\u001b[32m2026-02-24 01:18:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mFlavor SCS-2V-4 created\u001b[0m",
                                "\u001b[32m2026-02-24 01:18:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mFlavor SCS-2V-4-10 created\u001b[0m",
                                "\u001b[32m2026-02-24 01:18:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mFlavor SCS-2V-8 created\u001b[0m",
                                "\u001b[32m2026-02-24 01:18:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mFlavor SCS-2V-8-20 created\u001b[0m",
                                "\u001b[32m2026-02-24 01:18:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mFlavor SCS-2V-16 created\u001b[0m",
                                "\u001b[32m2026-02-24 01:18:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mFlavor SCS-2V-16-50 created\u001b[0m",
                                "\u001b[32m2026-02-24 01:19:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mFlavor SCS-4V-8 created\u001b[0m",
                                "\u001b[32m2026-02-24 01:19:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mFlavor SCS-4V-8-20 created\u001b[0m",
                                "\u001b[32m2026-02-24 01:19:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mFlavor SCS-4V-16 created\u001b[0m",
                                "\u001b[32m2026-02-24 01:19:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mFlavor SCS-4V-16-50 created\u001b[0m",
                                "\u001b[32m2026-02-24 01:19:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mFlavor SCS-4V-32 created\u001b[0m",
                                "\u001b[32m2026-02-24 01:19:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mFlavor SCS-4V-32-100 created\u001b[0m",
                                "\u001b[32m2026-02-24 01:19:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mFlavor SCS-8V-16 created\u001b[0m",
                                "\u001b[32m2026-02-24 01:19:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mFlavor SCS-8V-16-50 created\u001b[0m",
                                "\u001b[32m2026-02-24 01:19:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mFlavor SCS-8V-32 created\u001b[0m",
                                "\u001b[32m2026-02-24 01:19:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mFlavor SCS-8V-32-100 created\u001b[0m",
                                "\u001b[32m2026-02-24 01:19:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mFlavor SCS-16V-32 created\u001b[0m",
                                "\u001b[32m2026-02-24 01:19:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mFlavor SCS-16V-32-100 created\u001b[0m",
                                "\u001b[32m2026-02-24 01:19:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mFlavor SCS-2V-4-20s created\u001b[0m",
                                "\u001b[32m2026-02-24 01:19:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mFlavor SCS-4V-8-50s created\u001b[0m",
                                "\u001b[32m2026-02-24 01:19:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mFlavor SCS-8V-32-100s created\u001b[0m",
                                "\u001b[32m2026-02-24 01:19:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTrying to run play bootstrap-basic in environment openstack\u001b[0m",
                                "\u001b[32m2026-02-24 01:19:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mPrepare task for execution of bootstrap-basic.\u001b[0m",
                                "\u001b[32m2026-02-24 01:19:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 82896383-66d5-4d0c-8894-a5a03fb45b9b (bootstrap-basic) was prepared for execution.\u001b[0m",
                                "\u001b[32m2026-02-24 01:19:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task 82896383-66d5-4d0c-8894-a5a03fb45b9b (bootstrap-basic) has been started and output is visible here.\u001b[0m",
                                "",
                                "PLAY [Bootstrap basic OpenStack services] **************************************",
                                "",
                                "TASK [Gathering Facts] *********************************************************",
                                "Tuesday 24 February 2026  01:19:20 +0000 (0:00:00.088)       0:00:00.088 ****** ",
                                "\u001b[0;32mok: [localhost]\u001b[0m",
                                "",
                                "TASK [Get volume type LUKS] ****************************************************",
                                "Tuesday 24 February 2026  01:19:23 +0000 (0:00:02.430)       0:00:02.519 ****** ",
                                "\u001b[0;32mok: [localhost]\u001b[0m",
                                "",
                                "TASK [Create volume type LUKS] *************************************************",
                                "Tuesday 24 February 2026  01:19:32 +0000 (0:00:09.307)       0:00:11.827 ****** ",
                                "\u001b[0;33mchanged: [localhost]\u001b[0m",
                                "",
                                "TASK [Create public network] ***************************************************",
                                "Tuesday 24 February 2026  01:19:41 +0000 (0:00:09.179)       0:00:21.007 ****** ",
                                "\u001b[0;33mchanged: [localhost]\u001b[0m",
                                "",
                                "TASK [Set public network to default] *******************************************",
                                "Tuesday 24 February 2026  01:19:47 +0000 (0:00:05.756)       0:00:26.764 ****** ",
                                "\u001b[0;33mchanged: [localhost]\u001b[0m",
                                "",
                                "TASK [Create public subnet] ****************************************************",
                                "Tuesday 24 February 2026  01:19:55 +0000 (0:00:07.786)       0:00:34.550 ****** ",
                                "\u001b[0;33mchanged: [localhost]\u001b[0m",
                                "",
                                "TASK [Create default IPv4 subnet pool] *****************************************",
                                "Tuesday 24 February 2026  01:20:00 +0000 (0:00:05.176)       0:00:39.726 ****** ",
                                "\u001b[0;33mchanged: [localhost]\u001b[0m",
                                "",
                                "TASK [Create manager role] *****************************************************",
                                "Tuesday 24 February 2026  01:20:04 +0000 (0:00:04.270)       0:00:43.997 ****** ",
                                "\u001b[0;32mok: [localhost]\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;33mlocalhost\u001b[0m                  : \u001b[0;32mok=8   \u001b[0m \u001b[0;33mchanged=5   \u001b[0m unreachable=0    failed=0    skipped=0    rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Tuesday 24 February 2026  01:20:08 +0000 (0:00:03.997)       0:00:47.995 ****** ",
                                "=============================================================================== ",
                                "Get volume type LUKS ---------------------------------------------------- 9.31s",
                                "Create volume type LUKS ------------------------------------------------- 9.18s",
                                "Set public network to default ------------------------------------------- 7.79s",
                                "Create public network --------------------------------------------------- 5.76s",
                                "Create public subnet ---------------------------------------------------- 5.18s",
                                "Create default IPv4 subnet pool ----------------------------------------- 4.27s",
                                "Create manager role ----------------------------------------------------- 4.00s",
                                "Gathering Facts --------------------------------------------------------- 2.43s",
                                "\u001b[32m2026-02-24 01:20:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task 88a6f34e-11d1-4ee3-b4dd-23abeb2c3966 (image-manager) has been started and output is visible here.\u001b[0m",
                                "\u001b[32m2026-02-24 01:20:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mProcessing image 'Cirros 0.6.2'\u001b[0m",
                                "\u001b[32m2026-02-24 01:20:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTested URL https://github.com/cirros-dev/cirros/releases/download/0.6.2/cirros-0.6.2-x86_64-disk.img: 302\u001b[0m",
                                "\u001b[32m2026-02-24 01:20:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mImporting image Cirros 0.6.2\u001b[0m",
                                "\u001b[32m2026-02-24 01:20:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mImporting from URL https://github.com/cirros-dev/cirros/releases/download/0.6.2/cirros-0.6.2-x86_64-disk.img\u001b[0m",
                                "\u001b[32m2026-02-24 01:20:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWaiting for image to leave queued state...\u001b[0m",
                                "\u001b[32m2026-02-24 01:20:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWaiting for import to complete...\u001b[0m",
                                "\u001b[32m2026-02-24 01:20:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mImport of 'Cirros 0.6.2' successfully completed, reloading images\u001b[0m",
                                "\u001b[32m2026-02-24 01:20:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mChecking parameters of 'Cirros 0.6.2'\u001b[0m",
                                "\u001b[32m2026-02-24 01:20:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting internal_version = 0.6.2\u001b[0m",
                                "\u001b[32m2026-02-24 01:20:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting image_original_user = cirros\u001b[0m",
                                "\u001b[32m2026-02-24 01:20:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mAdding tag os:cirros\u001b[0m",
                                "\u001b[32m2026-02-24 01:20:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property architecture: x86_64\u001b[0m",
                                "\u001b[32m2026-02-24 01:20:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property hw_disk_bus: scsi\u001b[0m",
                                "\u001b[32m2026-02-24 01:20:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property hw_rng_model: virtio\u001b[0m",
                                "\u001b[32m2026-02-24 01:20:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property hw_scsi_model: virtio-scsi\u001b[0m",
                                "\u001b[32m2026-02-24 01:20:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property hw_watchdog_action: reset\u001b[0m",
                                "\u001b[32m2026-02-24 01:20:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property hypervisor_type: qemu\u001b[0m",
                                "\u001b[32m2026-02-24 01:20:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property os_distro: cirros\u001b[0m",
                                "\u001b[32m2026-02-24 01:20:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property os_purpose: minimal\u001b[0m",
                                "\u001b[32m2026-02-24 01:20:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property replace_frequency: never\u001b[0m",
                                "\u001b[32m2026-02-24 01:20:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property uuid_validity: none\u001b[0m",
                                "\u001b[32m2026-02-24 01:20:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property provided_until: none\u001b[0m",
                                "\u001b[32m2026-02-24 01:20:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property image_description: Cirros\u001b[0m",
                                "\u001b[32m2026-02-24 01:20:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property image_name: Cirros\u001b[0m",
                                "\u001b[32m2026-02-24 01:20:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property internal_version: 0.6.2\u001b[0m",
                                "\u001b[32m2026-02-24 01:20:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property image_original_user: cirros\u001b[0m",
                                "\u001b[32m2026-02-24 01:20:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property os_version: 0.6.2\u001b[0m",
                                "\u001b[32m2026-02-24 01:20:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property image_source: https://github.com/cirros-dev/cirros/releases/download/0.6.2/cirros-0.6.2-x86_64-disk.img\u001b[0m",
                                "\u001b[32m2026-02-24 01:20:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property image_build_date: 2023-05-30\u001b[0m",
                                "\u001b[32m2026-02-24 01:20:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mChecking status of 'Cirros 0.6.2'\u001b[0m",
                                "\u001b[32m2026-02-24 01:20:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mChecking visibility of 'Cirros 0.6.2'\u001b[0m",
                                "\u001b[32m2026-02-24 01:20:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting visibility of 'Cirros 0.6.2' to 'public'\u001b[0m",
                                "\u001b[32m2026-02-24 01:20:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mProcessing image 'Cirros 0.6.3'\u001b[0m",
                                "\u001b[32m2026-02-24 01:20:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTested URL https://github.com/cirros-dev/cirros/releases/download/0.6.3/cirros-0.6.3-x86_64-disk.img: 302\u001b[0m",
                                "\u001b[32m2026-02-24 01:20:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mImporting image Cirros 0.6.3\u001b[0m",
                                "\u001b[32m2026-02-24 01:20:35\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mImporting from URL https://github.com/cirros-dev/cirros/releases/download/0.6.3/cirros-0.6.3-x86_64-disk.img\u001b[0m",
                                "\u001b[32m2026-02-24 01:20:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWaiting for image to leave queued state...\u001b[0m",
                                "\u001b[32m2026-02-24 01:20:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWaiting for import to complete...\u001b[0m",
                                "\u001b[32m2026-02-24 01:20:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mImport of 'Cirros 0.6.3' successfully completed, reloading images\u001b[0m",
                                "\u001b[32m2026-02-24 01:20:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mChecking parameters of 'Cirros 0.6.3'\u001b[0m",
                                "\u001b[32m2026-02-24 01:20:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting internal_version = 0.6.3\u001b[0m",
                                "\u001b[32m2026-02-24 01:20:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting image_original_user = cirros\u001b[0m",
                                "\u001b[32m2026-02-24 01:20:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mAdding tag os:cirros\u001b[0m",
                                "\u001b[32m2026-02-24 01:20:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property architecture: x86_64\u001b[0m",
                                "\u001b[32m2026-02-24 01:20:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property hw_disk_bus: scsi\u001b[0m",
                                "\u001b[32m2026-02-24 01:20:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property hw_rng_model: virtio\u001b[0m",
                                "\u001b[32m2026-02-24 01:20:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property hw_scsi_model: virtio-scsi\u001b[0m",
                                "\u001b[32m2026-02-24 01:20:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property hw_watchdog_action: reset\u001b[0m",
                                "\u001b[32m2026-02-24 01:20:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property hypervisor_type: qemu\u001b[0m",
                                "\u001b[32m2026-02-24 01:20:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property os_distro: cirros\u001b[0m",
                                "\u001b[32m2026-02-24 01:20:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property os_purpose: minimal\u001b[0m",
                                "\u001b[32m2026-02-24 01:20:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property replace_frequency: never\u001b[0m",
                                "\u001b[32m2026-02-24 01:20:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property uuid_validity: none\u001b[0m",
                                "\u001b[32m2026-02-24 01:20:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property provided_until: none\u001b[0m",
                                "\u001b[32m2026-02-24 01:20:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property image_description: Cirros\u001b[0m",
                                "\u001b[32m2026-02-24 01:20:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property image_name: Cirros\u001b[0m",
                                "\u001b[32m2026-02-24 01:20:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property internal_version: 0.6.3\u001b[0m",
                                "\u001b[32m2026-02-24 01:20:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property image_original_user: cirros\u001b[0m",
                                "\u001b[32m2026-02-24 01:20:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property os_version: 0.6.3\u001b[0m",
                                "\u001b[32m2026-02-24 01:20:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property image_source: https://github.com/cirros-dev/cirros/releases/download/0.6.3/cirros-0.6.3-x86_64-disk.img\u001b[0m",
                                "\u001b[32m2026-02-24 01:20:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property image_build_date: 2024-09-26\u001b[0m",
                                "\u001b[32m2026-02-24 01:20:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mChecking status of 'Cirros 0.6.3'\u001b[0m",
                                "\u001b[32m2026-02-24 01:20:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mChecking visibility of 'Cirros 0.6.3'\u001b[0m",
                                "\u001b[32m2026-02-24 01:20:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting visibility of 'Cirros 0.6.3' to 'public'\u001b[0m",
                                "+ sh -c /opt/configuration/scripts/bootstrap/301-openstack-octavia-amhpora-image.sh",
                                "\u001b[32m2026-02-24 01:21:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mdate: 2026-02-23\u001b[0m",
                                "\u001b[32m2026-02-24 01:21:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mimage: octavia-amphora-haproxy-2024.2.20260223.qcow2\u001b[0m",
                                "\u001b[32m2026-02-24 01:21:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1murl: https://nbg1.your-objectstorage.com/osism/openstack-octavia-amphora-image/octavia-amphora-haproxy-2024.2.20260223.qcow2\u001b[0m",
                                "\u001b[32m2026-02-24 01:21:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mchecksum_url: https://nbg1.your-objectstorage.com/osism/openstack-octavia-amphora-image/octavia-amphora-haproxy-2024.2.20260223.qcow2.CHECKSUM\u001b[0m",
                                "\u001b[32m2026-02-24 01:21:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mchecksum: 8fa1c871cfb41c51427d243d56beff3796ad151e7381526e195fa2edff653330\u001b[0m",
                                "\u001b[32m2026-02-24 01:21:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task 7553e2cd-a545-433c-af4d-91b53ff89f84 (image-manager) has been started and output is visible here.\u001b[0m",
                                "\u001b[32m2026-02-24 01:21:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mProcessing image 'OpenStack Octavia Amphora 2026-02-23'\u001b[0m",
                                "\u001b[32m2026-02-24 01:21:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTested URL https://nbg1.your-objectstorage.com/osism/openstack-octavia-amphora-image/octavia-amphora-haproxy-2024.2.20260223.qcow2: 200\u001b[0m",
                                "\u001b[32m2026-02-24 01:21:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mImporting image OpenStack Octavia Amphora 2026-02-23\u001b[0m",
                                "\u001b[32m2026-02-24 01:21:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mImporting from URL https://nbg1.your-objectstorage.com/osism/openstack-octavia-amphora-image/octavia-amphora-haproxy-2024.2.20260223.qcow2\u001b[0m",
                                "\u001b[32m2026-02-24 01:21:04\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWaiting for image to leave queued state...\u001b[0m",
                                "\u001b[32m2026-02-24 01:21:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWaiting for import to complete...\u001b[0m",
                                "\u001b[32m2026-02-24 01:21:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWaiting for import to complete...\u001b[0m",
                                "\u001b[32m2026-02-24 01:21:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWaiting for import to complete...\u001b[0m",
                                "\u001b[32m2026-02-24 01:21:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWaiting for import to complete...\u001b[0m",
                                "\u001b[32m2026-02-24 01:21:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWaiting for import to complete...\u001b[0m",
                                "\u001b[32m2026-02-24 01:21:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mWaiting for import to complete...\u001b[0m",
                                "\u001b[32m2026-02-24 01:22:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mImport of 'OpenStack Octavia Amphora 2026-02-23' successfully completed, reloading images\u001b[0m",
                                "\u001b[32m2026-02-24 01:22:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mChecking parameters of 'OpenStack Octavia Amphora 2026-02-23'\u001b[0m",
                                "\u001b[32m2026-02-24 01:22:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting internal_version = 2026-02-23\u001b[0m",
                                "\u001b[32m2026-02-24 01:22:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting image_original_user = ubuntu\u001b[0m",
                                "\u001b[32m2026-02-24 01:22:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mAdding tag amphora\u001b[0m",
                                "\u001b[32m2026-02-24 01:22:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mAdding tag os:ubuntu\u001b[0m",
                                "\u001b[32m2026-02-24 01:22:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property architecture: x86_64\u001b[0m",
                                "\u001b[32m2026-02-24 01:22:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property hw_disk_bus: scsi\u001b[0m",
                                "\u001b[32m2026-02-24 01:22:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property hw_rng_model: virtio\u001b[0m",
                                "\u001b[32m2026-02-24 01:22:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property hw_scsi_model: virtio-scsi\u001b[0m",
                                "\u001b[32m2026-02-24 01:22:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property hw_watchdog_action: reset\u001b[0m",
                                "\u001b[32m2026-02-24 01:22:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property hypervisor_type: qemu\u001b[0m",
                                "\u001b[32m2026-02-24 01:22:09\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property os_distro: ubuntu\u001b[0m",
                                "\u001b[32m2026-02-24 01:22:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property replace_frequency: quarterly\u001b[0m",
                                "\u001b[32m2026-02-24 01:22:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property uuid_validity: last-1\u001b[0m",
                                "\u001b[32m2026-02-24 01:22:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property provided_until: none\u001b[0m",
                                "\u001b[32m2026-02-24 01:22:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property os_purpose: network\u001b[0m",
                                "\u001b[32m2026-02-24 01:22:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property image_description: OpenStack Octavia Amphora\u001b[0m",
                                "\u001b[32m2026-02-24 01:22:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property image_name: OpenStack Octavia Amphora\u001b[0m",
                                "\u001b[32m2026-02-24 01:22:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property internal_version: 2026-02-23\u001b[0m",
                                "\u001b[32m2026-02-24 01:22:11\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property image_original_user: ubuntu\u001b[0m",
                                "\u001b[32m2026-02-24 01:22:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property os_version: 2026-02-23\u001b[0m",
                                "\u001b[32m2026-02-24 01:22:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property image_source: https://nbg1.your-objectstorage.com/osism/openstack-octavia-amphora-image/octavia-amphora-haproxy-2024.2.20260223.qcow2\u001b[0m",
                                "\u001b[32m2026-02-24 01:22:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSetting property image_build_date: 2026-02-23\u001b[0m",
                                "\u001b[32m2026-02-24 01:22:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mChecking status of 'OpenStack Octavia Amphora 2026-02-23'\u001b[0m",
                                "\u001b[32m2026-02-24 01:22:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mChecking visibility of 'OpenStack Octavia Amphora 2026-02-23'\u001b[0m",
                                "\u001b[32m2026-02-24 01:22:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mProcessing image 'Cirros 0.6.3' (removal candidate)\u001b[0m",
                                "\u001b[32m2026-02-24 01:22:13\u001b[0m | \u001b[33m\u001b[1mWARNING \u001b[0m | \u001b[33m\u001b[1mNo image definition found for 'Cirros 0.6.3', image will be ignored\u001b[0m",
                                "\u001b[32m2026-02-24 01:22:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mProcessing image 'Cirros 0.6.2' (removal candidate)\u001b[0m",
                                "\u001b[32m2026-02-24 01:22:13\u001b[0m | \u001b[33m\u001b[1mWARNING \u001b[0m | \u001b[33m\u001b[1mNo image definition found for 'Cirros 0.6.2', image will be ignored\u001b[0m"
                            ],
                            "zuul_log_id": "0242ac1f-6606-1b2b-a63a-00000000001d-1-orchestrator"
                        }
                    },
                    "task": {
                        "duration": {
                            "end": "2026-02-24T01:22:14.045683Z",
                            "start": "2026-02-24T01:18:47.872288Z"
                        },
                        "id": "0242ac1f-6606-1b2b-a63a-00000000001d",
                        "name": "Bootstrap services"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "ansible.builtin.command",
                            "changed": true,
                            "cmd": [
                                "ssh",
                                "-i",
                                "/home/zuul-testbed02/src/github.com/osism/testbed/terraform/.id_rsa.ci",
                                "dragon@81.163.192.195",
                                "/opt/configuration/scripts/check.sh"
                            ],
                            "delta": "0:17:19.862231",
                            "end": "2026-02-24 01:39:34.236455",
                            "invocation": {
                                "module_args": {
                                    "_raw_params": "ssh -i /home/zuul-testbed02/src/github.com/osism/testbed/terraform/.id_rsa.ci dragon@81.163.192.195 /opt/configuration/scripts/check.sh",
                                    "_uses_shell": false,
                                    "argv": null,
                                    "chdir": null,
                                    "creates": null,
                                    "executable": null,
                                    "expand_argument_vars": true,
                                    "removes": null,
                                    "stdin": null,
                                    "stdin_add_newline": true,
                                    "strip_empty_ends": true,
                                    "zuul_ansible_split_streams": false,
                                    "zuul_log_id": "0242ac1f-6606-1b2b-a63a-00000000001e-1-orchestrator",
                                    "zuul_output_max_bytes": 1073741824
                                }
                            },
                            "msg": "",
                            "rc": 0,
                            "start": "2026-02-24 01:22:14.374224",
                            "stderr": "",
                            "stderr_lines": [],
                            "stdout": "+ set -e\n+ source /opt/configuration/scripts/include.sh\n++ export INTERACTIVE=false\n++ INTERACTIVE=false\n++ export OSISM_APPLY_RETRY=1\n++ OSISM_APPLY_RETRY=1\n+ source /opt/configuration/scripts/manager-version.sh\n+++ awk '-F: ' '/^manager_version:/ { print $2 }' /opt/configuration/environments/manager/configuration.yml\n\n# CHECK\n\n++ export MANAGER_VERSION=latest\n++ MANAGER_VERSION=latest\n+ echo\n+ echo '# CHECK'\n+ echo\n+ for node in testbed-manager testbed-node-0 testbed-node-1 testbed-node-2\n++ semver latest 5.0.0\n\n## Containers @ testbed-manager\n\n+ [[ -1 -eq -1 ]]\n+ [[ latest != \\l\\a\\t\\e\\s\\t ]]\n+ echo\n+ echo '## Containers @ testbed-manager'\n+ echo\n+ osism container testbed-manager ps\n\u001b[32m2026-02-24 01:22:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mCreating empty known_hosts file: /share/known_hosts\u001b[0m\nCONTAINER ID   IMAGE                                                           COMMAND                  CREATED             STATUS                       PORTS                                                                             NAMES\ndee2ca207554   registry.osism.tech/kolla/prometheus-blackbox-exporter:2024.2   \"dumb-init --single-\u2026\"   12 minutes ago      Up 12 minutes                                                                                                  prometheus_blackbox_exporter\ne96a2bb2a6f6   registry.osism.tech/kolla/prometheus-alertmanager:2024.2        \"dumb-init --single-\u2026\"   12 minutes ago      Up 12 minutes                                                                                                  prometheus_alertmanager\ne7894d6147dc   registry.osism.tech/kolla/prometheus-cadvisor:2024.2            \"dumb-init --single-\u2026\"   13 minutes ago      Up 13 minutes                                                                                                  prometheus_cadvisor\neb5b37ad51bd   registry.osism.tech/kolla/prometheus-node-exporter:2024.2       \"dumb-init --single-\u2026\"   13 minutes ago      Up 13 minutes                                                                                                  prometheus_node_exporter\nd0d10a5ed6b3   registry.osism.tech/kolla/prometheus-v2-server:2024.2           \"dumb-init --single-\u2026\"   13 minutes ago      Up 13 minutes                                                                                                  prometheus_server\n420b7e422cda   registry.osism.tech/osism/cephclient:reef                       \"/usr/bin/dumb-init \u2026\"   19 minutes ago      Up 19 minutes                                                                                                  cephclient\n0a0ceb50f3d9   registry.osism.tech/kolla/cron:2024.2                           \"dumb-init --single-\u2026\"   32 minutes ago      Up 32 minutes                                                                                                  cron\nb6386485e219   registry.osism.tech/kolla/kolla-toolbox:2024.2                  \"dumb-init --single-\u2026\"   33 minutes ago      Up 32 minutes                                                                                                  kolla_toolbox\n7cfd91f0f86a   registry.osism.tech/kolla/fluentd:2024.2                        \"dumb-init --single-\u2026\"   33 minutes ago      Up 33 minutes                                                                                                  fluentd\ndcad4d685626   phpmyadmin/phpmyadmin:5.2                                       \"/docker-entrypoint.\u2026\"   34 minutes ago      Up 33 minutes (healthy)      80/tcp                                                                            phpmyadmin\n513977b148a1   registry.osism.tech/osism/openstackclient:2024.2                \"/usr/bin/dumb-init \u2026\"   34 minutes ago      Up 34 minutes                                                                                                  openstackclient\nefec17d7241e   registry.osism.tech/osism/homer:v25.10.1                        \"/bin/sh /entrypoint\u2026\"   34 minutes ago      Up 34 minutes (healthy)      8080/tcp                                                                          homer\n0546b63110c2   registry.osism.tech/dockerhub/ubuntu/squid:6.1-23.10_beta       \"entrypoint.sh -f /e\u2026\"   58 minutes ago      Up 57 minutes (healthy)      192.168.16.5:3128->3128/tcp                                                       squid\nd16968ccbd7e   registry.osism.tech/osism/inventory-reconciler:latest           \"/sbin/tini -- /entr\u2026\"   About an hour ago   Up 40 minutes (healthy)                                                                                        manager-inventory_reconciler-1\n125c6d34e4e3   registry.osism.tech/osism/osism-ansible:latest                  \"/entrypoint.sh osis\u2026\"   About an hour ago   Up 41 minutes (healthy)                                                                                        osism-ansible\n7a3c2a4639c8   registry.osism.tech/osism/ceph-ansible:reef                     \"/entrypoint.sh osis\u2026\"   About an hour ago   Up 41 minutes (healthy)                                                                                        ceph-ansible\n04b4496a5f9c   registry.osism.tech/osism/kolla-ansible:2024.2                  \"/entrypoint.sh osis\u2026\"   About an hour ago   Up 41 minutes (healthy)                                                                                        kolla-ansible\nd9037511a0a8   registry.osism.tech/osism/osism-kubernetes:latest               \"/entrypoint.sh osis\u2026\"   About an hour ago   Up 41 minutes (healthy)                                                                                        osism-kubernetes\n3c94059e4b0f   registry.osism.tech/osism/ara-server:1.7.3                      \"sh -c '/wait && /ru\u2026\"   About an hour ago   Up 41 minutes (healthy)      8000/tcp                                                                          manager-ara-server-1\n2de2ca3ef170   registry.osism.tech/dockerhub/library/mariadb:11.8.4            \"docker-entrypoint.s\u2026\"   About an hour ago   Up 41 minutes (healthy)      3306/tcp                                                                          manager-mariadb-1\n13fae06f8b51   registry.osism.tech/osism/osism:latest                          \"/sbin/tini -- osism\u2026\"   About an hour ago   Up 41 minutes (healthy)                                                                                        manager-openstack-1\n515cac366103   registry.osism.tech/osism/osism:latest                          \"/sbin/tini -- osism\u2026\"   About an hour ago   Up 41 minutes (healthy)                                                                                        manager-beat-1\nfadf80797961   registry.osism.tech/osism/osism:latest                          \"/sbin/tini -- osism\u2026\"   About an hour ago   Up 41 minutes (healthy)      192.168.16.5:8000->8000/tcp                                                       manager-api-1\n3c464a2e2250   registry.osism.tech/osism/osism:latest                          \"/sbin/tini -- sleep\u2026\"   About an hour ago   Up 41 minutes (healthy)                                                                                        osismclient\ne3ec25c34f93   registry.osism.tech/dockerhub/library/redis:7.4.7-alpine        \"docker-entrypoint.s\u2026\"   About an hour ago   Up 41 minutes (healthy)      6379/tcp                                                                          manager-redis-1\n86c5161f8563   registry.osism.tech/osism/osism:latest                          \"/sbin/tini -- osism\u2026\"   About an hour ago   Up 41 minutes (healthy)                                                                                        manager-listener-1\nd2b38759b77b   registry.osism.tech/osism/osism:latest                          \"/sbin/tini -- osism\u2026\"   About an hour ago   Up 41 minutes (healthy)                                                                                        manager-flower-1\n4aca83748015   registry.osism.tech/osism/osism-frontend:latest                 \"docker-entrypoint.s\u2026\"   About an hour ago   Up 41 minutes                192.168.16.5:3000->3000/tcp                                                       osism-frontend\n6bb13d8ad793   registry.osism.tech/dockerhub/library/traefik:v3.5.0            \"/entrypoint.sh trae\u2026\"   About an hour ago   Up About an hour (healthy)   192.168.16.5:80->80/tcp, 192.168.16.5:443->443/tcp, 192.168.16.5:8122->8080/tcp   traefik\n\n## Images @ testbed-manager\n\n+ echo\n+ echo '## Images @ testbed-manager'\n+ echo\n+ osism container testbed-manager images\nREPOSITORY                                               TAG              IMAGE ID       CREATED             SIZE\nregistry.osism.tech/osism/osism-ansible                  latest           b31da8624ff4   About an hour ago   613MB\nregistry.osism.tech/osism/kolla-ansible                  2024.2           c655c028deb8   About an hour ago   610MB\nregistry.osism.tech/osism/ceph-ansible                   reef             f45fb23f980e   About an hour ago   559MB\nregistry.osism.tech/osism/osism-kubernetes               latest           093bdc689455   About an hour ago   1.22GB\nregistry.osism.tech/osism/osism-frontend                 latest           bd2b8e9cfdde   About an hour ago   232MB\nregistry.osism.tech/osism/osism                          latest           99863fdf241c   About an hour ago   409MB\nregistry.osism.tech/osism/inventory-reconciler           latest           66f25a2df9f3   About an hour ago   335MB\nregistry.osism.tech/osism/osism-ansible                  <none>           c6f06501950d   6 hours ago         613MB\nregistry.osism.tech/osism/openstackclient                2024.2           3e84edb92a41   21 hours ago        239MB\nregistry.osism.tech/osism/cephclient                     reef             ebb02e0df028   22 hours ago        453MB\nregistry.osism.tech/kolla/kolla-toolbox                  2024.2           f894a5bc6d0b   24 hours ago        673MB\nregistry.osism.tech/kolla/cron                           2024.2           5b921837d137   24 hours ago        271MB\nregistry.osism.tech/kolla/fluentd                        2024.2           743bb34b6cee   24 hours ago        584MB\nregistry.osism.tech/kolla/prometheus-node-exporter       2024.2           278eaa73d02c   24 hours ago        311MB\nregistry.osism.tech/kolla/prometheus-alertmanager        2024.2           23416dcdd72e   24 hours ago        409MB\nregistry.osism.tech/kolla/prometheus-blackbox-exporter   2024.2           ffe2e0b2e550   24 hours ago        313MB\nregistry.osism.tech/kolla/prometheus-v2-server           2024.2           290ded05e43a   24 hours ago        844MB\nregistry.osism.tech/kolla/prometheus-cadvisor            2024.2           4e4cb8e72dfc   24 hours ago        363MB\nregistry.osism.tech/dockerhub/library/redis              7.4.7-alpine     e08bd8d5a677   3 weeks ago         41.4MB\nregistry.osism.tech/osism/homer                          v25.10.1         ea34b371c716   2 months ago        11.5MB\nregistry.osism.tech/dockerhub/library/mariadb            11.8.4           70745dd8f1d0   3 months ago        334MB\nphpmyadmin/phpmyadmin                                    5.2              e66b1f5a8c58   4 months ago        742MB\nregistry.osism.tech/osism/ara-server                     1.7.3            d1b687333f2f   6 months ago        275MB\nregistry.osism.tech/dockerhub/library/traefik            v3.5.0           11cc59587f6a   7 months ago        226MB\nregistry.osism.tech/dockerhub/ubuntu/squid               6.1-23.10_beta   34b6bbbcf74b   20 months ago       146MB\n+ for node in testbed-manager testbed-node-0 testbed-node-1 testbed-node-2\n++ semver latest 5.0.0\n\n## Containers @ testbed-node-0\n\n+ [[ -1 -eq -1 ]]\n+ [[ latest != \\l\\a\\t\\e\\s\\t ]]\n+ echo\n+ echo '## Containers @ testbed-node-0'\n+ echo\n+ osism container testbed-node-0 ps\nCONTAINER ID   IMAGE                                                                COMMAND                  CREATED          STATUS                    PORTS     NAMES\nc446d4dc5ee8   registry.osism.tech/kolla/octavia-worker:2024.2                      \"dumb-init --single-\u2026\"   4 minutes ago    Up 4 minutes (healthy)              octavia_worker\nb5c92ab41dc7   registry.osism.tech/kolla/octavia-housekeeping:2024.2                \"dumb-init --single-\u2026\"   4 minutes ago    Up 4 minutes (healthy)              octavia_housekeeping\n55b4f5aec8dd   registry.osism.tech/kolla/octavia-health-manager:2024.2              \"dumb-init --single-\u2026\"   5 minutes ago    Up 5 minutes (healthy)              octavia_health_manager\n47566925e4d0   registry.osism.tech/kolla/octavia-driver-agent:2024.2                \"dumb-init --single-\u2026\"   5 minutes ago    Up 5 minutes                        octavia_driver_agent\n93adbcd52864   registry.osism.tech/kolla/octavia-api:2024.2                         \"dumb-init --single-\u2026\"   5 minutes ago    Up 5 minutes (healthy)              octavia_api\nfcc8d20bbbb8   registry.osism.tech/kolla/nova-novncproxy:2024.2                     \"dumb-init --single-\u2026\"   8 minutes ago    Up 8 minutes (healthy)              nova_novncproxy\n2e88e7afc50d   registry.osism.tech/kolla/nova-conductor:2024.2                      \"dumb-init --single-\u2026\"   8 minutes ago    Up 8 minutes (healthy)              nova_conductor\n28ee07355de0   registry.osism.tech/kolla/nova-api:2024.2                            \"dumb-init --single-\u2026\"   9 minutes ago    Up 9 minutes (healthy)              nova_api\n2b4f54d58446   registry.osism.tech/kolla/nova-scheduler:2024.2                      \"dumb-init --single-\u2026\"   10 minutes ago   Up 10 minutes (healthy)             nova_scheduler\ne514848de6bb   registry.osism.tech/kolla/grafana:2024.2                             \"dumb-init --single-\u2026\"   10 minutes ago   Up 10 minutes                       grafana\neeaca894092b   registry.osism.tech/kolla/cinder-backup:2024.2                       \"dumb-init --single-\u2026\"   11 minutes ago   Up 11 minutes (healthy)             cinder_backup\n1702c0db4a81   registry.osism.tech/kolla/cinder-volume:2024.2                       \"dumb-init --single-\u2026\"   11 minutes ago   Up 11 minutes (healthy)             cinder_volume\n15cceafd9c02   registry.osism.tech/kolla/cinder-scheduler:2024.2                    \"dumb-init --single-\u2026\"   12 minutes ago   Up 12 minutes (healthy)             cinder_scheduler\n8dba7f6fd2b6   registry.osism.tech/kolla/glance-api:2024.2                          \"dumb-init --single-\u2026\"   12 minutes ago   Up 12 minutes (healthy)             glance_api\n980ec8d288d0   registry.osism.tech/kolla/cinder-api:2024.2                          \"dumb-init --single-\u2026\"   12 minutes ago   Up 12 minutes (healthy)             cinder_api\n82e10dbf1d99   registry.osism.tech/kolla/prometheus-elasticsearch-exporter:2024.2   \"dumb-init --single-\u2026\"   12 minutes ago   Up 12 minutes                       prometheus_elasticsearch_exporter\n8e3ac52b5f37   registry.osism.tech/kolla/prometheus-cadvisor:2024.2                 \"dumb-init --single-\u2026\"   13 minutes ago   Up 13 minutes                       prometheus_cadvisor\nd3146db51139   registry.osism.tech/kolla/prometheus-memcached-exporter:2024.2       \"dumb-init --single-\u2026\"   13 minutes ago   Up 13 minutes                       prometheus_memcached_exporter\n48bf8f58f2f7   registry.osism.tech/kolla/prometheus-mysqld-exporter:2024.2          \"dumb-init --single-\u2026\"   13 minutes ago   Up 13 minutes                       prometheus_mysqld_exporter\ndb99b39df5fa   registry.osism.tech/kolla/prometheus-node-exporter:2024.2            \"dumb-init --single-\u2026\"   13 minutes ago   Up 13 minutes                       prometheus_node_exporter\n6dc45ba06456   registry.osism.tech/kolla/magnum-conductor:2024.2                    \"dumb-init --single-\u2026\"   14 minutes ago   Up 14 minutes (healthy)             magnum_conductor\n46f7079c6b5f   registry.osism.tech/kolla/magnum-api:2024.2                          \"dumb-init --single-\u2026\"   15 minutes ago   Up 15 minutes (healthy)             magnum_api\n331ecd41b313   registry.osism.tech/kolla/neutron-server:2024.2                      \"dumb-init --single-\u2026\"   15 minutes ago   Up 15 minutes (healthy)             neutron_server\ndc28d708b8d5   registry.osism.tech/kolla/designate-worker:2024.2                    \"dumb-init --single-\u2026\"   16 minutes ago   Up 16 minutes (healthy)             designate_worker\n0c627847d85c   registry.osism.tech/kolla/designate-mdns:2024.2                      \"dumb-init --single-\u2026\"   16 minutes ago   Up 16 minutes (healthy)             designate_mdns\n9a1784bcbe3c   registry.osism.tech/kolla/placement-api:2024.2                       \"dumb-init --single-\u2026\"   16 minutes ago   Up 16 minutes (healthy)             placement_api\n2baf5f5c4bde   registry.osism.tech/kolla/designate-producer:2024.2                  \"dumb-init --single-\u2026\"   16 minutes ago   Up 16 minutes (healthy)             designate_producer\nacdad461bfd6   registry.osism.tech/kolla/designate-central:2024.2                   \"dumb-init --single-\u2026\"   17 minutes ago   Up 16 minutes (healthy)             designate_central\nc5a13c72704a   registry.osism.tech/kolla/designate-api:2024.2                       \"dumb-init --single-\u2026\"   17 minutes ago   Up 17 minutes (healthy)             designate_api\n1866693c5939   registry.osism.tech/kolla/barbican-worker:2024.2                     \"dumb-init --single-\u2026\"   17 minutes ago   Up 17 minutes (healthy)             barbican_worker\nd55d06c19e8f   registry.osism.tech/kolla/barbican-keystone-listener:2024.2          \"dumb-init --single-\u2026\"   17 minutes ago   Up 17 minutes (healthy)             barbican_keystone_listener\n3babadf8056e   registry.osism.tech/kolla/designate-backend-bind9:2024.2             \"dumb-init --single-\u2026\"   17 minutes ago   Up 17 minutes (healthy)             designate_backend_bind9\ne5ae1024a5a0   registry.osism.tech/kolla/barbican-api:2024.2                        \"dumb-init --single-\u2026\"   17 minutes ago   Up 17 minutes (healthy)             barbican_api\nedd213c2b417   registry.osism.tech/osism/ceph-daemon:reef                           \"/usr/bin/ceph-mgr -\u2026\"   18 minutes ago   Up 18 minutes                       ceph-mgr-testbed-node-0\n3616d3cf4fc2   registry.osism.tech/kolla/keystone:2024.2                            \"dumb-init --single-\u2026\"   21 minutes ago   Up 21 minutes (healthy)             keystone\n6ee631dff16c   registry.osism.tech/kolla/keystone-fernet:2024.2                     \"dumb-init --single-\u2026\"   21 minutes ago   Up 21 minutes (healthy)             keystone_fernet\ndafeea7513bf   registry.osism.tech/kolla/keystone-ssh:2024.2                        \"dumb-init --single-\u2026\"   21 minutes ago   Up 21 minutes (healthy)             keystone_ssh\n08587638fa20   registry.osism.tech/kolla/horizon:2024.2                             \"dumb-init --single-\u2026\"   21 minutes ago   Up 21 minutes (healthy)             horizon\n0f4762768487   registry.osism.tech/kolla/mariadb-server:2024.2                      \"dumb-init -- kolla_\u2026\"   23 minutes ago   Up 23 minutes (healthy)             mariadb\n341fc807413d   registry.osism.tech/kolla/opensearch-dashboards:2024.2               \"dumb-init --single-\u2026\"   24 minutes ago   Up 24 minutes (healthy)             opensearch_dashboards\n9e484d669985   registry.osism.tech/kolla/opensearch:2024.2                          \"dumb-init --single-\u2026\"   25 minutes ago   Up 25 minutes (healthy)             opensearch\n4aad693a4e67   registry.osism.tech/osism/ceph-daemon:reef                           \"/usr/bin/ceph-crash\"    25 minutes ago   Up 25 minutes                       ceph-crash-testbed-node-0\n12a27e2732ed   registry.osism.tech/kolla/keepalived:2024.2                          \"dumb-init --single-\u2026\"   26 minutes ago   Up 26 minutes                       keepalived\nfeeddafe9200   registry.osism.tech/kolla/proxysql:2024.2                            \"dumb-init --single-\u2026\"   26 minutes ago   Up 26 minutes (healthy)             proxysql\nae60ab2ca74f   registry.osism.tech/kolla/haproxy:2024.2                             \"dumb-init --single-\u2026\"   26 minutes ago   Up 26 minutes (healthy)             haproxy\n0bf73eb1bc2a   registry.osism.tech/kolla/ovn-northd:2024.2                          \"dumb-init --single-\u2026\"   30 minutes ago   Up 30 minutes                       ovn_northd\n9c61d16c643d   registry.osism.tech/kolla/ovn-sb-db-server:2024.2                    \"dumb-init --single-\u2026\"   30 minutes ago   Up 30 minutes                       ovn_sb_db\n17111b2e006e   registry.osism.tech/kolla/ovn-nb-db-server:2024.2                    \"dumb-init --single-\u2026\"   30 minutes ago   Up 30 minutes                       ovn_nb_db\nf43c50bae847   registry.osism.tech/kolla/ovn-controller:2024.2                      \"dumb-init --single-\u2026\"   31 minutes ago   Up 31 minutes                       ovn_controller\n1cb145726d67   registry.osism.tech/osism/ceph-daemon:reef                           \"/usr/bin/ceph-mon -\u2026\"   31 minutes ago   Up 31 minutes                       ceph-mon-testbed-node-0\n99bb95cd6961   registry.osism.tech/kolla/rabbitmq:2024.2                            \"dumb-init --single-\u2026\"   31 minutes ago   Up 31 minutes (healthy)             rabbitmq\n77f4d5303968   registry.osism.tech/kolla/openvswitch-vswitchd:2024.2                \"dumb-init --single-\u2026\"   32 minutes ago   Up 31 minutes (healthy)             openvswitch_vswitchd\na9f9d1ebe587   registry.osism.tech/kolla/redis-sentinel:2024.2                      \"dumb-init --single-\u2026\"   32 minutes ago   Up 32 minutes (healthy)             redis_sentinel\n14dbe26fb9bf   registry.osism.tech/kolla/openvswitch-db-server:2024.2               \"dumb-init --single-\u2026\"   32 minutes ago   Up 32 minutes (healthy)             openvswitch_db\ne40fd89e362d   registry.osism.tech/kolla/redis:2024.2                               \"dumb-init --single-\u2026\"   32 minutes ago   Up 32 minutes (healthy)             redis\n96f285897c5e   registry.osism.tech/kolla/memcached:2024.2                           \"dumb-init --single-\u2026\"   32 minutes ago   Up 32 minutes (healthy)             memcached\n2236df7422f3   registry.osism.tech/kolla/cron:2024.2                                \"dumb-init --single-\u2026\"   33 minutes ago   Up 33 minutes                       cron\n879bca2009ad   registry.osism.tech/kolla/kolla-toolbox:2024.2                       \"dumb-init --single-\u2026\"   33 minutes ago   Up 33 minutes                       kolla_toolbox\n7101ed2ee4e0   registry.osism.tech/kolla/fluentd:2024.2                             \"dumb-init --single-\u2026\"   34 minutes ago   Up 34 minutes                       fluentd\n\n## Images @ testbed-node-0\n\n+ echo\n+ echo '## Images @ testbed-node-0'\n+ echo\n+ osism container testbed-node-0 images\nREPOSITORY                                                    TAG       IMAGE ID       CREATED        SIZE\nregistry.osism.tech/osism/ceph-daemon                         reef      ff59b29334dd   22 hours ago   1.27GB\nregistry.osism.tech/kolla/memcached                           2024.2    3652d3a7de56   24 hours ago   272MB\nregistry.osism.tech/kolla/proxysql                            2024.2    dbe28c2fd6c3   24 hours ago   418MB\nregistry.osism.tech/kolla/kolla-toolbox                       2024.2    f894a5bc6d0b   24 hours ago   673MB\nregistry.osism.tech/kolla/haproxy                             2024.2    b1fd4c643197   24 hours ago   279MB\nregistry.osism.tech/kolla/cron                                2024.2    5b921837d137   24 hours ago   271MB\nregistry.osism.tech/kolla/rabbitmq                            2024.2    540f6409fbf5   24 hours ago   328MB\nregistry.osism.tech/kolla/keepalived                          2024.2    13679e653294   24 hours ago   282MB\nregistry.osism.tech/kolla/grafana                             2024.2    37fbf97e7f6f   24 hours ago   1.02GB\nregistry.osism.tech/kolla/fluentd                             2024.2    743bb34b6cee   24 hours ago   584MB\nregistry.osism.tech/kolla/opensearch-dashboards               2024.2    c159d79f20f0   24 hours ago   1.53GB\nregistry.osism.tech/kolla/opensearch                          2024.2    80d69bb5f098   24 hours ago   1.56GB\nregistry.osism.tech/kolla/openvswitch-db-server               2024.2    7e08fe973c8e   24 hours ago   284MB\nregistry.osism.tech/kolla/openvswitch-vswitchd                2024.2    4d41d184aa52   24 hours ago   284MB\nregistry.osism.tech/kolla/redis                               2024.2    0fa4d8b23966   24 hours ago   278MB\nregistry.osism.tech/kolla/redis-sentinel                      2024.2    2052a297db29   24 hours ago   278MB\nregistry.osism.tech/kolla/horizon                             2024.2    4d92e9384e36   24 hours ago   1.15GB\nregistry.osism.tech/kolla/mariadb-server                      2024.2    5dbd7b932e63   24 hours ago   457MB\nregistry.osism.tech/kolla/prometheus-node-exporter            2024.2    278eaa73d02c   24 hours ago   311MB\nregistry.osism.tech/kolla/prometheus-elasticsearch-exporter   2024.2    13638ae36e57   24 hours ago   297MB\nregistry.osism.tech/kolla/prometheus-memcached-exporter       2024.2    8ff5a6b3cd16   24 hours ago   304MB\nregistry.osism.tech/kolla/prometheus-mysqld-exporter          2024.2    d9cb73d898bb   24 hours ago   306MB\nregistry.osism.tech/kolla/prometheus-cadvisor                 2024.2    4e4cb8e72dfc   24 hours ago   363MB\nregistry.osism.tech/kolla/ovn-nb-db-server                    2024.2    51d6ef378bbd   24 hours ago   846MB\nregistry.osism.tech/kolla/ovn-sb-db-server                    2024.2    25bb84f2289b   24 hours ago   846MB\nregistry.osism.tech/kolla/ovn-northd                          2024.2    47dae0f63bd5   24 hours ago   846MB\nregistry.osism.tech/kolla/ovn-controller                      2024.2    342e8da8ac3a   24 hours ago   846MB\nregistry.osism.tech/kolla/neutron-server                      2024.2    189daad12150   24 hours ago   1.17GB\nregistry.osism.tech/kolla/octavia-api                         2024.2    624955c2c115   24 hours ago   1.06GB\nregistry.osism.tech/kolla/octavia-housekeeping                2024.2    9b80c75ebf88   24 hours ago   1.03GB\nregistry.osism.tech/kolla/octavia-driver-agent                2024.2    3f4a0924d9ed   24 hours ago   1.06GB\nregistry.osism.tech/kolla/octavia-health-manager              2024.2    741914d697fa   24 hours ago   1.03GB\nregistry.osism.tech/kolla/octavia-worker                      2024.2    45d576cfc14a   24 hours ago   1.03GB\nregistry.osism.tech/kolla/skyline-apiserver                   2024.2    1f7f7078acd6   24 hours ago   995MB\nregistry.osism.tech/kolla/skyline-console                     2024.2    d6dc835952f4   24 hours ago   1.05GB\nregistry.osism.tech/kolla/cinder-backup                       2024.2    db0ba08b76e7   24 hours ago   1.42GB\nregistry.osism.tech/kolla/cinder-volume                       2024.2    aad4ba476f1c   24 hours ago   1.72GB\nregistry.osism.tech/kolla/cinder-scheduler                    2024.2    3102035cbd3f   24 hours ago   1.41GB\nregistry.osism.tech/kolla/cinder-api                          2024.2    ffa30f3933ed   24 hours ago   1.41GB\nregistry.osism.tech/kolla/placement-api                       2024.2    e4b48b9cb79e   24 hours ago   981MB\nregistry.osism.tech/kolla/designate-producer                  2024.2    e3b7812cd6ff   24 hours ago   990MB\nregistry.osism.tech/kolla/designate-central                   2024.2    8ab2ea11494a   24 hours ago   989MB\nregistry.osism.tech/kolla/designate-worker                    2024.2    13bd17ae9e62   24 hours ago   994MB\nregistry.osism.tech/kolla/designate-api                       2024.2    2e44006b8d8c   24 hours ago   990MB\nregistry.osism.tech/kolla/designate-backend-bind9             2024.2    7b60bbae1ff0   24 hours ago   994MB\nregistry.osism.tech/kolla/designate-mdns                      2024.2    a28e142a95bf   24 hours ago   990MB\nregistry.osism.tech/kolla/barbican-worker                     2024.2    0eec7e0f7b33   24 hours ago   996MB\nregistry.osism.tech/kolla/barbican-keystone-listener          2024.2    8c429b2daef4   24 hours ago   996MB\nregistry.osism.tech/kolla/barbican-api                        2024.2    9d4b161f78ee   24 hours ago   996MB\nregistry.osism.tech/kolla/keystone-fernet                     2024.2    3770c4e50a62   24 hours ago   1.04GB\nregistry.osism.tech/kolla/keystone-ssh                        2024.2    a75dcc713ad6   24 hours ago   1.05GB\nregistry.osism.tech/kolla/keystone                            2024.2    f3afb1708fa9   24 hours ago   1.07GB\nregistry.osism.tech/kolla/nova-conductor                      2024.2    7e0dfe6dfac8   24 hours ago   1.22GB\nregistry.osism.tech/kolla/nova-api                            2024.2    b141835a8ce3   24 hours ago   1.22GB\nregistry.osism.tech/kolla/nova-novncproxy                     2024.2    0e230dfe273c   24 hours ago   1.37GB\nregistry.osism.tech/kolla/nova-scheduler                      2024.2    1d5f182c3a75   24 hours ago   1.22GB\nregistry.osism.tech/kolla/aodh-listener                       2024.2    1d3d06d0cc72   24 hours ago   979MB\nregistry.osism.tech/kolla/aodh-api                            2024.2    d7ef40f551b0   24 hours ago   979MB\nregistry.osism.tech/kolla/aodh-notifier                       2024.2    826ad4a5837c   24 hours ago   979MB\nregistry.osism.tech/kolla/aodh-evaluator                      2024.2    9fa396903596   24 hours ago   979MB\nregistry.osism.tech/kolla/ceilometer-notification             2024.2    2d0736f4b996   24 hours ago   981MB\nregistry.osism.tech/kolla/ceilometer-central                  2024.2    112e904da6ac   24 hours ago   982MB\nregistry.osism.tech/kolla/magnum-api                          2024.2    5f8755d26e30   24 hours ago   1.13GB\nregistry.osism.tech/kolla/magnum-conductor                    2024.2    ad0e7877ae1a   24 hours ago   1.25GB\nregistry.osism.tech/kolla/glance-api                          2024.2    4e1be079f392   24 hours ago   1.1GB\n+ for node in testbed-manager testbed-node-0 testbed-node-1 testbed-node-2\n++ semver latest 5.0.0\n\n## Containers @ testbed-node-1\n\n+ [[ -1 -eq -1 ]]\n+ [[ latest != \\l\\a\\t\\e\\s\\t ]]\n+ echo\n+ echo '## Containers @ testbed-node-1'\n+ echo\n+ osism container testbed-node-1 ps\nCONTAINER ID   IMAGE                                                                COMMAND                  CREATED          STATUS                    PORTS     NAMES\n443c3105e0a7   registry.osism.tech/kolla/octavia-worker:2024.2                      \"dumb-init --single-\u2026\"   4 minutes ago    Up 4 minutes (healthy)              octavia_worker\n20d02ff22874   registry.osism.tech/kolla/octavia-housekeeping:2024.2                \"dumb-init --single-\u2026\"   4 minutes ago    Up 4 minutes (healthy)              octavia_housekeeping\nb063938944b8   registry.osism.tech/kolla/octavia-health-manager:2024.2              \"dumb-init --single-\u2026\"   5 minutes ago    Up 5 minutes (healthy)              octavia_health_manager\n715117e226ad   registry.osism.tech/kolla/octavia-driver-agent:2024.2                \"dumb-init --single-\u2026\"   5 minutes ago    Up 5 minutes                        octavia_driver_agent\n1fe566592e1b   registry.osism.tech/kolla/octavia-api:2024.2                         \"dumb-init --single-\u2026\"   5 minutes ago    Up 5 minutes (healthy)              octavia_api\n6f2a47747067   registry.osism.tech/kolla/nova-novncproxy:2024.2                     \"dumb-init --single-\u2026\"   8 minutes ago    Up 8 minutes (healthy)              nova_novncproxy\ndb660a20e692   registry.osism.tech/kolla/nova-conductor:2024.2                      \"dumb-init --single-\u2026\"   8 minutes ago    Up 8 minutes (healthy)              nova_conductor\n240b47367be8   registry.osism.tech/kolla/grafana:2024.2                             \"dumb-init --single-\u2026\"   10 minutes ago   Up 10 minutes                       grafana\n34998db525ca   registry.osism.tech/kolla/nova-api:2024.2                            \"dumb-init --single-\u2026\"   10 minutes ago   Up 10 minutes (healthy)             nova_api\n1de9015b70ff   registry.osism.tech/kolla/nova-scheduler:2024.2                      \"dumb-init --single-\u2026\"   10 minutes ago   Up 10 minutes (healthy)             nova_scheduler\n6b89e77fcb73   registry.osism.tech/kolla/cinder-backup:2024.2                       \"dumb-init --single-\u2026\"   11 minutes ago   Up 11 minutes (healthy)             cinder_backup\n83a200208cd2   registry.osism.tech/kolla/cinder-volume:2024.2                       \"dumb-init --single-\u2026\"   11 minutes ago   Up 11 minutes (healthy)             cinder_volume\n87ba4474b65d   registry.osism.tech/kolla/glance-api:2024.2                          \"dumb-init --single-\u2026\"   12 minutes ago   Up 12 minutes (healthy)             glance_api\n419a24ce1d7c   registry.osism.tech/kolla/cinder-scheduler:2024.2                    \"dumb-init --single-\u2026\"   12 minutes ago   Up 12 minutes (healthy)             cinder_scheduler\n8641a24ebad2   registry.osism.tech/kolla/cinder-api:2024.2                          \"dumb-init --single-\u2026\"   12 minutes ago   Up 12 minutes (healthy)             cinder_api\n982a8d255f2d   registry.osism.tech/kolla/prometheus-elasticsearch-exporter:2024.2   \"dumb-init --single-\u2026\"   13 minutes ago   Up 13 minutes                       prometheus_elasticsearch_exporter\n7202121cd714   registry.osism.tech/kolla/prometheus-cadvisor:2024.2                 \"dumb-init --single-\u2026\"   13 minutes ago   Up 13 minutes                       prometheus_cadvisor\n9b82bdefbd04   registry.osism.tech/kolla/prometheus-memcached-exporter:2024.2       \"dumb-init --single-\u2026\"   13 minutes ago   Up 13 minutes                       prometheus_memcached_exporter\n7f03efd478d8   registry.osism.tech/kolla/prometheus-mysqld-exporter:2024.2          \"dumb-init --single-\u2026\"   13 minutes ago   Up 13 minutes                       prometheus_mysqld_exporter\n5cf1ef52b333   registry.osism.tech/kolla/prometheus-node-exporter:2024.2            \"dumb-init --single-\u2026\"   13 minutes ago   Up 13 minutes                       prometheus_node_exporter\nd05b983ca5cd   registry.osism.tech/kolla/magnum-conductor:2024.2                    \"dumb-init --single-\u2026\"   15 minutes ago   Up 15 minutes (healthy)             magnum_conductor\n2bc61d574472   registry.osism.tech/kolla/magnum-api:2024.2                          \"dumb-init --single-\u2026\"   15 minutes ago   Up 15 minutes (healthy)             magnum_api\na12bdf928dd3   registry.osism.tech/kolla/neutron-server:2024.2                      \"dumb-init --single-\u2026\"   15 minutes ago   Up 15 minutes (healthy)             neutron_server\n696c642518c4   registry.osism.tech/kolla/designate-worker:2024.2                    \"dumb-init --single-\u2026\"   16 minutes ago   Up 16 minutes (healthy)             designate_worker\n3b21b349fbde   registry.osism.tech/kolla/placement-api:2024.2                       \"dumb-init --single-\u2026\"   16 minutes ago   Up 16 minutes (healthy)             placement_api\n274437c23da9   registry.osism.tech/kolla/designate-mdns:2024.2                      \"dumb-init --single-\u2026\"   16 minutes ago   Up 16 minutes (healthy)             designate_mdns\n76ab2f193443   registry.osism.tech/kolla/designate-producer:2024.2                  \"dumb-init --single-\u2026\"   16 minutes ago   Up 16 minutes (healthy)             designate_producer\n5c5d7a49f4bd   registry.osism.tech/kolla/designate-central:2024.2                   \"dumb-init --single-\u2026\"   17 minutes ago   Up 17 minutes (healthy)             designate_central\nd77a29bb1e26   registry.osism.tech/kolla/designate-api:2024.2                       \"dumb-init --single-\u2026\"   17 minutes ago   Up 17 minutes (healthy)             designate_api\nfc3dd7be7471   registry.osism.tech/kolla/barbican-worker:2024.2                     \"dumb-init --single-\u2026\"   17 minutes ago   Up 17 minutes (healthy)             barbican_worker\n336a8e22c323   registry.osism.tech/kolla/barbican-keystone-listener:2024.2          \"dumb-init --single-\u2026\"   17 minutes ago   Up 17 minutes (healthy)             barbican_keystone_listener\nb3eca58eb0ff   registry.osism.tech/kolla/designate-backend-bind9:2024.2             \"dumb-init --single-\u2026\"   17 minutes ago   Up 17 minutes (healthy)             designate_backend_bind9\nb19004b05880   registry.osism.tech/kolla/barbican-api:2024.2                        \"dumb-init --single-\u2026\"   17 minutes ago   Up 17 minutes (healthy)             barbican_api\n3534c242e3d3   registry.osism.tech/osism/ceph-daemon:reef                           \"/usr/bin/ceph-mgr -\u2026\"   18 minutes ago   Up 18 minutes                       ceph-mgr-testbed-node-1\n537e4ace13ae   registry.osism.tech/kolla/keystone:2024.2                            \"dumb-init --single-\u2026\"   21 minutes ago   Up 21 minutes (healthy)             keystone\n3e420c4d8d47   registry.osism.tech/kolla/horizon:2024.2                             \"dumb-init --single-\u2026\"   21 minutes ago   Up 21 minutes (healthy)             horizon\n3f909ec0fd73   registry.osism.tech/kolla/keystone-fernet:2024.2                     \"dumb-init --single-\u2026\"   21 minutes ago   Up 21 minutes (healthy)             keystone_fernet\nbed5bcc8f080   registry.osism.tech/kolla/keystone-ssh:2024.2                        \"dumb-init --single-\u2026\"   21 minutes ago   Up 21 minutes (healthy)             keystone_ssh\neb6cc91c9aeb   registry.osism.tech/kolla/opensearch-dashboards:2024.2               \"dumb-init --single-\u2026\"   23 minutes ago   Up 23 minutes (healthy)             opensearch_dashboards\n84651137feeb   registry.osism.tech/kolla/mariadb-server:2024.2                      \"dumb-init -- kolla_\u2026\"   24 minutes ago   Up 24 minutes (healthy)             mariadb\n479f29fcf95d   registry.osism.tech/kolla/opensearch:2024.2                          \"dumb-init --single-\u2026\"   24 minutes ago   Up 24 minutes (healthy)             opensearch\n2d3469b3e3ff   registry.osism.tech/osism/ceph-daemon:reef                           \"/usr/bin/ceph-crash\"    26 minutes ago   Up 26 minutes                       ceph-crash-testbed-node-1\nc19fe3d26618   registry.osism.tech/kolla/keepalived:2024.2                          \"dumb-init --single-\u2026\"   26 minutes ago   Up 26 minutes                       keepalived\n6204d745fe4b   registry.osism.tech/kolla/proxysql:2024.2                            \"dumb-init --single-\u2026\"   26 minutes ago   Up 26 minutes (healthy)             proxysql\n80e9102d54d3   registry.osism.tech/kolla/haproxy:2024.2                             \"dumb-init --single-\u2026\"   27 minutes ago   Up 27 minutes (healthy)             haproxy\n784e37f7ab24   registry.osism.tech/kolla/ovn-northd:2024.2                          \"dumb-init --single-\u2026\"   30 minutes ago   Up 29 minutes                       ovn_northd\n23c769291118   registry.osism.tech/kolla/ovn-sb-db-server:2024.2                    \"dumb-init --single-\u2026\"   30 minutes ago   Up 29 minutes                       ovn_sb_db\n575c84596223   registry.osism.tech/kolla/ovn-nb-db-server:2024.2                    \"dumb-init --single-\u2026\"   30 minutes ago   Up 29 minutes                       ovn_nb_db\n82884e1e4587   registry.osism.tech/kolla/ovn-controller:2024.2                      \"dumb-init --single-\u2026\"   30 minutes ago   Up 30 minutes                       ovn_controller\n7264d68f426f   registry.osism.tech/kolla/rabbitmq:2024.2                            \"dumb-init --single-\u2026\"   30 minutes ago   Up 30 minutes (healthy)             rabbitmq\n65c197c778ba   registry.osism.tech/osism/ceph-daemon:reef                           \"/usr/bin/ceph-mon -\u2026\"   31 minutes ago   Up 31 minutes                       ceph-mon-testbed-node-1\nb010803b2bf5   registry.osism.tech/kolla/openvswitch-vswitchd:2024.2                \"dumb-init --single-\u2026\"   32 minutes ago   Up 31 minutes (healthy)             openvswitch_vswitchd\ne7259e9ebf53   registry.osism.tech/kolla/redis-sentinel:2024.2                      \"dumb-init --single-\u2026\"   32 minutes ago   Up 32 minutes (healthy)             redis_sentinel\nd7292eec7f67   registry.osism.tech/kolla/openvswitch-db-server:2024.2               \"dumb-init --single-\u2026\"   32 minutes ago   Up 32 minutes (healthy)             openvswitch_db\n7d769c827d50   registry.osism.tech/kolla/redis:2024.2                               \"dumb-init --single-\u2026\"   32 minutes ago   Up 32 minutes (healthy)             redis\n7f70aeb5717e   registry.osism.tech/kolla/memcached:2024.2                           \"dumb-init --single-\u2026\"   32 minutes ago   Up 32 minutes (healthy)             memcached\n1d9ed2d48c5a   registry.osism.tech/kolla/cron:2024.2                                \"dumb-init --single-\u2026\"   33 minutes ago   Up 33 minutes                       cron\ne2a61650650f   registry.osism.tech/kolla/kolla-toolbox:2024.2                       \"dumb-init --single-\u2026\"   33 minutes ago   Up 33 minutes                       kolla_toolbox\nbaedf1310f6d   registry.osism.tech/kolla/fluentd:2024.2                             \"dumb-init --single-\u2026\"   33 minutes ago   Up 33 minutes                       fluentd\n\n## Images @ testbed-node-1\n\n+ echo\n+ echo '## Images @ testbed-node-1'\n+ echo\n+ osism container testbed-node-1 images\nREPOSITORY                                                    TAG       IMAGE ID       CREATED        SIZE\nregistry.osism.tech/osism/ceph-daemon                         reef      ff59b29334dd   22 hours ago   1.27GB\nregistry.osism.tech/kolla/memcached                           2024.2    3652d3a7de56   24 hours ago   272MB\nregistry.osism.tech/kolla/proxysql                            2024.2    dbe28c2fd6c3   24 hours ago   418MB\nregistry.osism.tech/kolla/kolla-toolbox                       2024.2    f894a5bc6d0b   24 hours ago   673MB\nregistry.osism.tech/kolla/haproxy                             2024.2    b1fd4c643197   24 hours ago   279MB\nregistry.osism.tech/kolla/cron                                2024.2    5b921837d137   24 hours ago   271MB\nregistry.osism.tech/kolla/rabbitmq                            2024.2    540f6409fbf5   24 hours ago   328MB\nregistry.osism.tech/kolla/keepalived                          2024.2    13679e653294   24 hours ago   282MB\nregistry.osism.tech/kolla/grafana                             2024.2    37fbf97e7f6f   24 hours ago   1.02GB\nregistry.osism.tech/kolla/fluentd                             2024.2    743bb34b6cee   24 hours ago   584MB\nregistry.osism.tech/kolla/opensearch-dashboards               2024.2    c159d79f20f0   24 hours ago   1.53GB\nregistry.osism.tech/kolla/opensearch                          2024.2    80d69bb5f098   24 hours ago   1.56GB\nregistry.osism.tech/kolla/openvswitch-vswitchd                2024.2    4d41d184aa52   24 hours ago   284MB\nregistry.osism.tech/kolla/openvswitch-db-server               2024.2    7e08fe973c8e   24 hours ago   284MB\nregistry.osism.tech/kolla/redis                               2024.2    0fa4d8b23966   24 hours ago   278MB\nregistry.osism.tech/kolla/redis-sentinel                      2024.2    2052a297db29   24 hours ago   278MB\nregistry.osism.tech/kolla/horizon                             2024.2    4d92e9384e36   24 hours ago   1.15GB\nregistry.osism.tech/kolla/mariadb-server                      2024.2    5dbd7b932e63   24 hours ago   457MB\nregistry.osism.tech/kolla/prometheus-node-exporter            2024.2    278eaa73d02c   24 hours ago   311MB\nregistry.osism.tech/kolla/prometheus-elasticsearch-exporter   2024.2    13638ae36e57   24 hours ago   297MB\nregistry.osism.tech/kolla/prometheus-memcached-exporter       2024.2    8ff5a6b3cd16   24 hours ago   304MB\nregistry.osism.tech/kolla/prometheus-mysqld-exporter          2024.2    d9cb73d898bb   24 hours ago   306MB\nregistry.osism.tech/kolla/prometheus-cadvisor                 2024.2    4e4cb8e72dfc   24 hours ago   363MB\nregistry.osism.tech/kolla/ovn-nb-db-server                    2024.2    51d6ef378bbd   24 hours ago   846MB\nregistry.osism.tech/kolla/ovn-sb-db-server                    2024.2    25bb84f2289b   24 hours ago   846MB\nregistry.osism.tech/kolla/ovn-controller                      2024.2    342e8da8ac3a   24 hours ago   846MB\nregistry.osism.tech/kolla/ovn-northd                          2024.2    47dae0f63bd5   24 hours ago   846MB\nregistry.osism.tech/kolla/neutron-server                      2024.2    189daad12150   24 hours ago   1.17GB\nregistry.osism.tech/kolla/octavia-api                         2024.2    624955c2c115   24 hours ago   1.06GB\nregistry.osism.tech/kolla/octavia-housekeeping                2024.2    9b80c75ebf88   24 hours ago   1.03GB\nregistry.osism.tech/kolla/octavia-driver-agent                2024.2    3f4a0924d9ed   24 hours ago   1.06GB\nregistry.osism.tech/kolla/octavia-health-manager              2024.2    741914d697fa   24 hours ago   1.03GB\nregistry.osism.tech/kolla/octavia-worker                      2024.2    45d576cfc14a   24 hours ago   1.03GB\nregistry.osism.tech/kolla/cinder-backup                       2024.2    db0ba08b76e7   24 hours ago   1.42GB\nregistry.osism.tech/kolla/cinder-volume                       2024.2    aad4ba476f1c   24 hours ago   1.72GB\nregistry.osism.tech/kolla/cinder-scheduler                    2024.2    3102035cbd3f   24 hours ago   1.41GB\nregistry.osism.tech/kolla/cinder-api                          2024.2    ffa30f3933ed   24 hours ago   1.41GB\nregistry.osism.tech/kolla/placement-api                       2024.2    e4b48b9cb79e   24 hours ago   981MB\nregistry.osism.tech/kolla/designate-producer                  2024.2    e3b7812cd6ff   24 hours ago   990MB\nregistry.osism.tech/kolla/designate-central                   2024.2    8ab2ea11494a   24 hours ago   989MB\nregistry.osism.tech/kolla/designate-worker                    2024.2    13bd17ae9e62   24 hours ago   994MB\nregistry.osism.tech/kolla/designate-api                       2024.2    2e44006b8d8c   24 hours ago   990MB\nregistry.osism.tech/kolla/designate-backend-bind9             2024.2    7b60bbae1ff0   24 hours ago   994MB\nregistry.osism.tech/kolla/designate-mdns                      2024.2    a28e142a95bf   24 hours ago   990MB\nregistry.osism.tech/kolla/barbican-worker                     2024.2    0eec7e0f7b33   24 hours ago   996MB\nregistry.osism.tech/kolla/barbican-keystone-listener          2024.2    8c429b2daef4   24 hours ago   996MB\nregistry.osism.tech/kolla/barbican-api                        2024.2    9d4b161f78ee   24 hours ago   996MB\nregistry.osism.tech/kolla/keystone-fernet                     2024.2    3770c4e50a62   24 hours ago   1.04GB\nregistry.osism.tech/kolla/keystone-ssh                        2024.2    a75dcc713ad6   24 hours ago   1.05GB\nregistry.osism.tech/kolla/keystone                            2024.2    f3afb1708fa9   24 hours ago   1.07GB\nregistry.osism.tech/kolla/nova-conductor                      2024.2    7e0dfe6dfac8   24 hours ago   1.22GB\nregistry.osism.tech/kolla/nova-api                            2024.2    b141835a8ce3   24 hours ago   1.22GB\nregistry.osism.tech/kolla/nova-novncproxy                     2024.2    0e230dfe273c   24 hours ago   1.37GB\nregistry.osism.tech/kolla/nova-scheduler                      2024.2    1d5f182c3a75   24 hours ago   1.22GB\nregistry.osism.tech/kolla/magnum-api                          2024.2    5f8755d26e30   24 hours ago   1.13GB\nregistry.osism.tech/kolla/magnum-conductor                    2024.2    ad0e7877ae1a   24 hours ago   1.25GB\nregistry.osism.tech/kolla/glance-api                          2024.2    4e1be079f392   24 hours ago   1.1GB\n+ for node in testbed-manager testbed-node-0 testbed-node-1 testbed-node-2\n++ semver latest 5.0.0\n\n## Containers @ testbed-node-2\n\n+ [[ -1 -eq -1 ]]\n+ [[ latest != \\l\\a\\t\\e\\s\\t ]]\n+ echo\n+ echo '## Containers @ testbed-node-2'\n+ echo\n+ osism container testbed-node-2 ps\nCONTAINER ID   IMAGE                                                                COMMAND                  CREATED          STATUS                    PORTS     NAMES\n204b517fdebb   registry.osism.tech/kolla/octavia-worker:2024.2                      \"dumb-init --single-\u2026\"   4 minutes ago    Up 4 minutes (healthy)              octavia_worker\n36a8f6d80767   registry.osism.tech/kolla/octavia-housekeeping:2024.2                \"dumb-init --single-\u2026\"   5 minutes ago    Up 5 minutes (healthy)              octavia_housekeeping\na0f8119d1238   registry.osism.tech/kolla/octavia-health-manager:2024.2              \"dumb-init --single-\u2026\"   5 minutes ago    Up 5 minutes (healthy)              octavia_health_manager\n90709743ce68   registry.osism.tech/kolla/octavia-driver-agent:2024.2                \"dumb-init --single-\u2026\"   5 minutes ago    Up 5 minutes                        octavia_driver_agent\na3dccaadc734   registry.osism.tech/kolla/octavia-api:2024.2                         \"dumb-init --single-\u2026\"   5 minutes ago    Up 5 minutes (healthy)              octavia_api\nf2dcf01d1373   registry.osism.tech/kolla/nova-novncproxy:2024.2                     \"dumb-init --single-\u2026\"   8 minutes ago    Up 8 minutes (healthy)              nova_novncproxy\n811663d32f5d   registry.osism.tech/kolla/nova-conductor:2024.2                      \"dumb-init --single-\u2026\"   8 minutes ago    Up 8 minutes (healthy)              nova_conductor\nc271d10de826   registry.osism.tech/kolla/grafana:2024.2                             \"dumb-init --single-\u2026\"   10 minutes ago   Up 10 minutes                       grafana\n58dceffbac3d   registry.osism.tech/kolla/nova-api:2024.2                            \"dumb-init --single-\u2026\"   10 minutes ago   Up 10 minutes (healthy)             nova_api\nacb91a3b3e52   registry.osism.tech/kolla/nova-scheduler:2024.2                      \"dumb-init --single-\u2026\"   10 minutes ago   Up 10 minutes (healthy)             nova_scheduler\n66c3acb647f6   registry.osism.tech/kolla/cinder-backup:2024.2                       \"dumb-init --single-\u2026\"   11 minutes ago   Up 11 minutes (healthy)             cinder_backup\n8ae7f28555cd   registry.osism.tech/kolla/cinder-volume:2024.2                       \"dumb-init --single-\u2026\"   11 minutes ago   Up 11 minutes (healthy)             cinder_volume\n9bf88c0ea1d0   registry.osism.tech/kolla/glance-api:2024.2                          \"dumb-init --single-\u2026\"   12 minutes ago   Up 12 minutes (healthy)             glance_api\ne5a3eec726f2   registry.osism.tech/kolla/cinder-scheduler:2024.2                    \"dumb-init --single-\u2026\"   12 minutes ago   Up 12 minutes (healthy)             cinder_scheduler\na36c539b639c   registry.osism.tech/kolla/cinder-api:2024.2                          \"dumb-init --single-\u2026\"   12 minutes ago   Up 12 minutes (healthy)             cinder_api\n827264021b9d   registry.osism.tech/kolla/prometheus-elasticsearch-exporter:2024.2   \"dumb-init --single-\u2026\"   13 minutes ago   Up 13 minutes                       prometheus_elasticsearch_exporter\nc0038ca67690   registry.osism.tech/kolla/prometheus-cadvisor:2024.2                 \"dumb-init --single-\u2026\"   13 minutes ago   Up 13 minutes                       prometheus_cadvisor\n1941c2999ee6   registry.osism.tech/kolla/prometheus-memcached-exporter:2024.2       \"dumb-init --single-\u2026\"   13 minutes ago   Up 13 minutes                       prometheus_memcached_exporter\nbac938ae89ca   registry.osism.tech/kolla/prometheus-mysqld-exporter:2024.2          \"dumb-init --single-\u2026\"   13 minutes ago   Up 13 minutes                       prometheus_mysqld_exporter\n031f59732dc3   registry.osism.tech/kolla/prometheus-node-exporter:2024.2            \"dumb-init --single-\u2026\"   14 minutes ago   Up 14 minutes                       prometheus_node_exporter\n9e744e3eaacd   registry.osism.tech/kolla/magnum-conductor:2024.2                    \"dumb-init --single-\u2026\"   15 minutes ago   Up 15 minutes (healthy)             magnum_conductor\n301fea6b9b29   registry.osism.tech/kolla/magnum-api:2024.2                          \"dumb-init --single-\u2026\"   15 minutes ago   Up 15 minutes (healthy)             magnum_api\n39d94878860b   registry.osism.tech/kolla/neutron-server:2024.2                      \"dumb-init --single-\u2026\"   15 minutes ago   Up 15 minutes (healthy)             neutron_server\n980b2f7e410d   registry.osism.tech/kolla/designate-worker:2024.2                    \"dumb-init --single-\u2026\"   16 minutes ago   Up 16 minutes (healthy)             designate_worker\n6ccc7bd0eff7   registry.osism.tech/kolla/designate-mdns:2024.2                      \"dumb-init --single-\u2026\"   16 minutes ago   Up 16 minutes (healthy)             designate_mdns\n7acc7ed7f354   registry.osism.tech/kolla/placement-api:2024.2                       \"dumb-init --single-\u2026\"   16 minutes ago   Up 16 minutes (healthy)             placement_api\n19ff8721f4df   registry.osism.tech/kolla/designate-producer:2024.2                  \"dumb-init --single-\u2026\"   17 minutes ago   Up 16 minutes (healthy)             designate_producer\nd3435344fc9d   registry.osism.tech/kolla/designate-central:2024.2                   \"dumb-init --single-\u2026\"   17 minutes ago   Up 17 minutes (healthy)             designate_central\ne99370646bc6   registry.osism.tech/kolla/designate-api:2024.2                       \"dumb-init --single-\u2026\"   17 minutes ago   Up 17 minutes (healthy)             designate_api\n26829716792d   registry.osism.tech/kolla/barbican-worker:2024.2                     \"dumb-init --single-\u2026\"   17 minutes ago   Up 17 minutes (healthy)             barbican_worker\nd93c9f61dfeb   registry.osism.tech/kolla/barbican-keystone-listener:2024.2          \"dumb-init --single-\u2026\"   17 minutes ago   Up 17 minutes (healthy)             barbican_keystone_listener\n6d90d9f2763b   registry.osism.tech/kolla/designate-backend-bind9:2024.2             \"dumb-init --single-\u2026\"   17 minutes ago   Up 17 minutes (healthy)             designate_backend_bind9\nb727b0e64d9f   registry.osism.tech/kolla/barbican-api:2024.2                        \"dumb-init --single-\u2026\"   17 minutes ago   Up 17 minutes (healthy)             barbican_api\nce6817b37e84   registry.osism.tech/osism/ceph-daemon:reef                           \"/usr/bin/ceph-mgr -\u2026\"   18 minutes ago   Up 18 minutes                       ceph-mgr-testbed-node-2\n1dfda4644c27   registry.osism.tech/kolla/keystone:2024.2                            \"dumb-init --single-\u2026\"   21 minutes ago   Up 21 minutes (healthy)             keystone\n03bc6395d015   registry.osism.tech/kolla/horizon:2024.2                             \"dumb-init --single-\u2026\"   21 minutes ago   Up 21 minutes (healthy)             horizon\n2ea35d17ff6d   registry.osism.tech/kolla/keystone-fernet:2024.2                     \"dumb-init --single-\u2026\"   21 minutes ago   Up 21 minutes (healthy)             keystone_fernet\nb1526b1936bc   registry.osism.tech/kolla/keystone-ssh:2024.2                        \"dumb-init --single-\u2026\"   21 minutes ago   Up 21 minutes (healthy)             keystone_ssh\nd954826f0c18   registry.osism.tech/kolla/opensearch-dashboards:2024.2               \"dumb-init --single-\u2026\"   23 minutes ago   Up 23 minutes (healthy)             opensearch_dashboards\nd2f01bbf78dc   registry.osism.tech/kolla/mariadb-server:2024.2                      \"dumb-init -- kolla_\u2026\"   24 minutes ago   Up 24 minutes (healthy)             mariadb\n8ae3b0ecb003   registry.osism.tech/kolla/opensearch:2024.2                          \"dumb-init --single-\u2026\"   25 minutes ago   Up 24 minutes (healthy)             opensearch\n41019e3dcdce   registry.osism.tech/osism/ceph-daemon:reef                           \"/usr/bin/ceph-crash\"    26 minutes ago   Up 26 minutes                       ceph-crash-testbed-node-2\n05144260ecef   registry.osism.tech/kolla/keepalived:2024.2                          \"dumb-init --single-\u2026\"   26 minutes ago   Up 26 minutes                       keepalived\ne61290ed4ec8   registry.osism.tech/kolla/proxysql:2024.2                            \"dumb-init --single-\u2026\"   26 minutes ago   Up 26 minutes (healthy)             proxysql\n6bc306d583d0   registry.osism.tech/kolla/haproxy:2024.2                             \"dumb-init --single-\u2026\"   27 minutes ago   Up 27 minutes (healthy)             haproxy\n109abbc59211   registry.osism.tech/kolla/ovn-northd:2024.2                          \"dumb-init --single-\u2026\"   30 minutes ago   Up 29 minutes                       ovn_northd\n1bb7fc0990bc   registry.osism.tech/kolla/ovn-sb-db-server:2024.2                    \"dumb-init --single-\u2026\"   30 minutes ago   Up 29 minutes                       ovn_sb_db\n7bf6dc9448f0   registry.osism.tech/kolla/ovn-nb-db-server:2024.2                    \"dumb-init --single-\u2026\"   30 minutes ago   Up 30 minutes                       ovn_nb_db\n9670df24e6f8   registry.osism.tech/kolla/rabbitmq:2024.2                            \"dumb-init --single-\u2026\"   30 minutes ago   Up 30 minutes (healthy)             rabbitmq\nc2f6ebfa804e   registry.osism.tech/kolla/ovn-controller:2024.2                      \"dumb-init --single-\u2026\"   30 minutes ago   Up 30 minutes                       ovn_controller\n839d7a58ab74   registry.osism.tech/osism/ceph-daemon:reef                           \"/usr/bin/ceph-mon -\u2026\"   31 minutes ago   Up 31 minutes                       ceph-mon-testbed-node-2\n006f681ca5fa   registry.osism.tech/kolla/openvswitch-vswitchd:2024.2                \"dumb-init --single-\u2026\"   32 minutes ago   Up 32 minutes (healthy)             openvswitch_vswitchd\n1e6b05f5a799   registry.osism.tech/kolla/redis-sentinel:2024.2                      \"dumb-init --single-\u2026\"   32 minutes ago   Up 32 minutes (healthy)             redis_sentinel\n5a83712dd552   registry.osism.tech/kolla/openvswitch-db-server:2024.2               \"dumb-init --single-\u2026\"   32 minutes ago   Up 32 minutes (healthy)             openvswitch_db\n2957a3cb2019   registry.osism.tech/kolla/redis:2024.2                               \"dumb-init --single-\u2026\"   32 minutes ago   Up 32 minutes (healthy)             redis\n695203b12a33   registry.osism.tech/kolla/memcached:2024.2                           \"dumb-init --single-\u2026\"   32 minutes ago   Up 32 minutes (healthy)             memcached\ne86e254e6196   registry.osism.tech/kolla/cron:2024.2                                \"dumb-init --single-\u2026\"   33 minutes ago   Up 33 minutes                       cron\n5a47ec47399b   registry.osism.tech/kolla/kolla-toolbox:2024.2                       \"dumb-init --single-\u2026\"   33 minutes ago   Up 33 minutes                       kolla_toolbox\n7a0d1026089c   registry.osism.tech/kolla/fluentd:2024.2                             \"dumb-init --single-\u2026\"   33 minutes ago   Up 33 minutes                       fluentd\n\n## Images @ testbed-node-2\n\n+ echo\n+ echo '## Images @ testbed-node-2'\n+ echo\n+ osism container testbed-node-2 images\nREPOSITORY                                                    TAG       IMAGE ID       CREATED        SIZE\nregistry.osism.tech/osism/ceph-daemon                         reef      ff59b29334dd   22 hours ago   1.27GB\nregistry.osism.tech/kolla/memcached                           2024.2    3652d3a7de56   24 hours ago   272MB\nregistry.osism.tech/kolla/proxysql                            2024.2    dbe28c2fd6c3   24 hours ago   418MB\nregistry.osism.tech/kolla/kolla-toolbox                       2024.2    f894a5bc6d0b   24 hours ago   673MB\nregistry.osism.tech/kolla/haproxy                             2024.2    b1fd4c643197   24 hours ago   279MB\nregistry.osism.tech/kolla/cron                                2024.2    5b921837d137   24 hours ago   271MB\nregistry.osism.tech/kolla/rabbitmq                            2024.2    540f6409fbf5   24 hours ago   328MB\nregistry.osism.tech/kolla/keepalived                          2024.2    13679e653294   24 hours ago   282MB\nregistry.osism.tech/kolla/grafana                             2024.2    37fbf97e7f6f   24 hours ago   1.02GB\nregistry.osism.tech/kolla/fluentd                             2024.2    743bb34b6cee   24 hours ago   584MB\nregistry.osism.tech/kolla/opensearch-dashboards               2024.2    c159d79f20f0   24 hours ago   1.53GB\nregistry.osism.tech/kolla/opensearch                          2024.2    80d69bb5f098   24 hours ago   1.56GB\nregistry.osism.tech/kolla/openvswitch-db-server               2024.2    7e08fe973c8e   24 hours ago   284MB\nregistry.osism.tech/kolla/openvswitch-vswitchd                2024.2    4d41d184aa52   24 hours ago   284MB\nregistry.osism.tech/kolla/redis                               2024.2    0fa4d8b23966   24 hours ago   278MB\nregistry.osism.tech/kolla/redis-sentinel                      2024.2    2052a297db29   24 hours ago   278MB\nregistry.osism.tech/kolla/horizon                             2024.2    4d92e9384e36   24 hours ago   1.15GB\nregistry.osism.tech/kolla/mariadb-server                      2024.2    5dbd7b932e63   24 hours ago   457MB\nregistry.osism.tech/kolla/prometheus-node-exporter            2024.2    278eaa73d02c   24 hours ago   311MB\nregistry.osism.tech/kolla/prometheus-elasticsearch-exporter   2024.2    13638ae36e57   24 hours ago   297MB\nregistry.osism.tech/kolla/prometheus-memcached-exporter       2024.2    8ff5a6b3cd16   24 hours ago   304MB\nregistry.osism.tech/kolla/prometheus-mysqld-exporter          2024.2    d9cb73d898bb   24 hours ago   306MB\nregistry.osism.tech/kolla/prometheus-cadvisor                 2024.2    4e4cb8e72dfc   24 hours ago   363MB\nregistry.osism.tech/kolla/ovn-nb-db-server                    2024.2    51d6ef378bbd   24 hours ago   846MB\nregistry.osism.tech/kolla/ovn-sb-db-server                    2024.2    25bb84f2289b   24 hours ago   846MB\nregistry.osism.tech/kolla/ovn-northd                          2024.2    47dae0f63bd5   24 hours ago   846MB\nregistry.osism.tech/kolla/ovn-controller                      2024.2    342e8da8ac3a   24 hours ago   846MB\nregistry.osism.tech/kolla/neutron-server                      2024.2    189daad12150   24 hours ago   1.17GB\nregistry.osism.tech/kolla/octavia-api                         2024.2    624955c2c115   24 hours ago   1.06GB\nregistry.osism.tech/kolla/octavia-housekeeping                2024.2    9b80c75ebf88   24 hours ago   1.03GB\nregistry.osism.tech/kolla/octavia-driver-agent                2024.2    3f4a0924d9ed   24 hours ago   1.06GB\nregistry.osism.tech/kolla/octavia-health-manager              2024.2    741914d697fa   24 hours ago   1.03GB\nregistry.osism.tech/kolla/octavia-worker                      2024.2    45d576cfc14a   24 hours ago   1.03GB\nregistry.osism.tech/kolla/cinder-backup                       2024.2    db0ba08b76e7   24 hours ago   1.42GB\nregistry.osism.tech/kolla/cinder-volume                       2024.2    aad4ba476f1c   24 hours ago   1.72GB\nregistry.osism.tech/kolla/cinder-scheduler                    2024.2    3102035cbd3f   24 hours ago   1.41GB\nregistry.osism.tech/kolla/cinder-api                          2024.2    ffa30f3933ed   24 hours ago   1.41GB\nregistry.osism.tech/kolla/placement-api                       2024.2    e4b48b9cb79e   24 hours ago   981MB\nregistry.osism.tech/kolla/designate-producer                  2024.2    e3b7812cd6ff   24 hours ago   990MB\nregistry.osism.tech/kolla/designate-central                   2024.2    8ab2ea11494a   24 hours ago   989MB\nregistry.osism.tech/kolla/designate-worker                    2024.2    13bd17ae9e62   24 hours ago   994MB\nregistry.osism.tech/kolla/designate-api                       2024.2    2e44006b8d8c   24 hours ago   990MB\nregistry.osism.tech/kolla/designate-backend-bind9             2024.2    7b60bbae1ff0   24 hours ago   994MB\nregistry.osism.tech/kolla/designate-mdns                      2024.2    a28e142a95bf   24 hours ago   990MB\nregistry.osism.tech/kolla/barbican-worker                     2024.2    0eec7e0f7b33   24 hours ago   996MB\nregistry.osism.tech/kolla/barbican-keystone-listener          2024.2    8c429b2daef4   24 hours ago   996MB\nregistry.osism.tech/kolla/barbican-api                        2024.2    9d4b161f78ee   24 hours ago   996MB\nregistry.osism.tech/kolla/keystone-fernet                     2024.2    3770c4e50a62   24 hours ago   1.04GB\nregistry.osism.tech/kolla/keystone-ssh                        2024.2    a75dcc713ad6   24 hours ago   1.05GB\nregistry.osism.tech/kolla/keystone                            2024.2    f3afb1708fa9   24 hours ago   1.07GB\nregistry.osism.tech/kolla/nova-conductor                      2024.2    7e0dfe6dfac8   24 hours ago   1.22GB\nregistry.osism.tech/kolla/nova-api                            2024.2    b141835a8ce3   24 hours ago   1.22GB\nregistry.osism.tech/kolla/nova-novncproxy                     2024.2    0e230dfe273c   24 hours ago   1.37GB\nregistry.osism.tech/kolla/nova-scheduler                      2024.2    1d5f182c3a75   24 hours ago   1.22GB\nregistry.osism.tech/kolla/magnum-api                          2024.2    5f8755d26e30   24 hours ago   1.13GB\nregistry.osism.tech/kolla/magnum-conductor                    2024.2    ad0e7877ae1a   24 hours ago   1.25GB\nregistry.osism.tech/kolla/glance-api                          2024.2    4e1be079f392   24 hours ago   1.1GB\n+ sh -c /opt/configuration/scripts/check-services.sh\n+ set -e\n+ source /opt/manager-vars.sh\n++ export NUMBER_OF_NODES=6\n++ NUMBER_OF_NODES=6\n++ export CEPH_VERSION=reef\n++ CEPH_VERSION=reef\n++ export CONFIGURATION_VERSION=main\n++ CONFIGURATION_VERSION=main\n++ export MANAGER_VERSION=latest\n++ MANAGER_VERSION=latest\n++ export OPENSTACK_VERSION=2024.2\n++ OPENSTACK_VERSION=2024.2\n++ export ARA=false\n++ ARA=false\n++ export DEPLOY_MODE=manager\n++ DEPLOY_MODE=manager\n++ export TEMPEST=true\n++ TEMPEST=true\n++ export IS_ZUUL=true\n++ IS_ZUUL=true\n++ export MANAGER_PUBLIC_IP_ADDRESS=81.163.192.195\n++ MANAGER_PUBLIC_IP_ADDRESS=81.163.192.195\n++ export EXTERNAL_API=false\n++ EXTERNAL_API=false\n++ export IMAGE_USER=ubuntu\n++ IMAGE_USER=ubuntu\n++ export IMAGE_NODE_USER=ubuntu\n++ IMAGE_NODE_USER=ubuntu\n++ export CEPH_STACK=ceph-ansible\n++ CEPH_STACK=ceph-ansible\n+ [[ ceph-ansible == \\c\\e\\p\\h\\-\\a\\n\\s\\i\\b\\l\\e ]]\n+ sh -c /opt/configuration/scripts/check/100-ceph-with-ansible.sh\n+ set -e\n+ source /opt/configuration/scripts/include.sh\n++ export INTERACTIVE=false\n++ INTERACTIVE=false\n++ export OSISM_APPLY_RETRY=1\n++ OSISM_APPLY_RETRY=1\n+ source /opt/configuration/scripts/manager-version.sh\n+++ awk '-F: ' '/^manager_version:/ { print $2 }' /opt/configuration/environments/manager/configuration.yml\n\n# Ceph status\n\n++ export MANAGER_VERSION=latest\n++ MANAGER_VERSION=latest\n+ echo\n+ echo '# Ceph status'\n+ echo\n+ ceph -s\n  cluster:\n    id:     11111111-1111-1111-1111-111111111111\n    health: HEALTH_OK\n \n  services:\n    mon: 3 daemons, quorum testbed-node-0,testbed-node-1,testbed-node-2 (age 31m)\n    mgr: testbed-node-0(active, since 17m), standbys: testbed-node-1, testbed-node-2\n    mds: 1/1 daemons up, 2 standby\n    osd: 6 osds: 6 up (since 27m), 6 in (since 27m)\n    rgw: 3 daemons active (3 hosts, 1 zones)\n \n  data:\n    volumes: 1/1 healthy\n    pools:   14 pools, 401 pgs\n    objects: 554 objects, 2.2 GiB\n    usage:   7.1 GiB used, 113 GiB / 120 GiB avail\n    pgs:     401 active+clean\n \n\n# Ceph versions\n\n+ echo\n+ echo '# Ceph versions'\n+ echo\n+ ceph versions\n{\n    \"mon\": {\n        \"ceph version 18.2.7 (6b0e988052ec84cf2d4a54ff9bbbc5e720b621ad) reef (stable)\": 3\n    },\n    \"mgr\": {\n        \"ceph version 18.2.7 (6b0e988052ec84cf2d4a54ff9bbbc5e720b621ad) reef (stable)\": 3\n    },\n    \"osd\": {\n        \"ceph version 18.2.7 (6b0e988052ec84cf2d4a54ff9bbbc5e720b621ad) reef (stable)\": 6\n    },\n    \"mds\": {\n        \"ceph version 18.2.7 (6b0e988052ec84cf2d4a54ff9bbbc5e720b621ad) reef (stable)\": 3\n    },\n    \"rgw\": {\n        \"ceph version 18.2.7 (6b0e988052ec84cf2d4a54ff9bbbc5e720b621ad) reef (stable)\": 3\n    },\n    \"overall\": {\n        \"ceph version 18.2.7 (6b0e988052ec84cf2d4a54ff9bbbc5e720b621ad) reef (stable)\": 18\n    }\n}\n\n# Ceph OSD tree\n\n+ echo\n+ echo '# Ceph OSD tree'\n+ echo\n+ ceph osd df tree\nID  CLASS  WEIGHT   REWEIGHT  SIZE     RAW USE  DATA     OMAP     META     AVAIL    %USE  VAR   PGS  STATUS  TYPE NAME              \n-1         0.11691         -  120 GiB  7.1 GiB  6.7 GiB    6 KiB  412 MiB  113 GiB  5.90  1.00    -          root default           \n-5         0.03897         -   40 GiB  2.4 GiB  2.2 GiB    2 KiB  143 MiB   38 GiB  5.91  1.00    -              host testbed-node-3\n 0    hdd  0.01949   1.00000   20 GiB  1.1 GiB  1.0 GiB    1 KiB   74 MiB   19 GiB  5.57  0.94  190      up          osd.0          \n 4    hdd  0.01949   1.00000   20 GiB  1.3 GiB  1.2 GiB    1 KiB   70 MiB   19 GiB  6.26  1.06  202      up          osd.4          \n-7         0.03897         -   40 GiB  2.4 GiB  2.2 GiB    2 KiB  143 MiB   38 GiB  5.91  1.00    -              host testbed-node-4\n 1    hdd  0.01949   1.00000   20 GiB  1.5 GiB  1.4 GiB    1 KiB   74 MiB   19 GiB  7.41  1.26  197      up          osd.1          \n 5    hdd  0.01949   1.00000   20 GiB  904 MiB  835 MiB    1 KiB   70 MiB   19 GiB  4.42  0.75  191      up          osd.5          \n-3         0.03897         -   40 GiB  2.3 GiB  2.2 GiB    2 KiB  126 MiB   38 GiB  5.87  0.99    -              host testbed-node-5\n 2    hdd  0.01949   1.00000   20 GiB  984 MiB  914 MiB    1 KiB   70 MiB   19 GiB  4.81  0.82  196      up          osd.2          \n 3    hdd  0.01949   1.00000   20 GiB  1.4 GiB  1.3 GiB    1 KiB   56 MiB   19 GiB  6.93  1.17  194      up          osd.3          \n                       TOTAL  120 GiB  7.1 GiB  6.7 GiB  9.3 KiB  412 MiB  113 GiB  5.90                                            \nMIN/MAX VAR: 0.75/1.26  STDDEV: 1.08\n\n# Ceph monitor status\n\n+ echo\n+ echo '# Ceph monitor status'\n+ echo\n+ ceph mon stat\ne1: 3 mons at {testbed-node-0=[v2:192.168.16.10:3300/0,v1:192.168.16.10:6789/0],testbed-node-1=[v2:192.168.16.11:3300/0,v1:192.168.16.11:6789/0],testbed-node-2=[v2:192.168.16.12:3300/0,v1:192.168.16.12:6789/0]} removed_ranks: {} disallowed_leaders: {}, election epoch 6, leader 0 testbed-node-0, quorum 0,1,2 testbed-node-0,testbed-node-1,testbed-node-2\n\n# Ceph quorum status\n\n+ echo\n+ echo '# Ceph quorum status'\n+ echo\n+ ceph quorum_status\n+ jq\n{\n  \"election_epoch\": 6,\n  \"quorum\": [\n    0,\n    1,\n    2\n  ],\n  \"quorum_names\": [\n    \"testbed-node-0\",\n    \"testbed-node-1\",\n    \"testbed-node-2\"\n  ],\n  \"quorum_leader_name\": \"testbed-node-0\",\n  \"quorum_age\": 1877,\n  \"features\": {\n    \"quorum_con\": \"4540138322906710015\",\n    \"quorum_mon\": [\n      \"kraken\",\n      \"luminous\",\n      \"mimic\",\n      \"osdmap-prune\",\n      \"nautilus\",\n      \"octopus\",\n      \"pacific\",\n      \"elector-pinging\",\n      \"quincy\",\n      \"reef\"\n    ]\n  },\n  \"monmap\": {\n    \"epoch\": 1,\n    \"fsid\": \"11111111-1111-1111-1111-111111111111\",\n    \"modified\": \"2026-02-24T00:51:03.217233Z\",\n    \"created\": \"2026-02-24T00:51:03.217233Z\",\n    \"min_mon_release\": 18,\n    \"min_mon_release_name\": \"reef\",\n    \"election_strategy\": 1,\n    \"disallowed_leaders: \": \"\",\n    \"stretch_mode\": false,\n    \"tiebreaker_mon\": \"\",\n    \"removed_ranks: \": \"\",\n    \"features\": {\n      \"persistent\": [\n        \"kraken\",\n        \"luminous\",\n        \"mimic\",\n        \"osdmap-prune\",\n        \"nautilus\",\n        \"octopus\",\n        \"pacific\",\n        \"elector-pinging\",\n        \"quincy\",\n        \"reef\"\n      ],\n      \"optional\": []\n    },\n    \"mons\": [\n      {\n        \"rank\": 0,\n        \"name\": \"testbed-node-0\",\n        \"public_addrs\": {\n          \"addrvec\": [\n            {\n              \"type\": \"v2\",\n              \"addr\": \"192.168.16.10:3300\",\n              \"nonce\": 0\n            },\n            {\n              \"type\": \"v1\",\n              \"addr\": \"192.168.16.10:6789\",\n              \"nonce\": 0\n            }\n          ]\n        },\n        \"addr\": \"192.168.16.10:6789/0\",\n        \"public_addr\": \"192.168.16.10:6789/0\",\n        \"priority\": 0,\n        \"weight\": 0,\n        \"crush_location\": \"{}\"\n      },\n      {\n        \"rank\": 1,\n        \"name\": \"testbed-node-1\",\n        \"public_addrs\": {\n          \"addrvec\": [\n            {\n              \"type\": \"v2\",\n              \"addr\": \"192.168.16.11:3300\",\n              \"nonce\": 0\n            },\n            {\n              \"type\": \"v1\",\n              \"addr\": \"192.168.16.11:6789\",\n              \"nonce\": 0\n            }\n          ]\n        },\n        \"addr\": \"192.168.16.11:6789/0\",\n        \"public_addr\": \"192.168.16.11:6789/0\",\n        \"priority\": 0,\n        \"weight\": 0,\n        \"crush_location\": \"{}\"\n      },\n      {\n        \"rank\": 2,\n        \"name\": \"testbed-node-2\",\n        \"public_addrs\": {\n          \"addrvec\": [\n            {\n              \"type\": \"v2\",\n              \"addr\": \"192.168.16.12:3300\",\n              \"nonce\": 0\n            },\n            {\n              \"type\": \"v1\",\n              \"addr\": \"192.168.16.12:6789\",\n              \"nonce\": 0\n            }\n          ]\n        },\n        \"addr\": \"192.168.16.12:6789/0\",\n        \"public_addr\": \"192.168.16.12:6789/0\",\n        \"priority\": 0,\n        \"weight\": 0,\n        \"crush_location\": \"{}\"\n      }\n    ]\n  }\n}\n\n# Ceph free space status\n\n+ echo\n+ echo '# Ceph free space status'\n+ echo\n+ ceph df\n--- RAW STORAGE ---\nCLASS     SIZE    AVAIL     USED  RAW USED  %RAW USED\nhdd    120 GiB  113 GiB  7.1 GiB   7.1 GiB       5.90\nTOTAL  120 GiB  113 GiB  7.1 GiB   7.1 GiB       5.90\n \n--- POOLS ---\nPOOL                       ID  PGS   STORED  OBJECTS     USED  %USED  MAX AVAIL\n.mgr                        1    1  577 KiB        2  1.1 MiB      0     53 GiB\ncephfs_data                 2   32      0 B        0      0 B      0     35 GiB\ncephfs_metadata             3   16  4.4 KiB       22   96 KiB      0     35 GiB\ndefault.rgw.buckets.data    4   32      0 B        0      0 B      0     35 GiB\ndefault.rgw.buckets.index   5   32      0 B        0      0 B      0     35 GiB\ndefault.rgw.control         6   32      0 B        8      0 B      0     35 GiB\ndefault.rgw.log             7   32  3.6 KiB      209  408 KiB      0     35 GiB\ndefault.rgw.meta            8   32      0 B        0      0 B      0     35 GiB\n.rgw.root                   9   32  2.6 KiB        6   48 KiB      0     53 GiB\nbackups                    10   32     19 B        2   12 KiB      0     35 GiB\nvolumes                    11   32     19 B        2   12 KiB      0     35 GiB\nimages                     12   32  2.2 GiB      299  6.7 GiB   5.96     35 GiB\nmetrics                    13   32     19 B        2   12 KiB      0     35 GiB\nvms                        14   32     19 B        2   12 KiB      0     35 GiB\n++ semver latest 5.0.0\n+ [[ -1 -eq -1 ]]\n+ [[ latest != \\l\\a\\t\\e\\s\\t ]]\n+ [[ ! -e /etc/redhat-release ]]\n+ osism apply facts\n\u001b[32m2026-02-24 01:22:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mPrepare task for execution of facts.\u001b[0m\n\u001b[32m2026-02-24 01:22:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 87d1577d-7457-4d6b-809f-ff0d65828ac8 (facts) was prepared for execution.\u001b[0m\n\u001b[32m2026-02-24 01:22:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task 87d1577d-7457-4d6b-809f-ff0d65828ac8 (facts) has been started and output is visible here.\u001b[0m\n\nPLAY [Apply role facts] ********************************************************\n\nTASK [osism.commons.facts : Create custom facts directory] *********************\nTuesday 24 February 2026  01:22:52 +0000 (0:00:00.284)       0:00:00.284 ****** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [osism.commons.facts : Copy fact files] ***********************************\nTuesday 24 February 2026  01:22:53 +0000 (0:00:01.573)       0:00:01.857 ****** \n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nPLAY [Gather facts for all hosts] **********************************************\n\nTASK [Gathers facts about hosts] ***********************************************\nTuesday 24 February 2026  01:22:55 +0000 (0:00:01.635)       0:00:03.493 ****** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nPLAY [Gather facts for all hosts if using --limit] *****************************\n\nTASK [Gather facts for all hosts] **********************************************\nTuesday 24 February 2026  01:23:00 +0000 (0:00:05.428)       0:00:08.921 ****** \n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;32mtestbed-manager\u001b[0m            : \u001b[0;32mok=2   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   \n\u001b[0;32mtestbed-node-0\u001b[0m             : \u001b[0;32mok=2   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   \n\u001b[0;32mtestbed-node-1\u001b[0m             : \u001b[0;32mok=2   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   \n\u001b[0;32mtestbed-node-2\u001b[0m             : \u001b[0;32mok=2   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   \n\u001b[0;32mtestbed-node-3\u001b[0m             : \u001b[0;32mok=2   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   \n\u001b[0;32mtestbed-node-4\u001b[0m             : \u001b[0;32mok=2   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   \n\u001b[0;32mtestbed-node-5\u001b[0m             : \u001b[0;32mok=2   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nTuesday 24 February 2026  01:23:01 +0000 (0:00:00.650)       0:00:09.571 ****** \n=============================================================================== \nGathers facts about hosts ----------------------------------------------- 5.43s\nosism.commons.facts : Copy fact files ----------------------------------- 1.64s\nosism.commons.facts : Create custom facts directory --------------------- 1.57s\nGather facts for all hosts ---------------------------------------------- 0.65s\n+ osism validate ceph-mons\n\nPLAY [Ceph validate mons] ******************************************************\n\nTASK [Get timestamp for report file] *******************************************\nTuesday 24 February 2026  01:23:20 +0000 (0:00:00.517)       0:00:00.517 ****** \n\u001b[0;32mok: [testbed-node-0 -> testbed-manager(192.168.16.5)]\u001b[0m\n\nTASK [Create report output directory] ******************************************\nTuesday 24 February 2026  01:23:21 +0000 (0:00:00.994)       0:00:01.511 ****** \n\u001b[0;33mchanged: [testbed-node-0 -> testbed-manager(192.168.16.5)]\u001b[0m\n\nTASK [Define report vars] ******************************************************\nTuesday 24 February 2026  01:23:22 +0000 (0:00:01.123)       0:00:02.635 ****** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [Prepare test data for container existance test] **************************\nTuesday 24 February 2026  01:23:22 +0000 (0:00:00.129)       0:00:02.764 ****** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [Get container info] ******************************************************\nTuesday 24 February 2026  01:23:22 +0000 (0:00:00.336)       0:00:03.101 ****** \n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [Set test result to failed if container is missing] ***********************\nTuesday 24 February 2026  01:23:24 +0000 (0:00:01.113)       0:00:04.215 ****** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [Set test result to passed if container is existing] **********************\nTuesday 24 February 2026  01:23:24 +0000 (0:00:00.347)       0:00:04.562 ****** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [Prepare test data] *******************************************************\nTuesday 24 February 2026  01:23:24 +0000 (0:00:00.559)       0:00:05.122 ****** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [Set test result to failed if ceph-mon is not running] ********************\nTuesday 24 February 2026  01:23:25 +0000 (0:00:00.366)       0:00:05.488 ****** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [Set test result to passed if ceph-mon is running] ************************\nTuesday 24 February 2026  01:23:25 +0000 (0:00:00.335)       0:00:05.823 ****** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [Aggregate test results step one] *****************************************\nTuesday 24 February 2026  01:23:26 +0000 (0:00:00.545)       0:00:06.369 ****** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [Aggregate test results step two] *****************************************\nTuesday 24 February 2026  01:23:26 +0000 (0:00:00.281)       0:00:06.651 ****** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [Aggregate test results step three] ***************************************\nTuesday 24 February 2026  01:23:26 +0000 (0:00:00.266)       0:00:06.918 ****** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [Flush handlers] **********************************************************\nTuesday 24 February 2026  01:23:27 +0000 (0:00:00.264)       0:00:07.182 ****** \n\nTASK [Flush handlers] **********************************************************\nTuesday 24 February 2026  01:23:27 +0000 (0:00:00.077)       0:00:07.260 ****** \n\nTASK [Flush handlers] **********************************************************\nTuesday 24 February 2026  01:23:27 +0000 (0:00:00.073)       0:00:07.333 ****** \n\nTASK [Print report file information] *******************************************\nTuesday 24 February 2026  01:23:27 +0000 (0:00:00.074)       0:00:07.408 ****** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [Fail due to missing containers] ******************************************\nTuesday 24 February 2026  01:23:27 +0000 (0:00:00.264)       0:00:07.672 ****** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [Prepare quorum test vars] ************************************************\nTuesday 24 February 2026  01:23:27 +0000 (0:00:00.271)       0:00:07.944 ****** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [Get monmap info from one mon container] **********************************\nTuesday 24 February 2026  01:23:27 +0000 (0:00:00.134)       0:00:08.079 ****** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [Set quorum test data] ****************************************************\nTuesday 24 February 2026  01:23:29 +0000 (0:00:01.709)       0:00:09.788 ****** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [Fail quorum test if not all monitors are in quorum] **********************\nTuesday 24 February 2026  01:23:30 +0000 (0:00:00.585)       0:00:10.374 ****** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [Pass quorum test if all monitors are in quorum] **************************\nTuesday 24 February 2026  01:23:30 +0000 (0:00:00.130)       0:00:10.505 ****** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [Set fsid test vars] ******************************************************\nTuesday 24 February 2026  01:23:30 +0000 (0:00:00.426)       0:00:10.932 ****** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [Fail Cluster FSID test if FSID does not match configuration] *************\nTuesday 24 February 2026  01:23:31 +0000 (0:00:00.334)       0:00:11.267 ****** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [Pass Cluster FSID test if it matches configuration] **********************\nTuesday 24 February 2026  01:23:31 +0000 (0:00:00.123)       0:00:11.390 ****** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [Prepare status test vars] ************************************************\nTuesday 24 February 2026  01:23:31 +0000 (0:00:00.128)       0:00:11.519 ****** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [Gather status data] ******************************************************\nTuesday 24 February 2026  01:23:31 +0000 (0:00:00.128)       0:00:11.648 ****** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [Set health test data] ****************************************************\nTuesday 24 February 2026  01:23:32 +0000 (0:00:01.508)       0:00:13.157 ****** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [Fail cluster-health if health is not acceptable] *************************\nTuesday 24 February 2026  01:23:33 +0000 (0:00:00.380)       0:00:13.537 ****** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [Pass cluster-health if health is acceptable] *****************************\nTuesday 24 February 2026  01:23:33 +0000 (0:00:00.154)       0:00:13.692 ****** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [Fail cluster-health if health is not acceptable (strict)] ****************\nTuesday 24 February 2026  01:23:33 +0000 (0:00:00.161)       0:00:13.854 ****** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [Pass cluster-health if status is OK (strict)] ****************************\nTuesday 24 February 2026  01:23:34 +0000 (0:00:00.379)       0:00:14.233 ****** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [Set validation result to passed if no test failed] ***********************\nTuesday 24 February 2026  01:23:34 +0000 (0:00:00.142)       0:00:14.376 ****** \n\u001b[0;32mok: [testbed-node-0 -> testbed-manager(192.168.16.5)]\u001b[0m\n\nTASK [Set validation result to failed if a test failed] ************************\nTuesday 24 February 2026  01:23:34 +0000 (0:00:00.283)       0:00:14.659 ****** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [Aggregate test results step one] *****************************************\nTuesday 24 February 2026  01:23:34 +0000 (0:00:00.268)       0:00:14.928 ****** \n\u001b[0;32mok: [testbed-node-0 -> testbed-manager(192.168.16.5)]\u001b[0m\n\nTASK [Aggregate test results step two] *****************************************\nTuesday 24 February 2026  01:23:36 +0000 (0:00:01.921)       0:00:16.850 ****** \n\u001b[0;32mok: [testbed-node-0 -> testbed-manager(192.168.16.5)]\u001b[0m\n\nTASK [Aggregate test results step three] ***************************************\nTuesday 24 February 2026  01:23:36 +0000 (0:00:00.259)       0:00:17.109 ****** \n\u001b[0;33mchanged: [testbed-node-0 -> testbed-manager(192.168.16.5)]\u001b[0m\n\nTASK [Flush handlers] **********************************************************\nTuesday 24 February 2026  01:23:37 +0000 (0:00:00.312)       0:00:17.422 ****** \n\nTASK [Flush handlers] **********************************************************\nTuesday 24 February 2026  01:23:37 +0000 (0:00:00.077)       0:00:17.499 ****** \n\nTASK [Flush handlers] **********************************************************\nTuesday 24 February 2026  01:23:37 +0000 (0:00:00.072)       0:00:17.572 ****** \n\nRUNNING HANDLER [Write report file] ********************************************\nTuesday 24 February 2026  01:23:37 +0000 (0:00:00.079)       0:00:17.651 ****** \n\u001b[0;33mchanged: [testbed-node-0 -> testbed-manager(192.168.16.5)]\u001b[0m\n\nTASK [Print report file information] *******************************************\nTuesday 24 February 2026  01:23:39 +0000 (0:00:01.717)       0:00:19.369 ****** \n\u001b[0;32mok: [testbed-node-0 -> testbed-manager(192.168.16.5)] => {\u001b[0m\n\u001b[0;32m    \"msg\": [\u001b[0m\n\u001b[0;32m        \"Validator run completed.\",\u001b[0m\n\u001b[0;32m        \"You can find the report file here:\",\u001b[0m\n\u001b[0;32m        \"/opt/reports/validator/ceph-mons-validator-2026-02-24T01:23:21+00:00-report.json\",\u001b[0m\n\u001b[0;32m        \"on the following host:\",\u001b[0m\n\u001b[0;32m        \"testbed-manager\"\u001b[0m\n\u001b[0;32m    ]\u001b[0m\n\u001b[0;32m}\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=24  \u001b[0m \u001b[0;33mchanged=5   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=13  \u001b[0m rescued=0    ignored=0   \n\u001b[0;32mtestbed-node-1\u001b[0m             : \u001b[0;32mok=5   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   \n\u001b[0;32mtestbed-node-2\u001b[0m             : \u001b[0;32mok=5   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nTuesday 24 February 2026  01:23:40 +0000 (0:00:01.070)       0:00:20.440 ****** \n=============================================================================== \nAggregate test results step one ----------------------------------------- 1.92s\nWrite report file ------------------------------------------------------- 1.72s\nGet monmap info from one mon container ---------------------------------- 1.71s\nGather status data ------------------------------------------------------ 1.51s\nCreate report output directory ------------------------------------------ 1.12s\nGet container info ------------------------------------------------------ 1.11s\nPrint report file information ------------------------------------------- 1.07s\nGet timestamp for report file ------------------------------------------- 0.99s\nSet quorum test data ---------------------------------------------------- 0.59s\nSet test result to passed if container is existing ---------------------- 0.56s\nSet test result to passed if ceph-mon is running ------------------------ 0.55s\nPass quorum test if all monitors are in quorum -------------------------- 0.43s\nSet health test data ---------------------------------------------------- 0.38s\nFail cluster-health if health is not acceptable (strict) ---------------- 0.38s\nPrepare test data ------------------------------------------------------- 0.37s\nSet test result to failed if container is missing ----------------------- 0.35s\nPrepare test data for container existance test -------------------------- 0.34s\nSet test result to failed if ceph-mon is not running -------------------- 0.34s\nSet fsid test vars ------------------------------------------------------ 0.33s\nAggregate test results step three --------------------------------------- 0.31s\n+ osism validate ceph-mgrs\n\nPLAY [Ceph validate mgrs] ******************************************************\n\nTASK [Get timestamp for report file] *******************************************\nTuesday 24 February 2026  01:23:58 +0000 (0:00:00.537)       0:00:00.537 ****** \n\u001b[0;32mok: [testbed-node-0 -> testbed-manager(192.168.16.5)]\u001b[0m\n\nTASK [Create report output directory] ******************************************\nTuesday 24 February 2026  01:23:59 +0000 (0:00:01.035)       0:00:01.573 ****** \n\u001b[0;32mok: [testbed-node-0 -> testbed-manager(192.168.16.5)]\u001b[0m\n\nTASK [Define report vars] ******************************************************\nTuesday 24 February 2026  01:24:01 +0000 (0:00:01.228)       0:00:02.801 ****** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [Prepare test data for container existance test] **************************\nTuesday 24 February 2026  01:24:01 +0000 (0:00:00.148)       0:00:02.950 ****** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [Get container info] ******************************************************\nTuesday 24 February 2026  01:24:01 +0000 (0:00:00.352)       0:00:03.303 ****** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [Set test result to failed if container is missing] ***********************\nTuesday 24 February 2026  01:24:02 +0000 (0:00:01.087)       0:00:04.391 ****** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [Set test result to passed if container is existing] **********************\nTuesday 24 February 2026  01:24:03 +0000 (0:00:00.328)       0:00:04.719 ****** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [Prepare test data] *******************************************************\nTuesday 24 February 2026  01:24:03 +0000 (0:00:00.597)       0:00:05.317 ****** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [Set test result to failed if ceph-mgr is not running] ********************\nTuesday 24 February 2026  01:24:04 +0000 (0:00:00.337)       0:00:05.654 ****** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [Set test result to passed if ceph-mgr is running] ************************\nTuesday 24 February 2026  01:24:04 +0000 (0:00:00.330)       0:00:05.985 ****** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [Aggregate test results step one] *****************************************\nTuesday 24 February 2026  01:24:04 +0000 (0:00:00.552)       0:00:06.537 ****** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [Aggregate test results step two] *****************************************\nTuesday 24 February 2026  01:24:05 +0000 (0:00:00.306)       0:00:06.843 ****** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [Aggregate test results step three] ***************************************\nTuesday 24 February 2026  01:24:05 +0000 (0:00:00.278)       0:00:07.122 ****** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [Flush handlers] **********************************************************\nTuesday 24 February 2026  01:24:05 +0000 (0:00:00.278)       0:00:07.400 ****** \n\nTASK [Flush handlers] **********************************************************\nTuesday 24 February 2026  01:24:05 +0000 (0:00:00.071)       0:00:07.472 ****** \n\nTASK [Flush handlers] **********************************************************\nTuesday 24 February 2026  01:24:05 +0000 (0:00:00.072)       0:00:07.545 ****** \n\nTASK [Print report file information] *******************************************\nTuesday 24 February 2026  01:24:06 +0000 (0:00:00.087)       0:00:07.632 ****** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [Fail due to missing containers] ******************************************\nTuesday 24 February 2026  01:24:06 +0000 (0:00:00.274)       0:00:07.906 ****** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [Define mgr module test vars] *********************************************\nTuesday 24 February 2026  01:24:06 +0000 (0:00:00.255)       0:00:08.162 ****** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [Gather list of mgr modules] **********************************************\nTuesday 24 February 2026  01:24:06 +0000 (0:00:00.121)       0:00:08.283 ****** \n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\nTASK [Parse mgr module list from json] *****************************************\nTuesday 24 February 2026  01:24:08 +0000 (0:00:02.118)       0:00:10.402 ****** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [Extract list of enabled mgr modules] *************************************\nTuesday 24 February 2026  01:24:09 +0000 (0:00:00.514)       0:00:10.917 ****** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [Fail test if mgr modules are disabled that should be enabled] ************\nTuesday 24 February 2026  01:24:09 +0000 (0:00:00.372)       0:00:11.289 ****** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [Pass test if required mgr modules are enabled] ***************************\nTuesday 24 February 2026  01:24:09 +0000 (0:00:00.155)       0:00:11.444 ****** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\nTASK [Set validation result to passed if no test failed] ***********************\nTuesday 24 February 2026  01:24:10 +0000 (0:00:00.181)       0:00:11.626 ****** \n\u001b[0;32mok: [testbed-node-0 -> testbed-manager(192.168.16.5)]\u001b[0m\n\nTASK [Set validation result to failed if a test failed] ************************\nTuesday 24 February 2026  01:24:10 +0000 (0:00:00.284)       0:00:11.910 ****** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\nTASK [Aggregate test results step one] *****************************************\nTuesday 24 February 2026  01:24:10 +0000 (0:00:00.323)       0:00:12.233 ****** \n\u001b[0;32mok: [testbed-node-0 -> testbed-manager(192.168.16.5)]\u001b[0m\n\nTASK [Aggregate test results step two] *****************************************\nTuesday 24 February 2026  01:24:12 +0000 (0:00:01.460)       0:00:13.694 ****** \n\u001b[0;32mok: [testbed-node-0 -> testbed-manager(192.168.16.5)]\u001b[0m\n\nTASK [Aggregate test results step three] ***************************************\nTuesday 24 February 2026  01:24:12 +0000 (0:00:00.311)       0:00:14.005 ****** \n\u001b[0;33mchanged: [testbed-node-0 -> testbed-manager(192.168.16.5)]\u001b[0m\n\nTASK [Flush handlers] **********************************************************\nTuesday 24 February 2026  01:24:12 +0000 (0:00:00.292)       0:00:14.298 ****** \n\nTASK [Flush handlers] **********************************************************\nTuesday 24 February 2026  01:24:12 +0000 (0:00:00.089)       0:00:14.387 ****** \n\nTASK [Flush handlers] **********************************************************\nTuesday 24 February 2026  01:24:12 +0000 (0:00:00.073)       0:00:14.461 ****** \n\nRUNNING HANDLER [Write report file] ********************************************\nTuesday 24 February 2026  01:24:13 +0000 (0:00:00.372)       0:00:14.834 ****** \n\u001b[0;33mchanged: [testbed-node-0 -> testbed-manager(192.168.16.5)]\u001b[0m\n\nTASK [Print report file information] *******************************************\nTuesday 24 February 2026  01:24:14 +0000 (0:00:01.575)       0:00:16.409 ****** \n\u001b[0;32mok: [testbed-node-0 -> testbed-manager(192.168.16.5)] => {\u001b[0m\n\u001b[0;32m    \"msg\": [\u001b[0m\n\u001b[0;32m        \"Validator run completed.\",\u001b[0m\n\u001b[0;32m        \"You can find the report file here:\",\u001b[0m\n\u001b[0;32m        \"/opt/reports/validator/ceph-mgrs-validator-2026-02-24T01:23:59+00:00-report.json\",\u001b[0m\n\u001b[0;32m        \"on the following host:\",\u001b[0m\n\u001b[0;32m        \"testbed-manager\"\u001b[0m\n\u001b[0;32m    ]\u001b[0m\n\u001b[0;32m}\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=19  \u001b[0m \u001b[0;33mchanged=3   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=9   \u001b[0m rescued=0    ignored=0   \n\u001b[0;32mtestbed-node-1\u001b[0m             : \u001b[0;32mok=5   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   \n\u001b[0;32mtestbed-node-2\u001b[0m             : \u001b[0;32mok=5   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nTuesday 24 February 2026  01:24:15 +0000 (0:00:00.485)       0:00:16.895 ****** \n=============================================================================== \nGather list of mgr modules ---------------------------------------------- 2.12s\nWrite report file ------------------------------------------------------- 1.58s\nAggregate test results step one ----------------------------------------- 1.46s\nCreate report output directory ------------------------------------------ 1.23s\nGet container info ------------------------------------------------------ 1.09s\nGet timestamp for report file ------------------------------------------- 1.04s\nSet test result to passed if container is existing ---------------------- 0.60s\nSet test result to passed if ceph-mgr is running ------------------------ 0.55s\nFlush handlers ---------------------------------------------------------- 0.54s\nParse mgr module list from json ----------------------------------------- 0.51s\nPrint report file information ------------------------------------------- 0.49s\nExtract list of enabled mgr modules ------------------------------------- 0.37s\nPrepare test data for container existance test -------------------------- 0.35s\nPrepare test data ------------------------------------------------------- 0.34s\nSet test result to failed if ceph-mgr is not running -------------------- 0.33s\nSet test result to failed if container is missing ----------------------- 0.33s\nSet validation result to failed if a test failed ------------------------ 0.32s\nAggregate test results step two ----------------------------------------- 0.31s\nAggregate test results step one ----------------------------------------- 0.31s\nAggregate test results step three --------------------------------------- 0.29s\n+ osism validate ceph-osds\n\nPLAY [Ceph validate OSDs] ******************************************************\n\nTASK [Get timestamp for report file] *******************************************\nTuesday 24 February 2026  01:24:33 +0000 (0:00:00.439)       0:00:00.439 ****** \n\u001b[0;32mok: [testbed-node-3 -> testbed-manager(192.168.16.5)]\u001b[0m\n\nTASK [Get extra vars for Ceph configuration] ***********************************\nTuesday 24 February 2026  01:24:34 +0000 (0:00:00.766)       0:00:01.205 ****** \n\u001b[0;32mok: [testbed-node-3 -> testbed-manager(192.168.16.5)]\u001b[0m\n\nTASK [Create report output directory] ******************************************\nTuesday 24 February 2026  01:24:34 +0000 (0:00:00.457)       0:00:01.663 ****** \n\u001b[0;32mok: [testbed-node-3 -> testbed-manager(192.168.16.5)]\u001b[0m\n\nTASK [Define report vars] ******************************************************\nTuesday 24 February 2026  01:24:35 +0000 (0:00:00.733)       0:00:02.397 ****** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\nTASK [Define OSD test variables] ***********************************************\nTuesday 24 February 2026  01:24:35 +0000 (0:00:00.154)       0:00:02.551 ****** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Calculate OSD devices for each host] *************************************\nTuesday 24 February 2026  01:24:35 +0000 (0:00:00.130)       0:00:02.681 ****** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Define OSD test variables] ***********************************************\nTuesday 24 February 2026  01:24:36 +0000 (0:00:00.336)       0:00:03.018 ****** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\nTASK [Calculate OSD devices for each host] *************************************\nTuesday 24 February 2026  01:24:36 +0000 (0:00:00.171)       0:00:03.190 ****** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [Calculate total number of OSDs in cluster] *******************************\nTuesday 24 February 2026  01:24:36 +0000 (0:00:00.412)       0:00:03.602 ****** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\nTASK [Prepare test data] *******************************************************\nTuesday 24 February 2026  01:24:37 +0000 (0:00:00.979)       0:00:04.582 ****** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [Get list of ceph-osd containers on host] *********************************\nTuesday 24 February 2026  01:24:38 +0000 (0:00:00.343)       0:00:04.925 ****** \n\u001b[0;36mskipping: [testbed-node-3] => (item={'id': 'a507e39101423fe0593a98563102285dee1b9fd5e4e5d6a1d563e1f610290f5c', 'image': 'registry.osism.tech/kolla/nova-compute:2024.2', 'name': '/nova_compute', 'state': 'running', 'status': 'Up 7 minutes (healthy)'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'id': '15f1497f24ccac119e571c3ca4ed4919a800516c5394fe9fa749128c8c905fea', 'image': 'registry.osism.tech/kolla/nova-libvirt:2024.2', 'name': '/nova_libvirt', 'state': 'running', 'status': 'Up 7 minutes (healthy)'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'id': 'e2e543e63c1632d677daa97c5cbf37d6ea88a08477e7bc52d6aee09aadb8e6db', 'image': 'registry.osism.tech/kolla/nova-ssh:2024.2', 'name': '/nova_ssh', 'state': 'running', 'status': 'Up 8 minutes (healthy)'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'id': '4311ed30834d73e01f23c87613b03c0e1bd426d330434ac4d2c4d49561ff357d', 'image': 'registry.osism.tech/kolla/prometheus-libvirt-exporter:2024.2', 'name': '/prometheus_libvirt_exporter', 'state': 'running', 'status': 'Up 13 minutes'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'id': '8522e1cfe839b6a8a827f4a855f7c976c62d9037c614a70a5d252a6fb85efab0', 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2024.2', 'name': '/prometheus_cadvisor', 'state': 'running', 'status': 'Up 13 minutes'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'id': 'ef9cefa9065cb8fc86482d2449bb51824e0ef00be0eab0e05343b55eb0f777aa', 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2024.2', 'name': '/prometheus_node_exporter', 'state': 'running', 'status': 'Up 14 minutes'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'id': '6f45def46706eada0c2405777dcccf6426aed54715f85d10cd9f8e7453251571', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'name': '/neutron_ovn_metadata_agent', 'state': 'running', 'status': 'Up 15 minutes (healthy)'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'id': '4312c9fd5889f17b2cb033af2d2a0b48880ec04b3615d5a47f26b51a8883f40f', 'image': 'registry.osism.tech/osism/ceph-daemon:reef', 'name': '/ceph-rgw-default-testbed-node-3-rgw0', 'state': 'running', 'status': 'Up 24 minutes'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'id': '9c4e10611bcd267b7394da6deb0a1c6ee24c1bef4a599ad52f10da4eec0d178b', 'image': 'registry.osism.tech/osism/ceph-daemon:reef', 'name': '/ceph-mds-testbed-node-3', 'state': 'running', 'status': 'Up 25 minutes'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'id': '89ba6407e0e4d76b96f76ae5868d3f4282281fb62ff7b462fc5a6e487127dd67', 'image': 'registry.osism.tech/osism/ceph-daemon:reef', 'name': '/ceph-crash-testbed-node-3', 'state': 'running', 'status': 'Up 26 minutes'}) \u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item={'id': '922f562c3e4e19031c8fbcd3924f765bc113be0d1c47548542160560d2fb29e3', 'image': 'registry.osism.tech/osism/ceph-daemon:reef', 'name': '/ceph-osd-4', 'state': 'running', 'status': 'Up 27 minutes'})\u001b[0m\n\u001b[0;32mok: [testbed-node-3] => (item={'id': 'c03474f234450bddc2a0c9477ef1b7f0f4d0652f44427646b62c4249df7b2471', 'image': 'registry.osism.tech/osism/ceph-daemon:reef', 'name': '/ceph-osd-0', 'state': 'running', 'status': 'Up 27 minutes'})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'id': 'bf8e81e60e6eab0f3a4d0768532240dda904f417d047f3554ef601548d2b5a69', 'image': 'registry.osism.tech/kolla/ovn-controller:2024.2', 'name': '/ovn_controller', 'state': 'running', 'status': 'Up 31 minutes'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'id': '49db629377e40e37c32d4b4436dc13c468c1a978eca01e92607906d576581f3b', 'image': 'registry.osism.tech/kolla/openvswitch-vswitchd:2024.2', 'name': '/openvswitch_vswitchd', 'state': 'running', 'status': 'Up 32 minutes (healthy)'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'id': '996e6fbdfb74a64d1997bbb0740db739ff64078a51f72ec4a4ea33d7d8777b6c', 'image': 'registry.osism.tech/kolla/openvswitch-db-server:2024.2', 'name': '/openvswitch_db', 'state': 'running', 'status': 'Up 32 minutes (healthy)'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'id': '4801bd0b79b2061b5696fea9da6800affdedc6589db051c15f03f6d6b7503835', 'image': 'registry.osism.tech/kolla/cron:2024.2', 'name': '/cron', 'state': 'running', 'status': 'Up 33 minutes'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'id': '995e066d70d5885bcf887baeb9bf278c6f94e406c388c550b49f2e9f0ba5cb57', 'image': 'registry.osism.tech/kolla/kolla-toolbox:2024.2', 'name': '/kolla_toolbox', 'state': 'running', 'status': 'Up 33 minutes'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'id': '67f9710f4059105a2661569001496eb30077fafb374c8d5c28faaf4d5b9c3bc8', 'image': 'registry.osism.tech/kolla/fluentd:2024.2', 'name': '/fluentd', 'state': 'running', 'status': 'Up 34 minutes'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'id': '51e654a6545de29f8704424e2295db8cfe2186daf0112420d21978a218a4c137', 'image': 'registry.osism.tech/kolla/nova-compute:2024.2', 'name': '/nova_compute', 'state': 'running', 'status': 'Up 7 minutes (healthy)'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'id': 'c4c18371cdda7ff45dbf717e5d8105a59a8a8ba34b2476726087b395be68028f', 'image': 'registry.osism.tech/kolla/nova-libvirt:2024.2', 'name': '/nova_libvirt', 'state': 'running', 'status': 'Up 7 minutes (healthy)'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'id': '1ae7fc12531decf870cbc9bd51d380c679b4c44d3793b563313a86448871568e', 'image': 'registry.osism.tech/kolla/nova-ssh:2024.2', 'name': '/nova_ssh', 'state': 'running', 'status': 'Up 8 minutes (healthy)'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'id': '56a6fbeeda2fd1543c68e86dd827b240bd56da9da3fe640562a3582929753412', 'image': 'registry.osism.tech/kolla/prometheus-libvirt-exporter:2024.2', 'name': '/prometheus_libvirt_exporter', 'state': 'running', 'status': 'Up 13 minutes'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'id': '3d69e1cda5c13feb1b014dbdef99f1fcd9bbbccedfde2828586da0e597984609', 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2024.2', 'name': '/prometheus_cadvisor', 'state': 'running', 'status': 'Up 13 minutes'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'id': '000daa7f8fc0b537a11690f4d5b496a3209b7ca9097e55b7f512a96dc85df651', 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2024.2', 'name': '/prometheus_node_exporter', 'state': 'running', 'status': 'Up 14 minutes'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'id': 'd80759bb7fde2e88bb7ccff304e84416534f89ac07b37bfd4a6a6a86edd73a4a', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'name': '/neutron_ovn_metadata_agent', 'state': 'running', 'status': 'Up 15 minutes (healthy)'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'id': '28c367c08ce47d0c2e99bf9b57930ba34a58a39e7314a02f5966c7c168e6d688', 'image': 'registry.osism.tech/osism/ceph-daemon:reef', 'name': '/ceph-rgw-default-testbed-node-4-rgw0', 'state': 'running', 'status': 'Up 24 minutes'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'id': 'b7bd5696edbb4aab30ecda8e91d4fed8bb8b984bf0564e7f9ad3599fb86d3e11', 'image': 'registry.osism.tech/osism/ceph-daemon:reef', 'name': '/ceph-mds-testbed-node-4', 'state': 'running', 'status': 'Up 25 minutes'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'id': '1a44500c43c872bae8490a0656dccfd2f55b0e5cf94ab759e0063df75c1dbaa9', 'image': 'registry.osism.tech/osism/ceph-daemon:reef', 'name': '/ceph-crash-testbed-node-4', 'state': 'running', 'status': 'Up 26 minutes'}) \u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item={'id': '550dd349a35c2047ccf0a20cce01b054274e5147959a93ac2764c6642ef9d171', 'image': 'registry.osism.tech/osism/ceph-daemon:reef', 'name': '/ceph-osd-1', 'state': 'running', 'status': 'Up 27 minutes'})\u001b[0m\n\u001b[0;32mok: [testbed-node-4] => (item={'id': '914dde566f6c2bd3ee6bdd875e0979ef10f6f9583775e528ae1419f54cf48d77', 'image': 'registry.osism.tech/osism/ceph-daemon:reef', 'name': '/ceph-osd-5', 'state': 'running', 'status': 'Up 27 minutes'})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'id': '2257a99720c614cd19bf2235233dbac7df752935f30cfab15734ba0094276beb', 'image': 'registry.osism.tech/kolla/ovn-controller:2024.2', 'name': '/ovn_controller', 'state': 'running', 'status': 'Up 31 minutes'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'id': '1b73c76cbb63c8895690c0326076d5e803c9db53c64ee63c5394617d9fa58f78', 'image': 'registry.osism.tech/kolla/openvswitch-vswitchd:2024.2', 'name': '/openvswitch_vswitchd', 'state': 'running', 'status': 'Up 32 minutes (healthy)'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'id': '023e8a49ea84a98603face731bc8acae8c39ce5b679dccf7d971701da87f78d0', 'image': 'registry.osism.tech/kolla/openvswitch-db-server:2024.2', 'name': '/openvswitch_db', 'state': 'running', 'status': 'Up 32 minutes (healthy)'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'id': '3de1fb6f795eb0615d2c83b052d202be3f8549da3010f5851552db141df00893', 'image': 'registry.osism.tech/kolla/cron:2024.2', 'name': '/cron', 'state': 'running', 'status': 'Up 33 minutes'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'id': '528140ec3692082ea8f19d80356850ae605f1c70428f61b3f0d7e3cd3b9c4987', 'image': 'registry.osism.tech/kolla/kolla-toolbox:2024.2', 'name': '/kolla_toolbox', 'state': 'running', 'status': 'Up 33 minutes'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'id': 'e22fda3dfc45e78d5c2ead3e48f54abd1a03f359e55b1266341f7ea655e81457', 'image': 'registry.osism.tech/kolla/fluentd:2024.2', 'name': '/fluentd', 'state': 'running', 'status': 'Up 34 minutes'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'id': '9bcf861975e2d45d6534e072b676ef8f1326e2b533be31015f681afdd8de4ec8', 'image': 'registry.osism.tech/kolla/nova-compute:2024.2', 'name': '/nova_compute', 'state': 'running', 'status': 'Up 7 minutes (healthy)'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'id': '767dfe89db6361c0e18e57ebb9f64becb7d61836dd5c888c9b1d41f751ba69c8', 'image': 'registry.osism.tech/kolla/nova-libvirt:2024.2', 'name': '/nova_libvirt', 'state': 'running', 'status': 'Up 7 minutes (healthy)'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'id': '670d1815146d1f4cffdb4ff528e7b32e8177f2740870d7edb31a73035ead5b6c', 'image': 'registry.osism.tech/kolla/nova-ssh:2024.2', 'name': '/nova_ssh', 'state': 'running', 'status': 'Up 8 minutes (healthy)'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'id': '54d8627050388ea00f23210a5c17ba02818a23388e551196c8a36d0cce5e581c', 'image': 'registry.osism.tech/kolla/prometheus-libvirt-exporter:2024.2', 'name': '/prometheus_libvirt_exporter', 'state': 'running', 'status': 'Up 13 minutes'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'id': '492ed32d7a2d24a33b9974da752fb5738700b0dafdcbf9f84e5d3ac8621b44a3', 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2024.2', 'name': '/prometheus_cadvisor', 'state': 'running', 'status': 'Up 13 minutes'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'id': '71982626b1d88f167ce4c025cccdd5b3fa714c50224c77cc9ac8510773040646', 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2024.2', 'name': '/prometheus_node_exporter', 'state': 'running', 'status': 'Up 14 minutes'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'id': '69bd630a540ff812d7b1d0cdac1f044a77c92f56c71e3f38cb704883a3ff27b5', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'name': '/neutron_ovn_metadata_agent', 'state': 'running', 'status': 'Up 15 minutes (healthy)'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'id': '70034c20eb20d8d23f2a3a5361d234f36ac40a53065542d70db5dd74d0160900', 'image': 'registry.osism.tech/osism/ceph-daemon:reef', 'name': '/ceph-rgw-default-testbed-node-5-rgw0', 'state': 'running', 'status': 'Up 24 minutes'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'id': '81ca12344dd172735a5039a9a3401bb6ec32982f379782396844d65d3af4495d', 'image': 'registry.osism.tech/osism/ceph-daemon:reef', 'name': '/ceph-mds-testbed-node-5', 'state': 'running', 'status': 'Up 25 minutes'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'id': '83cad903f15d3dc2fc3fd10d328830b4be77538bafc32efd7a4af313d66b73a6', 'image': 'registry.osism.tech/osism/ceph-daemon:reef', 'name': '/ceph-crash-testbed-node-5', 'state': 'running', 'status': 'Up 26 minutes'}) \u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item={'id': 'df66acd5ab8adda5830fc22aabf350b2c0de0776b4d53792f54ba39d4a53c4d2', 'image': 'registry.osism.tech/osism/ceph-daemon:reef', 'name': '/ceph-osd-2', 'state': 'running', 'status': 'Up 27 minutes'})\u001b[0m\n\u001b[0;32mok: [testbed-node-5] => (item={'id': '1f23e7a3653884d283e6addfab83116f386c1e6a00e715e33c4bf8532d43e904', 'image': 'registry.osism.tech/osism/ceph-daemon:reef', 'name': '/ceph-osd-3', 'state': 'running', 'status': 'Up 27 minutes'})\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'id': 'fcaa772f75e20cec07ace286cfc1e6510a1f29a14f2b41fdf171038597901555', 'image': 'registry.osism.tech/kolla/ovn-controller:2024.2', 'name': '/ovn_controller', 'state': 'running', 'status': 'Up 31 minutes'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'id': 'ec81ef1b068903742aabc86658c38a8ddaec66bfe292ce40d4003dce62ca6137', 'image': 'registry.osism.tech/kolla/openvswitch-vswitchd:2024.2', 'name': '/openvswitch_vswitchd', 'state': 'running', 'status': 'Up 32 minutes (healthy)'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'id': '8bbdc8d1907348d07d4f3ce135a14d7288deef5386d718df5a4f0eccc2818f2f', 'image': 'registry.osism.tech/kolla/openvswitch-db-server:2024.2', 'name': '/openvswitch_db', 'state': 'running', 'status': 'Up 32 minutes (healthy)'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'id': '2f661e4839110de787c96265a8d88854e3c7cb0704897771bce4501fed471a65', 'image': 'registry.osism.tech/kolla/cron:2024.2', 'name': '/cron', 'state': 'running', 'status': 'Up 33 minutes'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'id': 'b9edb24895fb63536299193e977a3fb921cad196014b280edb3a908c2ce88bfc', 'image': 'registry.osism.tech/kolla/kolla-toolbox:2024.2', 'name': '/kolla_toolbox', 'state': 'running', 'status': 'Up 33 minutes'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'id': 'c3d85964fcbea496d2ab6b2cbe7e4b9729bc65ece7016929d5cd7f688c6c08de', 'image': 'registry.osism.tech/kolla/fluentd:2024.2', 'name': '/fluentd', 'state': 'running', 'status': 'Up 34 minutes'}) \u001b[0m\n\nTASK [Get count of ceph-osd containers on host] ********************************\nTuesday 24 February 2026  01:24:38 +0000 (0:00:00.522)       0:00:05.447 ****** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [Set test result to failed when count of containers is wrong] *************\nTuesday 24 February 2026  01:24:38 +0000 (0:00:00.315)       0:00:05.763 ****** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Set test result to passed if count matches] ******************************\nTuesday 24 February 2026  01:24:39 +0000 (0:00:00.534)       0:00:06.297 ****** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [Prepare test data] *******************************************************\nTuesday 24 February 2026  01:24:39 +0000 (0:00:00.337)       0:00:06.635 ****** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [Get list of ceph-osd containers that are not running] ********************\nTuesday 24 February 2026  01:24:40 +0000 (0:00:00.332)       0:00:06.967 ****** \n\u001b[0;36mskipping: [testbed-node-3] => (item={'name': 'ceph-osd-4', 'osd_id': '4', 'state': 'running'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3] => (item={'name': 'ceph-osd-0', 'osd_id': '0', 'state': 'running'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'name': 'ceph-osd-1', 'osd_id': '1', 'state': 'running'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4] => (item={'name': 'ceph-osd-5', 'osd_id': '5', 'state': 'running'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'name': 'ceph-osd-2', 'osd_id': '2', 'state': 'running'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5] => (item={'name': 'ceph-osd-3', 'osd_id': '3', 'state': 'running'}) \u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Get count of ceph-osd containers that are not running] *******************\nTuesday 24 February 2026  01:24:40 +0000 (0:00:00.369)       0:00:07.337 ****** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [Set test result to failed if an OSD is not running] **********************\nTuesday 24 February 2026  01:24:41 +0000 (0:00:00.577)       0:00:07.915 ****** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Set test result to failed if an OSD is not running] **********************\nTuesday 24 February 2026  01:24:41 +0000 (0:00:00.327)       0:00:08.242 ****** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Set test result to passed if all containers are running] *****************\nTuesday 24 February 2026  01:24:41 +0000 (0:00:00.315)       0:00:08.558 ****** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [Aggregate test results step one] *****************************************\nTuesday 24 February 2026  01:24:42 +0000 (0:00:00.421)       0:00:08.980 ****** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Aggregate test results step two] *****************************************\nTuesday 24 February 2026  01:24:42 +0000 (0:00:00.777)       0:00:09.757 ****** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Aggregate test results step three] ***************************************\nTuesday 24 February 2026  01:24:43 +0000 (0:00:00.289)       0:00:10.047 ****** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Flush handlers] **********************************************************\nTuesday 24 February 2026  01:24:43 +0000 (0:00:00.321)       0:00:10.368 ****** \n\nTASK [Flush handlers] **********************************************************\nTuesday 24 February 2026  01:24:43 +0000 (0:00:00.094)       0:00:10.462 ****** \n\nTASK [Flush handlers] **********************************************************\nTuesday 24 February 2026  01:24:43 +0000 (0:00:00.072)       0:00:10.535 ****** \n\nTASK [Print report file information] *******************************************\nTuesday 24 February 2026  01:24:43 +0000 (0:00:00.074)       0:00:10.609 ****** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Fail early due to containers not running] ********************************\nTuesday 24 February 2026  01:24:44 +0000 (0:00:00.308)       0:00:10.918 ****** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Prepare test data] *******************************************************\nTuesday 24 February 2026  01:24:44 +0000 (0:00:00.273)       0:00:11.191 ****** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [Set _mon_hostname fact] **************************************************\nTuesday 24 February 2026  01:24:44 +0000 (0:00:00.317)       0:00:11.508 ****** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\nTASK [Get ceph osd tree] *******************************************************\nTuesday 24 February 2026  01:24:45 +0000 (0:00:00.978)       0:00:12.487 ****** \n\u001b[0;33mchanged: [testbed-node-3 -> testbed-node-0(192.168.16.10)]\u001b[0m\n\nTASK [Parse osd tree from JSON] ************************************************\nTuesday 24 February 2026  01:24:47 +0000 (0:00:01.850)       0:00:14.337 ****** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\nTASK [Get OSDs that are not up or in] ******************************************\nTuesday 24 February 2026  01:24:47 +0000 (0:00:00.152)       0:00:14.490 ****** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\nTASK [Fail test if OSDs are not up or in] **************************************\nTuesday 24 February 2026  01:24:48 +0000 (0:00:00.373)       0:00:14.863 ****** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Pass test if OSDs are all up and in] *************************************\nTuesday 24 February 2026  01:24:48 +0000 (0:00:00.139)       0:00:15.003 ****** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\nTASK [Prepare test data] *******************************************************\nTuesday 24 February 2026  01:24:48 +0000 (0:00:00.167)       0:00:15.171 ****** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [List ceph LVM volumes and collect data] **********************************\nTuesday 24 February 2026  01:24:48 +0000 (0:00:00.336)       0:00:15.507 ****** \n\u001b[0;33mchanged: [testbed-node-3]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-4]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-5]\u001b[0m\n\nTASK [Parse LVM data as JSON] **************************************************\nTuesday 24 February 2026  01:24:51 +0000 (0:00:02.906)       0:00:18.414 ****** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [Get unencrypted and encrypted OSDs] **************************************\nTuesday 24 February 2026  01:24:51 +0000 (0:00:00.356)       0:00:18.770 ****** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [Fail if count of encrypted OSDs does not match] **************************\nTuesday 24 February 2026  01:24:52 +0000 (0:00:00.569)       0:00:19.339 ****** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Pass if count of encrypted OSDs equals count of OSDs] ********************\nTuesday 24 February 2026  01:24:52 +0000 (0:00:00.357)       0:00:19.697 ****** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [Fail if count of unencrypted OSDs does not match] ************************\nTuesday 24 February 2026  01:24:53 +0000 (0:00:00.665)       0:00:20.363 ****** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Pass if count of unencrypted OSDs equals count of OSDs] ******************\nTuesday 24 February 2026  01:24:53 +0000 (0:00:00.327)       0:00:20.690 ****** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Prepare test data] *******************************************************\nTuesday 24 February 2026  01:24:54 +0000 (0:00:00.368)       0:00:21.059 ****** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [Get CRUSH node data of each OSD host and root node childs] ***************\nTuesday 24 February 2026  01:24:54 +0000 (0:00:00.591)       0:00:21.650 ****** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [Calculate sub test expression results] ***********************************\nTuesday 24 February 2026  01:24:55 +0000 (0:00:00.963)       0:00:22.614 ****** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [Fail test if any sub test failed] ****************************************\nTuesday 24 February 2026  01:24:56 +0000 (0:00:00.361)       0:00:22.976 ****** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-4]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-5]\u001b[0m\n\nTASK [Pass test if no sub test failed] *****************************************\nTuesday 24 February 2026  01:24:56 +0000 (0:00:00.369)       0:00:23.345 ****** \n\u001b[0;32mok: [testbed-node-3]\u001b[0m\n\u001b[0;32mok: [testbed-node-4]\u001b[0m\n\u001b[0;32mok: [testbed-node-5]\u001b[0m\n\nTASK [Set validation result to passed if no test failed] ***********************\nTuesday 24 February 2026  01:24:57 +0000 (0:00:00.577)       0:00:23.923 ****** \n\u001b[0;32mok: [testbed-node-3 -> testbed-manager(192.168.16.5)]\u001b[0m\n\nTASK [Set validation result to failed if a test failed] ************************\nTuesday 24 February 2026  01:24:57 +0000 (0:00:00.330)       0:00:24.253 ****** \n\u001b[0;36mskipping: [testbed-node-3]\u001b[0m\n\nTASK [Aggregate test results step one] *****************************************\nTuesday 24 February 2026  01:24:57 +0000 (0:00:00.293)       0:00:24.547 ****** \n\u001b[0;32mok: [testbed-node-3 -> testbed-manager(192.168.16.5)]\u001b[0m\n\nTASK [Aggregate test results step two] *****************************************\nTuesday 24 February 2026  01:24:59 +0000 (0:00:01.883)       0:00:26.430 ****** \n\u001b[0;32mok: [testbed-node-3 -> testbed-manager(192.168.16.5)]\u001b[0m\n\nTASK [Aggregate test results step three] ***************************************\nTuesday 24 February 2026  01:24:59 +0000 (0:00:00.325)       0:00:26.756 ****** \n\u001b[0;33mchanged: [testbed-node-3 -> testbed-manager(192.168.16.5)]\u001b[0m\n\nTASK [Flush handlers] **********************************************************\nTuesday 24 February 2026  01:25:00 +0000 (0:00:00.308)       0:00:27.065 ****** \n\nTASK [Flush handlers] **********************************************************\nTuesday 24 February 2026  01:25:00 +0000 (0:00:00.089)       0:00:27.154 ****** \n\nTASK [Flush handlers] **********************************************************\nTuesday 24 February 2026  01:25:00 +0000 (0:00:00.090)       0:00:27.244 ****** \n\nRUNNING HANDLER [Write report file] ********************************************\nTuesday 24 February 2026  01:25:00 +0000 (0:00:00.098)       0:00:27.342 ****** \n\u001b[0;33mchanged: [testbed-node-3 -> testbed-manager(192.168.16.5)]\u001b[0m\n\nTASK [Print report file information] *******************************************\nTuesday 24 February 2026  01:25:02 +0000 (0:00:02.074)       0:00:29.417 ****** \n\u001b[0;32mok: [testbed-node-3 -> testbed-manager(192.168.16.5)] => {\u001b[0m\n\u001b[0;32m    \"msg\": [\u001b[0m\n\u001b[0;32m        \"Validator run completed.\",\u001b[0m\n\u001b[0;32m        \"You can find the report file here:\",\u001b[0m\n\u001b[0;32m        \"/opt/reports/validator/ceph-osds-validator-2026-02-24T01:24:34+00:00-report.json\",\u001b[0m\n\u001b[0;32m        \"on the following host:\",\u001b[0m\n\u001b[0;32m        \"testbed-manager\"\u001b[0m\n\u001b[0;32m    ]\u001b[0m\n\u001b[0;32m}\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;33mtestbed-node-3\u001b[0m             : \u001b[0;32mok=35  \u001b[0m \u001b[0;33mchanged=4   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=17  \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-4\u001b[0m             : \u001b[0;32mok=18  \u001b[0m \u001b[0;33mchanged=1   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=9   \u001b[0m rescued=0    ignored=0   \n\u001b[0;33mtestbed-node-5\u001b[0m             : \u001b[0;32mok=18  \u001b[0m \u001b[0;33mchanged=1   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=9   \u001b[0m rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nTuesday 24 February 2026  01:25:03 +0000 (0:00:00.745)       0:00:30.162 ****** \n=============================================================================== \nList ceph LVM volumes and collect data ---------------------------------- 2.91s\nWrite report file ------------------------------------------------------- 2.07s\nAggregate test results step one ----------------------------------------- 1.88s\nGet ceph osd tree ------------------------------------------------------- 1.85s\nCalculate total number of OSDs in cluster ------------------------------- 0.98s\nSet _mon_hostname fact -------------------------------------------------- 0.98s\nGet CRUSH node data of each OSD host and root node childs --------------- 0.96s\nAggregate test results step one ----------------------------------------- 0.78s\nGet timestamp for report file ------------------------------------------- 0.77s\nPrint report file information ------------------------------------------- 0.75s\nCreate report output directory ------------------------------------------ 0.73s\nPass if count of encrypted OSDs equals count of OSDs -------------------- 0.67s\nPrepare test data ------------------------------------------------------- 0.59s\nGet count of ceph-osd containers that are not running ------------------- 0.58s\nPass test if no sub test failed ----------------------------------------- 0.58s\nGet unencrypted and encrypted OSDs -------------------------------------- 0.57s\nSet test result to failed when count of containers is wrong ------------- 0.53s\nGet list of ceph-osd containers on host --------------------------------- 0.52s\nGet extra vars for Ceph configuration ----------------------------------- 0.46s\nSet test result to passed if all containers are running ----------------- 0.42s\n+ sh -c /opt/configuration/scripts/check/200-infrastructure.sh\n+ set -e\n+ source /opt/manager-vars.sh\n++ export NUMBER_OF_NODES=6\n++ NUMBER_OF_NODES=6\n++ export CEPH_VERSION=reef\n++ CEPH_VERSION=reef\n++ export CONFIGURATION_VERSION=main\n++ CONFIGURATION_VERSION=main\n++ export MANAGER_VERSION=latest\n++ MANAGER_VERSION=latest\n++ export OPENSTACK_VERSION=2024.2\n++ OPENSTACK_VERSION=2024.2\n++ export ARA=false\n++ ARA=false\n++ export DEPLOY_MODE=manager\n++ DEPLOY_MODE=manager\n++ export TEMPEST=true\n++ TEMPEST=true\n++ export IS_ZUUL=true\n++ IS_ZUUL=true\n++ export MANAGER_PUBLIC_IP_ADDRESS=81.163.192.195\n++ MANAGER_PUBLIC_IP_ADDRESS=81.163.192.195\n++ export EXTERNAL_API=false\n++ EXTERNAL_API=false\n++ export IMAGE_USER=ubuntu\n++ IMAGE_USER=ubuntu\n++ export IMAGE_NODE_USER=ubuntu\n++ IMAGE_NODE_USER=ubuntu\n++ export CEPH_STACK=ceph-ansible\n++ CEPH_STACK=ceph-ansible\n+ [[ -e /etc/redhat-release ]]\n+ source /etc/os-release\n++ PRETTY_NAME='Ubuntu 24.04.4 LTS'\n++ NAME=Ubuntu\n++ VERSION_ID=24.04\n++ VERSION='24.04.4 LTS (Noble Numbat)'\n++ VERSION_CODENAME=noble\n++ ID=ubuntu\n++ ID_LIKE=debian\n++ HOME_URL=https://www.ubuntu.com/\n++ SUPPORT_URL=https://help.ubuntu.com/\n++ BUG_REPORT_URL=https://bugs.launchpad.net/ubuntu/\n++ PRIVACY_POLICY_URL=https://www.ubuntu.com/legal/terms-and-policies/privacy-policy\n++ UBUNTU_CODENAME=noble\n++ LOGO=ubuntu-logo\n+ [[ ubuntu == \\u\\b\\u\\n\\t\\u ]]\n+ packages='libmonitoring-plugin-perl libwww-perl libjson-perl monitoring-plugins-basic mysql-client'\n+ dpkg -s libmonitoring-plugin-perl libwww-perl libjson-perl monitoring-plugins-basic mysql-client\n+ sudo apt-get install -y libmonitoring-plugin-perl libwww-perl libjson-perl monitoring-plugins-basic mysql-client\n\n# Status of Elasticsearch\n\n+ pushd /opt/configuration/contrib\n+ echo\n+ echo '# Status of Elasticsearch'\n+ echo\n+ bash nagios-plugins/check_elasticsearch -H api-int.testbed.osism.xyz -s\nOK - elasticsearch (kolla_logging) is running. status: green; timed_out: false; number_of_nodes: 3; number_of_data_nodes: 3; active_primary_shards: 9; active_shards: 22; relocating_shards: 0; initializing_shards: 0; delayed_unassigned_shards: 0; unassigned_shards: 0  | 'active_primary'=9 'active'=22 'relocating'=0 'init'=0 'delay_unass'=0 'unass'=0\n\n# Status of MariaDB\n\n+ echo\n+ echo '# Status of MariaDB'\n+ echo\n++ semver latest 10.0.0-0\n+ [[ -1 -ge 0 ]]\n+ [[ latest == \\l\\a\\t\\e\\s\\t ]]\n+ osism status database\n\u001b[32m2026-02-24 01:25:32\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mUnable to get ansible vault password\u001b[0m\n\u001b[32m2026-02-24 01:25:32\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mUnable to get vault secret: [Errno 2] No such file or directory: '/share/ansible_vault_password.key'\u001b[0m\n\u001b[32m2026-02-24 01:25:32\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mDropping encrypted entries\u001b[0m\n\u001b[32m2026-02-24 01:25:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mConnecting to MariaDB at 192.168.16.9 as root_shard_0...\u001b[0m\n\u001b[32m2026-02-24 01:25:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mCluster Status: Primary\u001b[0m\n\u001b[32m2026-02-24 01:25:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mConnected: ON\u001b[0m\n\u001b[32m2026-02-24 01:25:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mReady: ON\u001b[0m\n\u001b[32m2026-02-24 01:25:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mCluster Size: 3\u001b[0m\n\u001b[32m2026-02-24 01:25:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLocal State: Synced\u001b[0m\n\u001b[32m2026-02-24 01:25:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mCluster State UUID: ce650e1c-111b-11f1-87f2-a37335181952\u001b[0m\n\u001b[32m2026-02-24 01:25:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mCluster Members: 192.168.16.10:3306,192.168.16.11:3306,192.168.16.12:3306\u001b[0m\n\u001b[32m2026-02-24 01:25:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mGalera Version: 26.4.25(r7387a566)\u001b[0m\n\u001b[32m2026-02-24 01:25:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLocal Node UUID: 05d530b6-111c-11f1-88be-d640b4f94096\u001b[0m\n\u001b[32m2026-02-24 01:25:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mFlow Control Paused: 0.00%\u001b[0m\n\u001b[32m2026-02-24 01:25:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mRecv Queue Avg: 0\u001b[0m\n\u001b[32m2026-02-24 01:25:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSend Queue Avg: 0.000144238\u001b[0m\n\u001b[32m2026-02-24 01:25:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTransactions: 4696 local commits, 6874 replicated, 120 received\u001b[0m\n\u001b[32m2026-02-24 01:25:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mConflicts: 0 cert failures, 0 bf aborts\u001b[0m\n\u001b[32m2026-02-24 01:25:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mMariaDB Uptime: 26 minutes, 25 seconds\u001b[0m\n\u001b[32m2026-02-24 01:25:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mThreads: 125 connected, 1 running\u001b[0m\n\u001b[32m2026-02-24 01:25:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mQueries: 147247 total, 0 slow\u001b[0m\n\u001b[32m2026-02-24 01:25:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mAborted Connects: 148\u001b[0m\n\u001b[32m2026-02-24 01:25:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mMariaDB Galera Cluster validation PASSED\u001b[0m\n\n# Status of Prometheus\n\n+ echo\n+ echo '# Status of Prometheus'\n+ echo\n+ curl -s https://api-int.testbed.osism.xyz:9091/-/healthy\nUnauthorized\n+ curl -s https://api-int.testbed.osism.xyz:9091/-/ready\nUnauthorized\n\n# Status of RabbitMQ\n\n+ echo\n+ echo '# Status of RabbitMQ'\n+ echo\n++ semver latest 10.0.0-0\n+ [[ -1 -ge 0 ]]\n+ [[ latest == \\l\\a\\t\\e\\s\\t ]]\n+ osism status messaging\n\u001b[32m2026-02-24 01:25:57\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mUnable to get ansible vault password\u001b[0m\n\u001b[32m2026-02-24 01:25:57\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mUnable to get vault secret: [Errno 2] No such file or directory: '/share/ansible_vault_password.key'\u001b[0m\n\u001b[32m2026-02-24 01:25:57\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mDropping encrypted entries\u001b[0m\n\u001b[32m2026-02-24 01:25:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-0] Connecting to RabbitMQ Management API at 192.168.16.10:15672 as openstack...\u001b[0m\n\u001b[32m2026-02-24 01:25:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-0] RabbitMQ Version: 3.13.7\u001b[0m\n\u001b[32m2026-02-24 01:25:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-0] Erlang Version: 26.2.5.15\u001b[0m\n\u001b[32m2026-02-24 01:25:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-0] Cluster Name: rabbit@testbed-node-0\u001b[0m\n\u001b[32m2026-02-24 01:25:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-0] Cluster Size: 3\u001b[0m\n\u001b[32m2026-02-24 01:25:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-0] Nodes: rabbit@testbed-node-0, rabbit@testbed-node-1, rabbit@testbed-node-2\u001b[0m\n\u001b[32m2026-02-24 01:25:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-0] Running Nodes: rabbit@testbed-node-0, rabbit@testbed-node-1, rabbit@testbed-node-2\u001b[0m\n\u001b[32m2026-02-24 01:25:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-0] Partitions: None (healthy)\u001b[0m\n\u001b[32m2026-02-24 01:25:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-0] Connections: 194, Channels: 193, Queues: 173\u001b[0m\n\u001b[32m2026-02-24 01:25:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-0] Messages: 218 total, 218 ready, 0 unacked\u001b[0m\n\u001b[32m2026-02-24 01:25:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-0] Message Rates: 6.6/s publish, 6.0/s deliver\u001b[0m\n\u001b[32m2026-02-24 01:25:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-0] Disk Free: 58.2 GB (limit: 0.0 GB)\u001b[0m\n\u001b[32m2026-02-24 01:25:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-0] Memory Used: 0.18 GB (limit: 12.54 GB)\u001b[0m\n\u001b[32m2026-02-24 01:25:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-0] File Descriptors: 113/1024\u001b[0m\n\u001b[32m2026-02-24 01:25:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-0] Sockets: 67/832\u001b[0m\n\u001b[32m2026-02-24 01:25:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-1] Connecting to RabbitMQ Management API at 192.168.16.11:15672 as openstack...\u001b[0m\n\u001b[32m2026-02-24 01:25:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-1] RabbitMQ Version: 3.13.7\u001b[0m\n\u001b[32m2026-02-24 01:25:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-1] Erlang Version: 26.2.5.15\u001b[0m\n\u001b[32m2026-02-24 01:25:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-1] Cluster Name: rabbit@testbed-node-1\u001b[0m\n\u001b[32m2026-02-24 01:25:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-1] Cluster Size: 3\u001b[0m\n\u001b[32m2026-02-24 01:25:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-1] Nodes: rabbit@testbed-node-0, rabbit@testbed-node-1, rabbit@testbed-node-2\u001b[0m\n\u001b[32m2026-02-24 01:25:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-1] Running Nodes: rabbit@testbed-node-0, rabbit@testbed-node-1, rabbit@testbed-node-2\u001b[0m\n\u001b[32m2026-02-24 01:25:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-1] Partitions: None (healthy)\u001b[0m\n\u001b[32m2026-02-24 01:25:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-1] Connections: 194, Channels: 193, Queues: 173\u001b[0m\n\u001b[32m2026-02-24 01:25:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-1] Messages: 218 total, 218 ready, 0 unacked\u001b[0m\n\u001b[32m2026-02-24 01:25:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-1] Message Rates: 6.6/s publish, 6.0/s deliver\u001b[0m\n\u001b[32m2026-02-24 01:25:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-1] Disk Free: 58.5 GB (limit: 0.0 GB)\u001b[0m\n\u001b[32m2026-02-24 01:25:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-1] Memory Used: 0.18 GB (limit: 12.54 GB)\u001b[0m\n\u001b[32m2026-02-24 01:25:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-1] File Descriptors: 119/1024\u001b[0m\n\u001b[32m2026-02-24 01:25:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-1] Sockets: 71/832\u001b[0m\n\u001b[32m2026-02-24 01:25:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-2] Connecting to RabbitMQ Management API at 192.168.16.12:15672 as openstack...\u001b[0m\n\u001b[32m2026-02-24 01:25:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-2] RabbitMQ Version: 3.13.7\u001b[0m\n\u001b[32m2026-02-24 01:25:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-2] Erlang Version: 26.2.5.15\u001b[0m\n\u001b[32m2026-02-24 01:25:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-2] Cluster Name: rabbit@testbed-node-2\u001b[0m\n\u001b[32m2026-02-24 01:25:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-2] Cluster Size: 3\u001b[0m\n\u001b[32m2026-02-24 01:25:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-2] Nodes: rabbit@testbed-node-0, rabbit@testbed-node-1, rabbit@testbed-node-2\u001b[0m\n\u001b[32m2026-02-24 01:25:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-2] Running Nodes: rabbit@testbed-node-0, rabbit@testbed-node-1, rabbit@testbed-node-2\u001b[0m\n\u001b[32m2026-02-24 01:25:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-2] Partitions: None (healthy)\u001b[0m\n\u001b[32m2026-02-24 01:25:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-2] Connections: 194, Channels: 193, Queues: 173\u001b[0m\n\u001b[32m2026-02-24 01:25:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-2] Messages: 218 total, 218 ready, 0 unacked\u001b[0m\n\u001b[32m2026-02-24 01:25:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-2] Message Rates: 6.6/s publish, 6.0/s deliver\u001b[0m\n\u001b[32m2026-02-24 01:25:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-2] Disk Free: 58.5 GB (limit: 0.0 GB)\u001b[0m\n\u001b[32m2026-02-24 01:25:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-2] Memory Used: 0.17 GB (limit: 12.54 GB)\u001b[0m\n\u001b[32m2026-02-24 01:25:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-2] File Descriptors: 102/1024\u001b[0m\n\u001b[32m2026-02-24 01:25:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-2] Sockets: 56/832\u001b[0m\n\u001b[32m2026-02-24 01:25:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mRabbitMQ Cluster validation PASSED\u001b[0m\n\n# Status of Redis\n\n+ echo\n+ echo '# Status of Redis'\n+ echo\n+ /usr/lib/nagios/plugins/check_tcp -H 192.168.16.10 -p 6379 -A -E -s 'AUTH QHNA1SZRlOKzLADhUd5ZDgpHfQe6dNfr3bwEdY24\\r\\nPING\\r\\nINFO replication\\r\\nQUIT\\r\\n' -e PONG -e role:master -e slave0:ip=192.168.16.1 -e,port=6379 -j\nTCP OK - 0.002 second response time on 192.168.16.10 port 6379|time=0.002036s;;;0.000000;10.000000\n\n# Create backup of MariaDB database\n\n+ popd\n+ echo\n+ echo '# Create backup of MariaDB database'\n+ echo\n+ osism apply mariadb_backup -e mariadb_backup_type=full\n\u001b[32m2026-02-24 01:26:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mPrepare task for execution of mariadb_backup.\u001b[0m\n\u001b[32m2026-02-24 01:26:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 00dcac0d-65ca-4e85-892e-f92ee847cc55 (mariadb_backup) was prepared for execution.\u001b[0m\n\u001b[32m2026-02-24 01:26:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task 00dcac0d-65ca-4e85-892e-f92ee847cc55 (mariadb_backup) has been started and output is visible here.\u001b[0m\n\nPLAY [Group hosts based on configuration] **************************************\n\nTASK [Group hosts based on Kolla action] ***************************************\nTuesday 24 February 2026  01:26:05 +0000 (0:00:00.210)       0:00:00.210 ****** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [Group hosts based on enabled services] ***********************************\nTuesday 24 February 2026  01:26:05 +0000 (0:00:00.362)       0:00:00.572 ****** \n\u001b[0;32mok: [testbed-node-0] => (item=enable_mariadb_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-1] => (item=enable_mariadb_True)\u001b[0m\n\u001b[0;32mok: [testbed-node-2] => (item=enable_mariadb_True)\u001b[0m\n\nPLAY [Apply role mariadb] ******************************************************\n\nTASK [mariadb : Group MariaDB hosts based on shards] ***************************\nTuesday 24 February 2026  01:26:06 +0000 (0:00:00.627)       0:00:01.200 ****** \n\u001b[0;32mok: [testbed-node-0] => (item=testbed-node-0)\u001b[0m\n\u001b[0;32mok: [testbed-node-0] => (item=testbed-node-1)\u001b[0m\n\u001b[0;32mok: [testbed-node-0] => (item=testbed-node-2)\u001b[0m\n\nTASK [mariadb : include_tasks] *************************************************\nTuesday 24 February 2026  01:26:06 +0000 (0:00:00.443)       0:00:01.643 ****** \n\u001b[0;36mincluded: /ansible/roles/mariadb/tasks/backup.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m\n\nTASK [mariadb : Get MariaDB container facts] ***********************************\nTuesday 24 February 2026  01:26:07 +0000 (0:00:00.646)       0:00:02.290 ****** \n\u001b[0;32mok: [testbed-node-0]\u001b[0m\n\u001b[0;32mok: [testbed-node-1]\u001b[0m\n\u001b[0;32mok: [testbed-node-2]\u001b[0m\n\nTASK [mariadb : Taking full database backup via Mariabackup] *******************\nTuesday 24 February 2026  01:26:11 +0000 (0:00:03.663)       0:00:05.953 ****** \n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\u001b[0;33mchanged: [testbed-node-0]\u001b[0m\n\u001b[1;35m[WARNING]: Could not match supplied host pattern, ignoring: mariadb_restart\u001b[0m\n\nPLAY [Restart mariadb services] ************************************************\n\u001b[0;36mskipping: no hosts matched\u001b[0m\n\u001b[1;35m[WARNING]: Could not match supplied host pattern, ignoring: mariadb_start\u001b[0m\n\nPLAY [Start mariadb services] **************************************************\n\u001b[0;36mskipping: no hosts matched\u001b[0m\n\u001b[1;35m[WARNING]: Could not match supplied host pattern, ignoring:\u001b[0m\n\u001b[1;35mmariadb_bootstrap_restart\u001b[0m\n\nPLAY [Restart bootstrap mariadb service] ***************************************\n\u001b[0;36mskipping: no hosts matched\u001b[0m\n\nPLAY [Apply mariadb post-configuration] ****************************************\n\nTASK [Include mariadb post-deploy.yml] *****************************************\nTuesday 24 February 2026  01:26:28 +0000 (0:00:17.742)       0:00:23.696 ****** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nTASK [Include mariadb post-upgrade.yml] ****************************************\nTuesday 24 February 2026  01:26:29 +0000 (0:00:00.356)       0:00:24.052 ****** \n\u001b[0;36mskipping: [testbed-node-0]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-1]\u001b[0m\n\u001b[0;36mskipping: [testbed-node-2]\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=6   \u001b[0m \u001b[0;33mchanged=1   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   \n\u001b[0;32mtestbed-node-1\u001b[0m             : \u001b[0;32mok=4   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=3   \u001b[0m rescued=0    ignored=0   \n\u001b[0;32mtestbed-node-2\u001b[0m             : \u001b[0;32mok=4   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=3   \u001b[0m rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nTuesday 24 February 2026  01:26:29 +0000 (0:00:00.471)       0:00:24.524 ****** \n=============================================================================== \nmariadb : Taking full database backup via Mariabackup ------------------ 17.74s\nmariadb : Get MariaDB container facts ----------------------------------- 3.66s\nmariadb : include_tasks ------------------------------------------------- 0.65s\nGroup hosts based on enabled services ----------------------------------- 0.63s\nInclude mariadb post-upgrade.yml ---------------------------------------- 0.47s\nmariadb : Group MariaDB hosts based on shards --------------------------- 0.44s\nGroup hosts based on Kolla action --------------------------------------- 0.36s\nInclude mariadb post-deploy.yml ----------------------------------------- 0.36s\n+ sh -c /opt/configuration/scripts/check/300-openstack.sh\n+ set -e\n+ source /opt/configuration/scripts/include.sh\n++ export INTERACTIVE=false\n++ INTERACTIVE=false\n++ export OSISM_APPLY_RETRY=1\n++ OSISM_APPLY_RETRY=1\n+ source /opt/configuration/scripts/manager-version.sh\n+++ awk '-F: ' '/^manager_version:/ { print $2 }' /opt/configuration/environments/manager/configuration.yml\n\n# OpenStack endpoints\n\n++ export MANAGER_VERSION=latest\n++ MANAGER_VERSION=latest\n+ export OS_CLOUD=admin\n+ OS_CLOUD=admin\n+ echo\n+ echo '# OpenStack endpoints'\n+ echo\n+ openstack endpoint list\n+----------------------------------+-----------+--------------+-----------------+---------+-----------+---------------------------------------------------------------------+\n| ID                               | Region    | Service Name | Service Type    | Enabled | Interface | URL                                                                 |\n+----------------------------------+-----------+--------------+-----------------+---------+-----------+---------------------------------------------------------------------+\n| 0f79775df2984381b69a39a3c8f528a6 | RegionOne | neutron      | network         | True    | public    | https://api.testbed.osism.xyz:9696                                  |\n| 11d66025a00741ed894131de48a5dae2 | RegionOne | glance       | image           | True    | public    | https://api.testbed.osism.xyz:9292                                  |\n| 1339ce61b4c044e6b62d23025297cb02 | RegionOne | designate    | dns             | True    | public    | https://api.testbed.osism.xyz:9001                                  |\n| 39efb8ff32f1445a98f7befaa27d645d | RegionOne | octavia      | load-balancer   | True    | public    | https://api.testbed.osism.xyz:9876                                  |\n| 3f7ee67821ba444da6a14c0aa2e4a2db | RegionOne | swift        | object-store    | True    | public    | https://api.testbed.osism.xyz:6780/swift/v1/AUTH_%(project_id)s     |\n| 4c07fb06d154427a9d22d05d70e3c3b3 | RegionOne | cinderv3     | volumev3        | True    | public    | https://api.testbed.osism.xyz:8776/v3/%(tenant_id)s                 |\n| 4f6115d6d2694b208e9151fee09a1055 | RegionOne | octavia      | load-balancer   | True    | internal  | https://api-int.testbed.osism.xyz:9876                              |\n| 5381b2fd05f24977b35e574551cc90e0 | RegionOne | placement    | placement       | True    | internal  | https://api-int.testbed.osism.xyz:8780                              |\n| 7886a9dedb5847ef884e8ab23b142b95 | RegionOne | keystone     | identity        | True    | public    | https://api.testbed.osism.xyz:5000                                  |\n| 820b19b0ce25401eb1312ddf139fc769 | RegionOne | swift        | object-store    | True    | internal  | https://api-int.testbed.osism.xyz:6780/swift/v1/AUTH_%(project_id)s |\n| 8a2ce0d906a14e28af196bca035f3420 | RegionOne | magnum       | container-infra | True    | public    | https://api.testbed.osism.xyz:9511/v1                               |\n| 9747730a38b14f92b3875abe2bd50c77 | RegionOne | magnum       | container-infra | True    | internal  | https://api-int.testbed.osism.xyz:9511/v1                           |\n| 98d65750f75b4f09ac1bfd8bf24814d2 | RegionOne | barbican     | key-manager     | True    | internal  | https://api-int.testbed.osism.xyz:9311                              |\n| 99f9e533d552453390d9520d0d1b791b | RegionOne | barbican     | key-manager     | True    | public    | https://api.testbed.osism.xyz:9311                                  |\n| a1b5fd82ee5145a98ca38241b04e6ab9 | RegionOne | placement    | placement       | True    | public    | https://api.testbed.osism.xyz:8780                                  |\n| ae914f5218ba421aa35104854b40493b | RegionOne | keystone     | identity        | True    | internal  | https://api-int.testbed.osism.xyz:5000                              |\n| bed9f6e75968413a94b053e6321fde45 | RegionOne | neutron      | network         | True    | internal  | https://api-int.testbed.osism.xyz:9696                              |\n| c17ded326a994063b389de5737dceafd | RegionOne | nova         | compute         | True    | public    | https://api.testbed.osism.xyz:8774/v2.1                             |\n| ddd8a701f1854e3fa079a918613cdb47 | RegionOne | designate    | dns             | True    | internal  | https://api-int.testbed.osism.xyz:9001                              |\n| e64331c9909c451581a31b1431037a8d | RegionOne | nova         | compute         | True    | internal  | https://api-int.testbed.osism.xyz:8774/v2.1                         |\n| ef0d0bd767bc4f0fb38578ef589d6c7e | RegionOne | cinderv3     | volumev3        | True    | internal  | https://api-int.testbed.osism.xyz:8776/v3/%(tenant_id)s             |\n| fc6fbb20a8434150a5b23f28ce920d9b | RegionOne | glance       | image           | True    | internal  | https://api-int.testbed.osism.xyz:9292                              |\n+----------------------------------+-----------+--------------+-----------------+---------+-----------+---------------------------------------------------------------------+\n\n# Cinder\n\n+ echo\n+ echo '# Cinder'\n+ echo\n+ openstack volume service list\n+------------------+----------------------------+----------+---------+-------+----------------------------+\n| Binary           | Host                       | Zone     | Status  | State | Updated At                 |\n+------------------+----------------------------+----------+---------+-------+----------------------------+\n| cinder-scheduler | testbed-node-0             | internal | enabled | up    | 2026-02-24T01:26:34.000000 |\n| cinder-scheduler | testbed-node-1             | internal | enabled | up    | 2026-02-24T01:26:33.000000 |\n| cinder-scheduler | testbed-node-2             | internal | enabled | up    | 2026-02-24T01:26:33.000000 |\n| cinder-volume    | testbed-node-0@rbd-volumes | nova     | enabled | up    | 2026-02-24T01:26:33.000000 |\n| cinder-volume    | testbed-node-2@rbd-volumes | nova     | enabled | up    | 2026-02-24T01:26:28.000000 |\n| cinder-volume    | testbed-node-1@rbd-volumes | nova     | enabled | up    | 2026-02-24T01:26:28.000000 |\n| cinder-backup    | testbed-node-0             | nova     | enabled | up    | 2026-02-24T01:26:30.000000 |\n| cinder-backup    | testbed-node-1             | nova     | enabled | up    | 2026-02-24T01:26:32.000000 |\n| cinder-backup    | testbed-node-2             | nova     | enabled | up    | 2026-02-24T01:26:32.000000 |\n+------------------+----------------------------+----------+---------+-------+----------------------------+\n\n# Neutron\n\n+ echo\n+ echo '# Neutron'\n+ echo\n+ openstack network agent list\n+--------------------------------------+------------------------------+----------------+-------------------+-------+-------+----------------------------+\n| ID                                   | Agent Type                   | Host           | Availability Zone | Alive | State | Binary                     |\n+--------------------------------------+------------------------------+----------------+-------------------+-------+-------+----------------------------+\n| testbed-node-3                       | OVN Controller agent         | testbed-node-3 |                   | :-)   | UP    | ovn-controller             |\n| testbed-node-5                       | OVN Controller agent         | testbed-node-5 |                   | :-)   | UP    | ovn-controller             |\n| testbed-node-0                       | OVN Controller Gateway agent | testbed-node-0 | nova              | :-)   | UP    | ovn-controller             |\n| testbed-node-2                       | OVN Controller Gateway agent | testbed-node-2 | nova              | :-)   | UP    | ovn-controller             |\n| testbed-node-4                       | OVN Controller agent         | testbed-node-4 |                   | :-)   | UP    | ovn-controller             |\n| testbed-node-1                       | OVN Controller Gateway agent | testbed-node-1 | nova              | :-)   | UP    | ovn-controller             |\n| e645415a-98f5-5758-8cd1-c47af282b5c0 | OVN Metadata agent           | testbed-node-3 |                   | :-)   | UP    | neutron-ovn-metadata-agent |\n| 4939696e-6092-5a33-bb73-b850064684df | OVN Metadata agent           | testbed-node-4 |                   | :-)   | UP    | neutron-ovn-metadata-agent |\n| 36b9d21c-9928-5c0a-9b27-73ac7a3e770c | OVN Metadata agent           | testbed-node-5 |                   | :-)   | UP    | neutron-ovn-metadata-agent |\n+--------------------------------------+------------------------------+----------------+-------------------+-------+-------+----------------------------+\n+ openstack network service provider list\n+---------------+------+---------+\n| Service Type  | Name | Default |\n+---------------+------+---------+\n| L3_ROUTER_NAT | ovn  | True    |\n+---------------+------+---------+\n\n# Nova\n\n+ echo\n+ echo '# Nova'\n+ echo\n+ openstack compute service list\n+--------------------------------------+----------------+----------------+----------+---------+-------+----------------------------+\n| ID                                   | Binary         | Host           | Zone     | Status  | State | Updated At                 |\n+--------------------------------------+----------------+----------------+----------+---------+-------+----------------------------+\n| 159da094-afa1-4e21-a1fa-329a2c038378 | nova-scheduler | testbed-node-0 | internal | enabled | up    | 2026-02-24T01:26:41.000000 |\n| 45821a88-2046-43d9-9358-eb37605d19f4 | nova-scheduler | testbed-node-2 | internal | enabled | up    | 2026-02-24T01:26:44.000000 |\n| e89363db-9a24-4ca8-9b4d-079af985725c | nova-scheduler | testbed-node-1 | internal | enabled | up    | 2026-02-24T01:26:44.000000 |\n| 3e23bf07-6948-4a8b-9ceb-004e5ce25e33 | nova-conductor | testbed-node-0 | internal | enabled | up    | 2026-02-24T01:26:46.000000 |\n| 9dea0c2d-25fa-4ca8-aa55-a2ea86ce238b | nova-conductor | testbed-node-2 | internal | enabled | up    | 2026-02-24T01:26:38.000000 |\n| f3aeba5e-1840-4fa1-9fd9-71f9ee8ebbfb | nova-conductor | testbed-node-1 | internal | enabled | up    | 2026-02-24T01:26:38.000000 |\n| 91bf9cdb-9bc1-45e9-9c85-614d018e902f | nova-compute   | testbed-node-4 | nova     | enabled | up    | 2026-02-24T01:26:38.000000 |\n| a10b69c5-9163-455e-b4ce-1c530538f267 | nova-compute   | testbed-node-3 | nova     | enabled | up    | 2026-02-24T01:26:38.000000 |\n| 2676ef64-9c2a-401a-8eb3-57e84238424a | nova-compute   | testbed-node-5 | nova     | enabled | up    | 2026-02-24T01:26:38.000000 |\n+--------------------------------------+----------------+----------------+----------+---------+-------+----------------------------+\n+ openstack hypervisor list\n+--------------------------------------+---------------------+-----------------+---------------+-------+\n| ID                                   | Hypervisor Hostname | Hypervisor Type | Host IP       | State |\n+--------------------------------------+---------------------+-----------------+---------------+-------+\n| 507d0ef5-91e2-4f9f-8f0f-2edc85623699 | testbed-node-4      | QEMU            | 192.168.16.14 | up    |\n| 0bcf54ec-2474-42b4-9bd1-39dbdc3f0488 | testbed-node-3      | QEMU            | 192.168.16.13 | up    |\n| 525e42fe-41e6-4fe2-9c2d-83a9aefd0b98 | testbed-node-5      | QEMU            | 192.168.16.15 | up    |\n+--------------------------------------+---------------------+-----------------+---------------+-------+\n\n# Run OpenStack test play\n\n+ echo\n+ echo '# Run OpenStack test play'\n+ echo\n+ osism apply --environment openstack test\n\u001b[32m2026-02-24 01:26:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTrying to run play test in environment openstack\u001b[0m\n\u001b[32m2026-02-24 01:27:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mPrepare task for execution of test.\u001b[0m\n\u001b[32m2026-02-24 01:27:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask bf061e28-2200-486a-a46f-075db86a3739 (test) was prepared for execution.\u001b[0m\n\u001b[32m2026-02-24 01:27:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task bf061e28-2200-486a-a46f-075db86a3739 (test) has been started and output is visible here.\u001b[0m\n\nPLAY [Create test project] *****************************************************\n\nTASK [Create test domain] ******************************************************\nTuesday 24 February 2026  01:27:08 +0000 (0:00:00.084)       0:00:00.084 ****** \n\u001b[0;33mchanged: [localhost]\u001b[0m\n\nTASK [Create test-admin user] **************************************************\nTuesday 24 February 2026  01:27:12 +0000 (0:00:04.304)       0:00:04.388 ****** \n\u001b[0;33mchanged: [localhost]\u001b[0m\n\nTASK [Add manager role to user test-admin] *************************************\nTuesday 24 February 2026  01:27:17 +0000 (0:00:04.771)       0:00:09.159 ****** \n\u001b[0;33mchanged: [localhost]\u001b[0m\n\nTASK [Create test project] *****************************************************\nTuesday 24 February 2026  01:27:24 +0000 (0:00:07.251)       0:00:16.411 ****** \n\u001b[0;33mchanged: [localhost]\u001b[0m\n\nTASK [Create test user] ********************************************************\nTuesday 24 February 2026  01:27:29 +0000 (0:00:04.585)       0:00:20.997 ****** \n\u001b[0;33mchanged: [localhost]\u001b[0m\n\nTASK [Add member roles to user test] *******************************************\nTuesday 24 February 2026  01:27:33 +0000 (0:00:04.540)       0:00:25.537 ****** \n\u001b[0;33mchanged: [localhost] => (item=load-balancer_member)\u001b[0m\n\u001b[0;33mchanged: [localhost] => (item=member)\u001b[0m\n\u001b[0;33mchanged: [localhost] => (item=creator)\u001b[0m\n\nTASK [Create test server group] ************************************************\nTuesday 24 February 2026  01:27:46 +0000 (0:00:12.578)       0:00:38.116 ****** \n\u001b[0;33mchanged: [localhost]\u001b[0m\n\nTASK [Create ssh security group] ***********************************************\nTuesday 24 February 2026  01:27:51 +0000 (0:00:04.802)       0:00:42.919 ****** \n\u001b[0;33mchanged: [localhost]\u001b[0m\n\nTASK [Add rule to ssh security group] ******************************************\nTuesday 24 February 2026  01:27:56 +0000 (0:00:05.379)       0:00:48.299 ****** \n\u001b[0;33mchanged: [localhost]\u001b[0m\n\nTASK [Create icmp security group] **********************************************\nTuesday 24 February 2026  01:28:01 +0000 (0:00:04.665)       0:00:52.964 ****** \n\u001b[0;33mchanged: [localhost]\u001b[0m\n\nTASK [Add rule to icmp security group] *****************************************\nTuesday 24 February 2026  01:28:05 +0000 (0:00:04.176)       0:00:57.141 ****** \n\u001b[0;33mchanged: [localhost]\u001b[0m\n\nTASK [Create test keypair] *****************************************************\nTuesday 24 February 2026  01:28:09 +0000 (0:00:04.370)       0:01:01.511 ****** \n\u001b[0;33mchanged: [localhost]\u001b[0m\n\nTASK [Create test network] *****************************************************\nTuesday 24 February 2026  01:28:14 +0000 (0:00:04.791)       0:01:06.302 ****** \n\u001b[0;33mchanged: [localhost]\u001b[0m\n\nTASK [Create test subnet] ******************************************************\nTuesday 24 February 2026  01:28:19 +0000 (0:00:05.117)       0:01:11.420 ****** \n\u001b[0;33mchanged: [localhost]\u001b[0m\n\nTASK [Create test router] ******************************************************\nTuesday 24 February 2026  01:28:25 +0000 (0:00:05.905)       0:01:17.326 ****** \n\u001b[0;33mchanged: [localhost]\u001b[0m\n\nPLAY [Manage test instances and volumes] ***************************************\n\nTASK [Get test server group] ***************************************************\nTuesday 24 February 2026  01:28:37 +0000 (0:00:11.931)       0:01:29.258 ****** \n\u001b[0;32mok: [localhost]\u001b[0m\n\nTASK [Detach test volume] ******************************************************\nTuesday 24 February 2026  01:28:42 +0000 (0:00:04.430)       0:01:33.688 ****** \n\u001b[0;36mskipping: [localhost]\u001b[0m\n\nTASK [Delete test volume] ******************************************************\nTuesday 24 February 2026  01:28:42 +0000 (0:00:00.062)       0:01:33.751 ****** \n\u001b[0;36mskipping: [localhost]\u001b[0m\n\nTASK [Delete test instances] ***************************************************\nTuesday 24 February 2026  01:28:42 +0000 (0:00:00.053)       0:01:33.804 ****** \n\u001b[0;36mskipping: [localhost] => (item=test-4) \u001b[0m\n\u001b[0;36mskipping: [localhost] => (item=test-3) \u001b[0m\n\u001b[0;36mskipping: [localhost] => (item=test-2) \u001b[0m\n\u001b[0;36mskipping: [localhost] => (item=test-1) \u001b[0m\n\u001b[0;36mskipping: [localhost] => (item=test) \u001b[0m\n\u001b[0;36mskipping: [localhost]\u001b[0m\n\nTASK [Wait for instance deletion to complete] **********************************\nTuesday 24 February 2026  01:28:42 +0000 (0:00:00.191)       0:01:33.996 ****** \n\u001b[0;36mskipping: [localhost]\u001b[0m\n\nTASK [Create test instances] ***************************************************\nTuesday 24 February 2026  01:28:42 +0000 (0:00:00.176)       0:01:34.173 ****** \n\u001b[0;33mchanged: [localhost] => (item=test)\u001b[0m\n\u001b[0;33mchanged: [localhost] => (item=test-1)\u001b[0m\n\u001b[0;33mchanged: [localhost] => (item=test-2)\u001b[0m\n\u001b[0;33mchanged: [localhost] => (item=test-3)\u001b[0m\n\u001b[0;33mchanged: [localhost] => (item=test-4)\u001b[0m\n\nTASK [Wait for instance creation to complete] **********************************\nTuesday 24 February 2026  01:28:48 +0000 (0:00:05.595)       0:01:39.769 ****** \n\u001b[1;30mFAILED - RETRYING: [localhost]: Wait for instance creation to complete (60 retries left).\u001b[0m\n\u001b[1;30mFAILED - RETRYING: [localhost]: Wait for instance creation to complete (59 retries left).\u001b[0m\n\u001b[1;30mFAILED - RETRYING: [localhost]: Wait for instance creation to complete (58 retries left).\u001b[0m\n\u001b[1;30mFAILED - RETRYING: [localhost]: Wait for instance creation to complete (57 retries left).\u001b[0m\n\u001b[0;33mchanged: [localhost] => (item={'failed': 0, 'started': 1, 'finished': 0, 'ansible_job_id': 'j138772053794.2657', 'results_file': '/ansible/.ansible_async/j138772053794.2657', 'changed': True, 'item': 'test', 'ansible_loop_var': 'item'})\u001b[0m\n\u001b[1;30mFAILED - RETRYING: [localhost]: Wait for instance creation to complete (60 retries left).\u001b[0m\n\u001b[0;33mchanged: [localhost] => (item={'failed': 0, 'started': 1, 'finished': 0, 'ansible_job_id': 'j906118553359.2682', 'results_file': '/ansible/.ansible_async/j906118553359.2682', 'changed': True, 'item': 'test-1', 'ansible_loop_var': 'item'})\u001b[0m\n\u001b[0;33mchanged: [localhost] => (item={'failed': 0, 'started': 1, 'finished': 0, 'ansible_job_id': 'j356775953011.2707', 'results_file': '/ansible/.ansible_async/j356775953011.2707', 'changed': True, 'item': 'test-2', 'ansible_loop_var': 'item'})\u001b[0m\n\u001b[0;33mchanged: [localhost] => (item={'failed': 0, 'started': 1, 'finished': 0, 'ansible_job_id': 'j955926461069.2739', 'results_file': '/ansible/.ansible_async/j955926461069.2739', 'changed': True, 'item': 'test-3', 'ansible_loop_var': 'item'})\u001b[0m\n\u001b[0;33mchanged: [localhost] => (item={'failed': 0, 'started': 1, 'finished': 0, 'ansible_job_id': 'j269502165248.2764', 'results_file': '/ansible/.ansible_async/j269502165248.2764', 'changed': True, 'item': 'test-4', 'ansible_loop_var': 'item'})\u001b[0m\n\nTASK [Add metadata to instances] ***********************************************\nTuesday 24 February 2026  01:29:46 +0000 (0:00:58.612)       0:02:38.381 ****** \n\u001b[0;33mchanged: [localhost] => (item=test)\u001b[0m\n\u001b[0;33mchanged: [localhost] => (item=test-1)\u001b[0m\n\u001b[0;33mchanged: [localhost] => (item=test-2)\u001b[0m\n\u001b[0;33mchanged: [localhost] => (item=test-3)\u001b[0m\n\u001b[0;33mchanged: [localhost] => (item=test-4)\u001b[0m\n\nTASK [Wait for metadata to be added] *******************************************\nTuesday 24 February 2026  01:29:51 +0000 (0:00:04.868)       0:02:43.249 ****** \n\u001b[1;30mFAILED - RETRYING: [localhost]: Wait for metadata to be added (30 retries left).\u001b[0m\n\u001b[0;33mchanged: [localhost] => (item={'failed': 0, 'started': 1, 'finished': 0, 'ansible_job_id': 'j239672281548.2875', 'results_file': '/ansible/.ansible_async/j239672281548.2875', 'changed': True, 'item': 'test', 'ansible_loop_var': 'item'})\u001b[0m\n\u001b[0;33mchanged: [localhost] => (item={'failed': 0, 'started': 1, 'finished': 0, 'ansible_job_id': 'j691765774518.2900', 'results_file': '/ansible/.ansible_async/j691765774518.2900', 'changed': True, 'item': 'test-1', 'ansible_loop_var': 'item'})\u001b[0m\n\u001b[0;33mchanged: [localhost] => (item={'failed': 0, 'started': 1, 'finished': 0, 'ansible_job_id': 'j210524382123.2925', 'results_file': '/ansible/.ansible_async/j210524382123.2925', 'changed': True, 'item': 'test-2', 'ansible_loop_var': 'item'})\u001b[0m\n\u001b[0;33mchanged: [localhost] => (item={'failed': 0, 'started': 1, 'finished': 0, 'ansible_job_id': 'j82192225245.2950', 'results_file': '/ansible/.ansible_async/j82192225245.2950', 'changed': True, 'item': 'test-3', 'ansible_loop_var': 'item'})\u001b[0m\n\u001b[0;33mchanged: [localhost] => (item={'failed': 0, 'started': 1, 'finished': 0, 'ansible_job_id': 'j577716052385.2975', 'results_file': '/ansible/.ansible_async/j577716052385.2975', 'changed': True, 'item': 'test-4', 'ansible_loop_var': 'item'})\u001b[0m\n\nTASK [Add tag to instances] ****************************************************\nTuesday 24 February 2026  01:30:01 +0000 (0:00:09.945)       0:02:53.195 ****** \n\u001b[0;33mchanged: [localhost] => (item=test)\u001b[0m\n\u001b[0;33mchanged: [localhost] => (item=test-1)\u001b[0m\n\u001b[0;33mchanged: [localhost] => (item=test-2)\u001b[0m\n\u001b[0;33mchanged: [localhost] => (item=test-3)\u001b[0m\n\u001b[0;33mchanged: [localhost] => (item=test-4)\u001b[0m\n\nTASK [Wait for tags to be added] ***********************************************\nTuesday 24 February 2026  01:30:06 +0000 (0:00:05.186)       0:02:58.381 ****** \n\u001b[1;30mFAILED - RETRYING: [localhost]: Wait for tags to be added (30 retries left).\u001b[0m\n\u001b[0;33mchanged: [localhost] => (item={'failed': 0, 'started': 1, 'finished': 0, 'ansible_job_id': 'j460463240913.3044', 'results_file': '/ansible/.ansible_async/j460463240913.3044', 'changed': True, 'item': 'test', 'ansible_loop_var': 'item'})\u001b[0m\n\u001b[0;33mchanged: [localhost] => (item={'failed': 0, 'started': 1, 'finished': 0, 'ansible_job_id': 'j7366433596.3069', 'results_file': '/ansible/.ansible_async/j7366433596.3069', 'changed': True, 'item': 'test-1', 'ansible_loop_var': 'item'})\u001b[0m\n\u001b[0;33mchanged: [localhost] => (item={'failed': 0, 'started': 1, 'finished': 0, 'ansible_job_id': 'j721472555207.3095', 'results_file': '/ansible/.ansible_async/j721472555207.3095', 'changed': True, 'item': 'test-2', 'ansible_loop_var': 'item'})\u001b[0m\n\u001b[0;33mchanged: [localhost] => (item={'failed': 0, 'started': 1, 'finished': 0, 'ansible_job_id': 'j842733406981.3121', 'results_file': '/ansible/.ansible_async/j842733406981.3121', 'changed': True, 'item': 'test-3', 'ansible_loop_var': 'item'})\u001b[0m\n\u001b[0;33mchanged: [localhost] => (item={'failed': 0, 'started': 1, 'finished': 0, 'ansible_job_id': 'j622682305399.3147', 'results_file': '/ansible/.ansible_async/j622682305399.3147', 'changed': True, 'item': 'test-4', 'ansible_loop_var': 'item'})\u001b[0m\n\nTASK [Create test volume] ******************************************************\nTuesday 24 February 2026  01:30:17 +0000 (0:00:10.389)       0:03:08.770 ****** \n\u001b[0;33mchanged: [localhost]\u001b[0m\n\nTASK [Attach test volume] ******************************************************\nTuesday 24 February 2026  01:30:23 +0000 (0:00:06.680)       0:03:15.451 ****** \n\u001b[0;33mchanged: [localhost]\u001b[0m\n\nTASK [Create floating ip address] **********************************************\nTuesday 24 February 2026  01:30:37 +0000 (0:00:13.569)       0:03:29.021 ****** \n\u001b[0;32mok: [localhost]\u001b[0m\n\nTASK [Print floating ip address] ***********************************************\nTuesday 24 February 2026  01:30:42 +0000 (0:00:04.934)       0:03:33.956 ****** \n\u001b[0;32mok: [localhost] => {\u001b[0m\n\u001b[0;32m    \"msg\": \"192.168.112.184\"\u001b[0m\n\u001b[0;32m}\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;33mlocalhost\u001b[0m                  : \u001b[0;32mok=26  \u001b[0m \u001b[0;33mchanged=23  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=4   \u001b[0m rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nTuesday 24 February 2026  01:30:42 +0000 (0:00:00.045)       0:03:34.001 ****** \n=============================================================================== \nWait for instance creation to complete --------------------------------- 58.61s\nAttach test volume ----------------------------------------------------- 13.57s\nAdd member roles to user test ------------------------------------------ 12.58s\nCreate test router ----------------------------------------------------- 11.93s\nWait for tags to be added ---------------------------------------------- 10.39s\nWait for metadata to be added ------------------------------------------- 9.94s\nAdd manager role to user test-admin ------------------------------------- 7.25s\nCreate test volume ------------------------------------------------------ 6.68s\nCreate test subnet ------------------------------------------------------ 5.91s\nCreate test instances --------------------------------------------------- 5.60s\nCreate ssh security group ----------------------------------------------- 5.38s\nAdd tag to instances ---------------------------------------------------- 5.19s\nCreate test network ----------------------------------------------------- 5.12s\nCreate floating ip address ---------------------------------------------- 4.93s\nAdd metadata to instances ----------------------------------------------- 4.87s\nCreate test server group ------------------------------------------------ 4.80s\nCreate test keypair ----------------------------------------------------- 4.79s\nCreate test-admin user -------------------------------------------------- 4.77s\nAdd rule to ssh security group ------------------------------------------ 4.67s\nCreate test project ----------------------------------------------------- 4.59s\n+ server_list\n+ openstack --os-cloud test server list\n+--------------------------------------+--------+--------+---------------------------------------+--------------------------+----------+\n| ID                                   | Name   | Status | Networks                              | Image                    | Flavor   |\n+--------------------------------------+--------+--------+---------------------------------------+--------------------------+----------+\n| 0b9309c0-5fce-46a5-86f1-8d5360f5077a | test-4 | ACTIVE | test=192.168.112.193, 192.168.200.119 | N/A (booted from volume) | SCS-1L-1 |\n| 36006f27-69d5-4fba-95d5-a09ad6e3d2b6 | test-3 | ACTIVE | test=192.168.112.192, 192.168.200.217 | N/A (booted from volume) | SCS-1L-1 |\n| 5ca62725-a636-4f3b-bb5e-7b2c661006b6 | test-1 | ACTIVE | test=192.168.112.182, 192.168.200.206 | N/A (booted from volume) | SCS-1L-1 |\n| 73a9cbe2-a4c7-4bb9-be1d-0f4606a87824 | test   | ACTIVE | test=192.168.112.184, 192.168.200.68  | N/A (booted from volume) | SCS-1L-1 |\n| e237ead4-6261-44c0-95f1-03bd2fd83f6a | test-2 | ACTIVE | test=192.168.112.141, 192.168.200.188 | N/A (booted from volume) | SCS-1L-1 |\n+--------------------------------------+--------+--------+---------------------------------------+--------------------------+----------+\n+ openstack --os-cloud test server show test\n+-------------------------------------+-------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+\n| Field                               | Value                                                                                                                                                                                                                                                                                                                                                                               |\n+-------------------------------------+-------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+\n| OS-DCF:diskConfig                   | MANUAL                                                                                                                                                                                                                                                                                                                                                                              |\n| OS-EXT-AZ:availability_zone         | nova                                                                                                                                                                                                                                                                                                                                                                                |\n| OS-EXT-SRV-ATTR:host                | None                                                                                                                                                                                                                                                                                                                                                                                |\n| OS-EXT-SRV-ATTR:hostname            | test                                                                                                                                                                                                                                                                                                                                                                                |\n| OS-EXT-SRV-ATTR:hypervisor_hostname | None                                                                                                                                                                                                                                                                                                                                                                                |\n| OS-EXT-SRV-ATTR:instance_name       | None                                                                                                                                                                                                                                                                                                                                                                                |\n| OS-EXT-SRV-ATTR:kernel_id           | None                                                                                                                                                                                                                                                                                                                                                                                |\n| OS-EXT-SRV-ATTR:launch_index        | None                                                                                                                                                                                                                                                                                                                                                                                |\n| OS-EXT-SRV-ATTR:ramdisk_id          | None                                                                                                                                                                                                                                                                                                                                                                                |\n| OS-EXT-SRV-ATTR:reservation_id      | None                                                                                                                                                                                                                                                                                                                                                                                |\n| OS-EXT-SRV-ATTR:root_device_name    | None                                                                                                                                                                                                                                                                                                                                                                                |\n| OS-EXT-SRV-ATTR:user_data           | None                                                                                                                                                                                                                                                                                                                                                                                |\n| OS-EXT-STS:power_state              | Running                                                                                                                                                                                                                                                                                                                                                                             |\n| OS-EXT-STS:task_state               | None                                                                                                                                                                                                                                                                                                                                                                                |\n| OS-EXT-STS:vm_state                 | active                                                                                                                                                                                                                                                                                                                                                                              |\n| OS-SRV-USG:launched_at              | 2026-02-24T01:29:21.000000                                                                                                                                                                                                                                                                                                                                                          |\n| OS-SRV-USG:terminated_at            | None                                                                                                                                                                                                                                                                                                                                                                                |\n| accessIPv4                          |                                                                                                                                                                                                                                                                                                                                                                                     |\n| accessIPv6                          |                                                                                                                                                                                                                                                                                                                                                                                     |\n| addresses                           | test=192.168.112.184, 192.168.200.68                                                                                                                                                                                                                                                                                                                                                |\n| config_drive                        |                                                                                                                                                                                                                                                                                                                                                                                     |\n| created                             | 2026-02-24T01:28:53Z                                                                                                                                                                                                                                                                                                                                                                |\n| description                         | None                                                                                                                                                                                                                                                                                                                                                                                |\n| flavor                              | description=, disk='0', ephemeral='0', extra_specs.hw_rng:allowed='true', extra_specs.scs:cpu-type='crowded-core', extra_specs.scs:disk0-type='network', extra_specs.scs:name-v1='SCS-1L:1', extra_specs.scs:name-v2='SCS-1L-1', id='SCS-1L-1', is_disabled=, is_public='True', location=, name='SCS-1L-1', original_name='SCS-1L-1', ram='1024', rxtx_factor=, swap='0', vcpus='1' |\n| hostId                              | 22a9aa40e3d2fc5564193a7c5a62e11757162f1fc18a38281aa3588f                                                                                                                                                                                                                                                                                                                            |\n| host_status                         | None                                                                                                                                                                                                                                                                                                                                                                                |\n| id                                  | 73a9cbe2-a4c7-4bb9-be1d-0f4606a87824                                                                                                                                                                                                                                                                                                                                                |\n| image                               | N/A (booted from volume)                                                                                                                                                                                                                                                                                                                                                            |\n| key_name                            | test                                                                                                                                                                                                                                                                                                                                                                                |\n| locked                              | False                                                                                                                                                                                                                                                                                                                                                                               |\n| locked_reason                       | None                                                                                                                                                                                                                                                                                                                                                                                |\n| name                                | test                                                                                                                                                                                                                                                                                                                                                                                |\n| pinned_availability_zone            | None                                                                                                                                                                                                                                                                                                                                                                                |\n| progress                            | 0                                                                                                                                                                                                                                                                                                                                                                                   |\n| project_id                          | b7fd8521139f4f698ef53541c2215403                                                                                                                                                                                                                                                                                                                                                    |\n| properties                          | hostname='test'                                                                                                                                                                                                                                                                                                                                                                     |\n| security_groups                     | name='icmp'                                                                                                                                                                                                                                                                                                                                                                         |\n|                                     | name='ssh'                                                                                                                                                                                                                                                                                                                                                                          |\n| server_groups                       | None                                                                                                                                                                                                                                                                                                                                                                                |\n| status                              | ACTIVE                                                                                                                                                                                                                                                                                                                                                                              |\n| tags                                | test                                                                                                                                                                                                                                                                                                                                                                                |\n| trusted_image_certificates          | None                                                                                                                                                                                                                                                                                                                                                                                |\n| updated                             | 2026-02-24T01:29:52Z                                                                                                                                                                                                                                                                                                                                                                |\n| user_id                             | e7bd38d58eed475ab28679fe129b0889                                                                                                                                                                                                                                                                                                                                                    |\n| volumes_attached                    | delete_on_termination='True', id='fc2483d8-4dc3-4f58-852c-acaadfc5f2d2'                                                                                                                                                                                                                                                                                                             |\n|                                     | delete_on_termination='False', id='1b918256-4d4e-4cf8-9f1c-3a729ddb7086'                                                                                                                                                                                                                                                                                                            |\n+-------------------------------------+-------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+\n+ openstack --os-cloud test server show test-1\n+-------------------------------------+-------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+\n| Field                               | Value                                                                                                                                                                                                                                                                                                                                                                               |\n+-------------------------------------+-------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+\n| OS-DCF:diskConfig                   | MANUAL                                                                                                                                                                                                                                                                                                                                                                              |\n| OS-EXT-AZ:availability_zone         | nova                                                                                                                                                                                                                                                                                                                                                                                |\n| OS-EXT-SRV-ATTR:host                | None                                                                                                                                                                                                                                                                                                                                                                                |\n| OS-EXT-SRV-ATTR:hostname            | test-1                                                                                                                                                                                                                                                                                                                                                                              |\n| OS-EXT-SRV-ATTR:hypervisor_hostname | None                                                                                                                                                                                                                                                                                                                                                                                |\n| OS-EXT-SRV-ATTR:instance_name       | None                                                                                                                                                                                                                                                                                                                                                                                |\n| OS-EXT-SRV-ATTR:kernel_id           | None                                                                                                                                                                                                                                                                                                                                                                                |\n| OS-EXT-SRV-ATTR:launch_index        | None                                                                                                                                                                                                                                                                                                                                                                                |\n| OS-EXT-SRV-ATTR:ramdisk_id          | None                                                                                                                                                                                                                                                                                                                                                                                |\n| OS-EXT-SRV-ATTR:reservation_id      | None                                                                                                                                                                                                                                                                                                                                                                                |\n| OS-EXT-SRV-ATTR:root_device_name    | None                                                                                                                                                                                                                                                                                                                                                                                |\n| OS-EXT-SRV-ATTR:user_data           | None                                                                                                                                                                                                                                                                                                                                                                                |\n| OS-EXT-STS:power_state              | Running                                                                                                                                                                                                                                                                                                                                                                             |\n| OS-EXT-STS:task_state               | None                                                                                                                                                                                                                                                                                                                                                                                |\n| OS-EXT-STS:vm_state                 | active                                                                                                                                                                                                                                                                                                                                                                              |\n| OS-SRV-USG:launched_at              | 2026-02-24T01:29:23.000000                                                                                                                                                                                                                                                                                                                                                          |\n| OS-SRV-USG:terminated_at            | None                                                                                                                                                                                                                                                                                                                                                                                |\n| accessIPv4                          |                                                                                                                                                                                                                                                                                                                                                                                     |\n| accessIPv6                          |                                                                                                                                                                                                                                                                                                                                                                                     |\n| addresses                           | test=192.168.112.182, 192.168.200.206                                                                                                                                                                                                                                                                                                                                               |\n| config_drive                        |                                                                                                                                                                                                                                                                                                                                                                                     |\n| created                             | 2026-02-24T01:28:53Z                                                                                                                                                                                                                                                                                                                                                                |\n| description                         | None                                                                                                                                                                                                                                                                                                                                                                                |\n| flavor                              | description=, disk='0', ephemeral='0', extra_specs.hw_rng:allowed='true', extra_specs.scs:cpu-type='crowded-core', extra_specs.scs:disk0-type='network', extra_specs.scs:name-v1='SCS-1L:1', extra_specs.scs:name-v2='SCS-1L-1', id='SCS-1L-1', is_disabled=, is_public='True', location=, name='SCS-1L-1', original_name='SCS-1L-1', ram='1024', rxtx_factor=, swap='0', vcpus='1' |\n| hostId                              | 22a9aa40e3d2fc5564193a7c5a62e11757162f1fc18a38281aa3588f                                                                                                                                                                                                                                                                                                                            |\n| host_status                         | None                                                                                                                                                                                                                                                                                                                                                                                |\n| id                                  | 5ca62725-a636-4f3b-bb5e-7b2c661006b6                                                                                                                                                                                                                                                                                                                                                |\n| image                               | N/A (booted from volume)                                                                                                                                                                                                                                                                                                                                                            |\n| key_name                            | test                                                                                                                                                                                                                                                                                                                                                                                |\n| locked                              | False                                                                                                                                                                                                                                                                                                                                                                               |\n| locked_reason                       | None                                                                                                                                                                                                                                                                                                                                                                                |\n| name                                | test-1                                                                                                                                                                                                                                                                                                                                                                              |\n| pinned_availability_zone            | None                                                                                                                                                                                                                                                                                                                                                                                |\n| progress                            | 0                                                                                                                                                                                                                                                                                                                                                                                   |\n| project_id                          | b7fd8521139f4f698ef53541c2215403                                                                                                                                                                                                                                                                                                                                                    |\n| properties                          | hostname='test-1'                                                                                                                                                                                                                                                                                                                                                                   |\n| security_groups                     | name='icmp'                                                                                                                                                                                                                                                                                                                                                                         |\n|                                     | name='ssh'                                                                                                                                                                                                                                                                                                                                                                          |\n| server_groups                       | None                                                                                                                                                                                                                                                                                                                                                                                |\n| status                              | ACTIVE                                                                                                                                                                                                                                                                                                                                                                              |\n| tags                                | test                                                                                                                                                                                                                                                                                                                                                                                |\n| trusted_image_certificates          | None                                                                                                                                                                                                                                                                                                                                                                                |\n| updated                             | 2026-02-24T01:29:53Z                                                                                                                                                                                                                                                                                                                                                                |\n| user_id                             | e7bd38d58eed475ab28679fe129b0889                                                                                                                                                                                                                                                                                                                                                    |\n| volumes_attached                    | delete_on_termination='True', id='2f4f7b85-91b4-430d-abe3-4e964e5dea87'                                                                                                                                                                                                                                                                                                             |\n+-------------------------------------+-------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+\n+ openstack --os-cloud test server show test-2\n+-------------------------------------+-------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+\n| Field                               | Value                                                                                                                                                                                                                                                                                                                                                                               |\n+-------------------------------------+-------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+\n| OS-DCF:diskConfig                   | MANUAL                                                                                                                                                                                                                                                                                                                                                                              |\n| OS-EXT-AZ:availability_zone         | nova                                                                                                                                                                                                                                                                                                                                                                                |\n| OS-EXT-SRV-ATTR:host                | None                                                                                                                                                                                                                                                                                                                                                                                |\n| OS-EXT-SRV-ATTR:hostname            | test-2                                                                                                                                                                                                                                                                                                                                                                              |\n| OS-EXT-SRV-ATTR:hypervisor_hostname | None                                                                                                                                                                                                                                                                                                                                                                                |\n| OS-EXT-SRV-ATTR:instance_name       | None                                                                                                                                                                                                                                                                                                                                                                                |\n| OS-EXT-SRV-ATTR:kernel_id           | None                                                                                                                                                                                                                                                                                                                                                                                |\n| OS-EXT-SRV-ATTR:launch_index        | None                                                                                                                                                                                                                                                                                                                                                                                |\n| OS-EXT-SRV-ATTR:ramdisk_id          | None                                                                                                                                                                                                                                                                                                                                                                                |\n| OS-EXT-SRV-ATTR:reservation_id      | None                                                                                                                                                                                                                                                                                                                                                                                |\n| OS-EXT-SRV-ATTR:root_device_name    | None                                                                                                                                                                                                                                                                                                                                                                                |\n| OS-EXT-SRV-ATTR:user_data           | None                                                                                                                                                                                                                                                                                                                                                                                |\n| OS-EXT-STS:power_state              | Running                                                                                                                                                                                                                                                                                                                                                                             |\n| OS-EXT-STS:task_state               | None                                                                                                                                                                                                                                                                                                                                                                                |\n| OS-EXT-STS:vm_state                 | active                                                                                                                                                                                                                                                                                                                                                                              |\n| OS-SRV-USG:launched_at              | 2026-02-24T01:29:21.000000                                                                                                                                                                                                                                                                                                                                                          |\n| OS-SRV-USG:terminated_at            | None                                                                                                                                                                                                                                                                                                                                                                                |\n| accessIPv4                          |                                                                                                                                                                                                                                                                                                                                                                                     |\n| accessIPv6                          |                                                                                                                                                                                                                                                                                                                                                                                     |\n| addresses                           | test=192.168.112.141, 192.168.200.188                                                                                                                                                                                                                                                                                                                                               |\n| config_drive                        |                                                                                                                                                                                                                                                                                                                                                                                     |\n| created                             | 2026-02-24T01:28:53Z                                                                                                                                                                                                                                                                                                                                                                |\n| description                         | None                                                                                                                                                                                                                                                                                                                                                                                |\n| flavor                              | description=, disk='0', ephemeral='0', extra_specs.hw_rng:allowed='true', extra_specs.scs:cpu-type='crowded-core', extra_specs.scs:disk0-type='network', extra_specs.scs:name-v1='SCS-1L:1', extra_specs.scs:name-v2='SCS-1L-1', id='SCS-1L-1', is_disabled=, is_public='True', location=, name='SCS-1L-1', original_name='SCS-1L-1', ram='1024', rxtx_factor=, swap='0', vcpus='1' |\n| hostId                              | 660187b7e152ec0335e13743959201b6eb9412e4dde6e4a1f5c8c5a7                                                                                                                                                                                                                                                                                                                            |\n| host_status                         | None                                                                                                                                                                                                                                                                                                                                                                                |\n| id                                  | e237ead4-6261-44c0-95f1-03bd2fd83f6a                                                                                                                                                                                                                                                                                                                                                |\n| image                               | N/A (booted from volume)                                                                                                                                                                                                                                                                                                                                                            |\n| key_name                            | test                                                                                                                                                                                                                                                                                                                                                                                |\n| locked                              | False                                                                                                                                                                                                                                                                                                                                                                               |\n| locked_reason                       | None                                                                                                                                                                                                                                                                                                                                                                                |\n| name                                | test-2                                                                                                                                                                                                                                                                                                                                                                              |\n| pinned_availability_zone            | None                                                                                                                                                                                                                                                                                                                                                                                |\n| progress                            | 0                                                                                                                                                                                                                                                                                                                                                                                   |\n| project_id                          | b7fd8521139f4f698ef53541c2215403                                                                                                                                                                                                                                                                                                                                                    |\n| properties                          | hostname='test-2'                                                                                                                                                                                                                                                                                                                                                                   |\n| security_groups                     | name='icmp'                                                                                                                                                                                                                                                                                                                                                                         |\n|                                     | name='ssh'                                                                                                                                                                                                                                                                                                                                                                          |\n| server_groups                       | None                                                                                                                                                                                                                                                                                                                                                                                |\n| status                              | ACTIVE                                                                                                                                                                                                                                                                                                                                                                              |\n| tags                                | test                                                                                                                                                                                                                                                                                                                                                                                |\n| trusted_image_certificates          | None                                                                                                                                                                                                                                                                                                                                                                                |\n| updated                             | 2026-02-24T01:29:54Z                                                                                                                                                                                                                                                                                                                                                                |\n| user_id                             | e7bd38d58eed475ab28679fe129b0889                                                                                                                                                                                                                                                                                                                                                    |\n| volumes_attached                    | delete_on_termination='True', id='461d2ade-dad0-452c-ab59-2b1683c48a38'                                                                                                                                                                                                                                                                                                             |\n+-------------------------------------+-------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+\n+ openstack --os-cloud test server show test-3\n+-------------------------------------+-------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+\n| Field                               | Value                                                                                                                                                                                                                                                                                                                                                                               |\n+-------------------------------------+-------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+\n| OS-DCF:diskConfig                   | MANUAL                                                                                                                                                                                                                                                                                                                                                                              |\n| OS-EXT-AZ:availability_zone         | nova                                                                                                                                                                                                                                                                                                                                                                                |\n| OS-EXT-SRV-ATTR:host                | None                                                                                                                                                                                                                                                                                                                                                                                |\n| OS-EXT-SRV-ATTR:hostname            | test-3                                                                                                                                                                                                                                                                                                                                                                              |\n| OS-EXT-SRV-ATTR:hypervisor_hostname | None                                                                                                                                                                                                                                                                                                                                                                                |\n| OS-EXT-SRV-ATTR:instance_name       | None                                                                                                                                                                                                                                                                                                                                                                                |\n| OS-EXT-SRV-ATTR:kernel_id           | None                                                                                                                                                                                                                                                                                                                                                                                |\n| OS-EXT-SRV-ATTR:launch_index        | None                                                                                                                                                                                                                                                                                                                                                                                |\n| OS-EXT-SRV-ATTR:ramdisk_id          | None                                                                                                                                                                                                                                                                                                                                                                                |\n| OS-EXT-SRV-ATTR:reservation_id      | None                                                                                                                                                                                                                                                                                                                                                                                |\n| OS-EXT-SRV-ATTR:root_device_name    | None                                                                                                                                                                                                                                                                                                                                                                                |\n| OS-EXT-SRV-ATTR:user_data           | None                                                                                                                                                                                                                                                                                                                                                                                |\n| OS-EXT-STS:power_state              | Running                                                                                                                                                                                                                                                                                                                                                                             |\n| OS-EXT-STS:task_state               | None                                                                                                                                                                                                                                                                                                                                                                                |\n| OS-EXT-STS:vm_state                 | active                                                                                                                                                                                                                                                                                                                                                                              |\n| OS-SRV-USG:launched_at              | 2026-02-24T01:29:23.000000                                                                                                                                                                                                                                                                                                                                                          |\n| OS-SRV-USG:terminated_at            | None                                                                                                                                                                                                                                                                                                                                                                                |\n| accessIPv4                          |                                                                                                                                                                                                                                                                                                                                                                                     |\n| accessIPv6                          |                                                                                                                                                                                                                                                                                                                                                                                     |\n| addresses                           | test=192.168.112.192, 192.168.200.217                                                                                                                                                                                                                                                                                                                                               |\n| config_drive                        |                                                                                                                                                                                                                                                                                                                                                                                     |\n| created                             | 2026-02-24T01:28:54Z                                                                                                                                                                                                                                                                                                                                                                |\n| description                         | None                                                                                                                                                                                                                                                                                                                                                                                |\n| flavor                              | description=, disk='0', ephemeral='0', extra_specs.hw_rng:allowed='true', extra_specs.scs:cpu-type='crowded-core', extra_specs.scs:disk0-type='network', extra_specs.scs:name-v1='SCS-1L:1', extra_specs.scs:name-v2='SCS-1L-1', id='SCS-1L-1', is_disabled=, is_public='True', location=, name='SCS-1L-1', original_name='SCS-1L-1', ram='1024', rxtx_factor=, swap='0', vcpus='1' |\n| hostId                              | 660187b7e152ec0335e13743959201b6eb9412e4dde6e4a1f5c8c5a7                                                                                                                                                                                                                                                                                                                            |\n| host_status                         | None                                                                                                                                                                                                                                                                                                                                                                                |\n| id                                  | 36006f27-69d5-4fba-95d5-a09ad6e3d2b6                                                                                                                                                                                                                                                                                                                                                |\n| image                               | N/A (booted from volume)                                                                                                                                                                                                                                                                                                                                                            |\n| key_name                            | test                                                                                                                                                                                                                                                                                                                                                                                |\n| locked                              | False                                                                                                                                                                                                                                                                                                                                                                               |\n| locked_reason                       | None                                                                                                                                                                                                                                                                                                                                                                                |\n| name                                | test-3                                                                                                                                                                                                                                                                                                                                                                              |\n| pinned_availability_zone            | None                                                                                                                                                                                                                                                                                                                                                                                |\n| progress                            | 0                                                                                                                                                                                                                                                                                                                                                                                   |\n| project_id                          | b7fd8521139f4f698ef53541c2215403                                                                                                                                                                                                                                                                                                                                                    |\n| properties                          | hostname='test-3'                                                                                                                                                                                                                                                                                                                                                                   |\n| security_groups                     | name='icmp'                                                                                                                                                                                                                                                                                                                                                                         |\n|                                     | name='ssh'                                                                                                                                                                                                                                                                                                                                                                          |\n| server_groups                       | None                                                                                                                                                                                                                                                                                                                                                                                |\n| status                              | ACTIVE                                                                                                                                                                                                                                                                                                                                                                              |\n| tags                                | test                                                                                                                                                                                                                                                                                                                                                                                |\n| trusted_image_certificates          | None                                                                                                                                                                                                                                                                                                                                                                                |\n| updated                             | 2026-02-24T01:29:54Z                                                                                                                                                                                                                                                                                                                                                                |\n| user_id                             | e7bd38d58eed475ab28679fe129b0889                                                                                                                                                                                                                                                                                                                                                    |\n| volumes_attached                    | delete_on_termination='True', id='17957e18-17d5-475f-a5a6-11013bd262d8'                                                                                                                                                                                                                                                                                                             |\n+-------------------------------------+-------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+\n+ openstack --os-cloud test server show test-4\n+-------------------------------------+-------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+\n| Field                               | Value                                                                                                                                                                                                                                                                                                                                                                               |\n+-------------------------------------+-------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+\n| OS-DCF:diskConfig                   | MANUAL                                                                                                                                                                                                                                                                                                                                                                              |\n| OS-EXT-AZ:availability_zone         | nova                                                                                                                                                                                                                                                                                                                                                                                |\n| OS-EXT-SRV-ATTR:host                | None                                                                                                                                                                                                                                                                                                                                                                                |\n| OS-EXT-SRV-ATTR:hostname            | test-4                                                                                                                                                                                                                                                                                                                                                                              |\n| OS-EXT-SRV-ATTR:hypervisor_hostname | None                                                                                                                                                                                                                                                                                                                                                                                |\n| OS-EXT-SRV-ATTR:instance_name       | None                                                                                                                                                                                                                                                                                                                                                                                |\n| OS-EXT-SRV-ATTR:kernel_id           | None                                                                                                                                                                                                                                                                                                                                                                                |\n| OS-EXT-SRV-ATTR:launch_index        | None                                                                                                                                                                                                                                                                                                                                                                                |\n| OS-EXT-SRV-ATTR:ramdisk_id          | None                                                                                                                                                                                                                                                                                                                                                                                |\n| OS-EXT-SRV-ATTR:reservation_id      | None                                                                                                                                                                                                                                                                                                                                                                                |\n| OS-EXT-SRV-ATTR:root_device_name    | None                                                                                                                                                                                                                                                                                                                                                                                |\n| OS-EXT-SRV-ATTR:user_data           | None                                                                                                                                                                                                                                                                                                                                                                                |\n| OS-EXT-STS:power_state              | Running                                                                                                                                                                                                                                                                                                                                                                             |\n| OS-EXT-STS:task_state               | None                                                                                                                                                                                                                                                                                                                                                                                |\n| OS-EXT-STS:vm_state                 | active                                                                                                                                                                                                                                                                                                                                                                              |\n| OS-SRV-USG:launched_at              | 2026-02-24T01:29:21.000000                                                                                                                                                                                                                                                                                                                                                          |\n| OS-SRV-USG:terminated_at            | None                                                                                                                                                                                                                                                                                                                                                                                |\n| accessIPv4                          |                                                                                                                                                                                                                                                                                                                                                                                     |\n| accessIPv6                          |                                                                                                                                                                                                                                                                                                                                                                                     |\n| addresses                           | test=192.168.112.193, 192.168.200.119                                                                                                                                                                                                                                                                                                                                               |\n| config_drive                        |                                                                                                                                                                                                                                                                                                                                                                                     |\n| created                             | 2026-02-24T01:28:55Z                                                                                                                                                                                                                                                                                                                                                                |\n| description                         | None                                                                                                                                                                                                                                                                                                                                                                                |\n| flavor                              | description=, disk='0', ephemeral='0', extra_specs.hw_rng:allowed='true', extra_specs.scs:cpu-type='crowded-core', extra_specs.scs:disk0-type='network', extra_specs.scs:name-v1='SCS-1L:1', extra_specs.scs:name-v2='SCS-1L-1', id='SCS-1L-1', is_disabled=, is_public='True', location=, name='SCS-1L-1', original_name='SCS-1L-1', ram='1024', rxtx_factor=, swap='0', vcpus='1' |\n| hostId                              | 22a9aa40e3d2fc5564193a7c5a62e11757162f1fc18a38281aa3588f                                                                                                                                                                                                                                                                                                                            |\n| host_status                         | None                                                                                                                                                                                                                                                                                                                                                                                |\n| id                                  | 0b9309c0-5fce-46a5-86f1-8d5360f5077a                                                                                                                                                                                                                                                                                                                                                |\n| image                               | N/A (booted from volume)                                                                                                                                                                                                                                                                                                                                                            |\n| key_name                            | test                                                                                                                                                                                                                                                                                                                                                                                |\n| locked                              | False                                                                                                                                                                                                                                                                                                                                                                               |\n| locked_reason                       | None                                                                                                                                                                                                                                                                                                                                                                                |\n| name                                | test-4                                                                                                                                                                                                                                                                                                                                                                              |\n| pinned_availability_zone            | None                                                                                                                                                                                                                                                                                                                                                                                |\n| progress                            | 0                                                                                                                                                                                                                                                                                                                                                                                   |\n| project_id                          | b7fd8521139f4f698ef53541c2215403                                                                                                                                                                                                                                                                                                                                                    |\n| properties                          | hostname='test-4'                                                                                                                                                                                                                                                                                                                                                                   |\n| security_groups                     | name='icmp'                                                                                                                                                                                                                                                                                                                                                                         |\n|                                     | name='ssh'                                                                                                                                                                                                                                                                                                                                                                          |\n| server_groups                       | None                                                                                                                                                                                                                                                                                                                                                                                |\n| status                              | ACTIVE                                                                                                                                                                                                                                                                                                                                                                              |\n| tags                                | test                                                                                                                                                                                                                                                                                                                                                                                |\n| trusted_image_certificates          | None                                                                                                                                                                                                                                                                                                                                                                                |\n| updated                             | 2026-02-24T01:29:55Z                                                                                                                                                                                                                                                                                                                                                                |\n| user_id                             | e7bd38d58eed475ab28679fe129b0889                                                                                                                                                                                                                                                                                                                                                    |\n| volumes_attached                    | delete_on_termination='True', id='90f82ae0-4c58-479a-837e-190419bc3244'                                                                                                                                                                                                                                                                                                             |\n+-------------------------------------+-------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+\n+ server_ping\n++ openstack --os-cloud test floating ip list --status ACTIVE -f value -c 'Floating IP Address'\n++ tr -d '\\r'\n+ for address in $(openstack --os-cloud test floating ip list --status ACTIVE -f value -c \"Floating IP Address\" | tr -d '\\r')\n+ ping -c3 192.168.112.193\nPING 192.168.112.193 (192.168.112.193) 56(84) bytes of data.\n64 bytes from 192.168.112.193: icmp_seq=1 ttl=63 time=8.78 ms\n64 bytes from 192.168.112.193: icmp_seq=2 ttl=63 time=2.35 ms\n64 bytes from 192.168.112.193: icmp_seq=3 ttl=63 time=2.29 ms\n\n--- 192.168.112.193 ping statistics ---\n3 packets transmitted, 3 received, 0% packet loss, time 2002ms\nrtt min/avg/max/mdev = 2.292/4.474/8.782/3.046 ms\n+ for address in $(openstack --os-cloud test floating ip list --status ACTIVE -f value -c \"Floating IP Address\" | tr -d '\\r')\n+ ping -c3 192.168.112.184\nPING 192.168.112.184 (192.168.112.184) 56(84) bytes of data.\n64 bytes from 192.168.112.184: icmp_seq=1 ttl=63 time=5.67 ms\n64 bytes from 192.168.112.184: icmp_seq=2 ttl=63 time=2.22 ms\n64 bytes from 192.168.112.184: icmp_seq=3 ttl=63 time=1.75 ms\n\n--- 192.168.112.184 ping statistics ---\n3 packets transmitted, 3 received, 0% packet loss, time 2003ms\nrtt min/avg/max/mdev = 1.749/3.210/5.667/1.747 ms\n+ for address in $(openstack --os-cloud test floating ip list --status ACTIVE -f value -c \"Floating IP Address\" | tr -d '\\r')\n+ ping -c3 192.168.112.182\nPING 192.168.112.182 (192.168.112.182) 56(84) bytes of data.\n64 bytes from 192.168.112.182: icmp_seq=1 ttl=63 time=4.92 ms\n64 bytes from 192.168.112.182: icmp_seq=2 ttl=63 time=2.41 ms\n64 bytes from 192.168.112.182: icmp_seq=3 ttl=63 time=1.83 ms\n\n--- 192.168.112.182 ping statistics ---\n3 packets transmitted, 3 received, 0% packet loss, time 2002ms\nrtt min/avg/max/mdev = 1.831/3.054/4.921/1.340 ms\n+ for address in $(openstack --os-cloud test floating ip list --status ACTIVE -f value -c \"Floating IP Address\" | tr -d '\\r')\n+ ping -c3 192.168.112.141\nPING 192.168.112.141 (192.168.112.141) 56(84) bytes of data.\n64 bytes from 192.168.112.141: icmp_seq=1 ttl=63 time=5.78 ms\n64 bytes from 192.168.112.141: icmp_seq=2 ttl=63 time=2.10 ms\n64 bytes from 192.168.112.141: icmp_seq=3 ttl=63 time=2.04 ms\n\n--- 192.168.112.141 ping statistics ---\n3 packets transmitted, 3 received, 0% packet loss, time 2003ms\nrtt min/avg/max/mdev = 2.037/3.305/5.782/1.751 ms\n+ for address in $(openstack --os-cloud test floating ip list --status ACTIVE -f value -c \"Floating IP Address\" | tr -d '\\r')\n+ ping -c3 192.168.112.192\nPING 192.168.112.192 (192.168.112.192) 56(84) bytes of data.\n64 bytes from 192.168.112.192: icmp_seq=1 ttl=63 time=5.25 ms\n64 bytes from 192.168.112.192: icmp_seq=2 ttl=63 time=1.97 ms\n64 bytes from 192.168.112.192: icmp_seq=3 ttl=63 time=1.64 ms\n\n--- 192.168.112.192 ping statistics ---\n3 packets transmitted, 3 received, 0% packet loss, time 2002ms\nrtt min/avg/max/mdev = 1.641/2.954/5.251/1.629 ms\n+ [[ latest == \\l\\a\\t\\e\\s\\t ]]\n+ compute_list\n+ osism manage compute list testbed-node-3\n\u001b[32m2026-02-24 01:31:17\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mUnable to get ansible vault password\u001b[0m\n\u001b[32m2026-02-24 01:31:17\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mUnable to get vault secret: [Errno 2] No such file or directory: '/share/ansible_vault_password.key'\u001b[0m\n\u001b[32m2026-02-24 01:31:17\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mDropping encrypted entries\u001b[0m\n+--------------------------------------+--------+----------+\n| ID                                   | Name   | Status   |\n|--------------------------------------+--------+----------|\n| 0b9309c0-5fce-46a5-86f1-8d5360f5077a | test-4 | ACTIVE   |\n| 5ca62725-a636-4f3b-bb5e-7b2c661006b6 | test-1 | ACTIVE   |\n| 73a9cbe2-a4c7-4bb9-be1d-0f4606a87824 | test   | ACTIVE   |\n+--------------------------------------+--------+----------+\n+ osism manage compute list testbed-node-4\n\u001b[32m2026-02-24 01:31:22\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mUnable to get ansible vault password\u001b[0m\n\u001b[32m2026-02-24 01:31:22\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mUnable to get vault secret: [Errno 2] No such file or directory: '/share/ansible_vault_password.key'\u001b[0m\n\u001b[32m2026-02-24 01:31:22\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mDropping encrypted entries\u001b[0m\n+--------------------------------------+--------+----------+\n| ID                                   | Name   | Status   |\n|--------------------------------------+--------+----------|\n| 36006f27-69d5-4fba-95d5-a09ad6e3d2b6 | test-3 | ACTIVE   |\n| e237ead4-6261-44c0-95f1-03bd2fd83f6a | test-2 | ACTIVE   |\n+--------------------------------------+--------+----------+\n+ osism manage compute list testbed-node-5\n\u001b[32m2026-02-24 01:31:26\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mUnable to get ansible vault password\u001b[0m\n\u001b[32m2026-02-24 01:31:26\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mUnable to get vault secret: [Errno 2] No such file or directory: '/share/ansible_vault_password.key'\u001b[0m\n\u001b[32m2026-02-24 01:31:26\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mDropping encrypted entries\u001b[0m\n+------+--------+----------+\n| ID   | Name   | Status   |\n|------+--------+----------|\n+------+--------+----------+\n+ osism manage compute migrate --yes --target testbed-node-3 testbed-node-4\n\u001b[32m2026-02-24 01:31:29\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mUnable to get ansible vault password\u001b[0m\n\u001b[32m2026-02-24 01:31:29\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mUnable to get vault secret: [Errno 2] No such file or directory: '/share/ansible_vault_password.key'\u001b[0m\n\u001b[32m2026-02-24 01:31:29\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mDropping encrypted entries\u001b[0m\n\u001b[32m2026-02-24 01:31:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migrating server 36006f27-69d5-4fba-95d5-a09ad6e3d2b6\u001b[0m\n\u001b[32m2026-02-24 01:31:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 36006f27-69d5-4fba-95d5-a09ad6e3d2b6 (test-3) is still in progress\u001b[0m\n\u001b[32m2026-02-24 01:31:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 36006f27-69d5-4fba-95d5-a09ad6e3d2b6 (test-3) is still in progress\u001b[0m\n\u001b[32m2026-02-24 01:31:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 36006f27-69d5-4fba-95d5-a09ad6e3d2b6 (test-3) is still in progress\u001b[0m\n\u001b[32m2026-02-24 01:31:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 36006f27-69d5-4fba-95d5-a09ad6e3d2b6 (test-3) is still in progress\u001b[0m\n\u001b[32m2026-02-24 01:31:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 36006f27-69d5-4fba-95d5-a09ad6e3d2b6 (test-3) is still in progress\u001b[0m\n\u001b[32m2026-02-24 01:31:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 36006f27-69d5-4fba-95d5-a09ad6e3d2b6 (test-3) is still in progress\u001b[0m\n\u001b[32m2026-02-24 01:31:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 36006f27-69d5-4fba-95d5-a09ad6e3d2b6 (test-3) is still in progress\u001b[0m\n\u001b[32m2026-02-24 01:32:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 36006f27-69d5-4fba-95d5-a09ad6e3d2b6 (test-3) is still in progress\u001b[0m\n\u001b[32m2026-02-24 01:32:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 36006f27-69d5-4fba-95d5-a09ad6e3d2b6 (test-3) completed with status ACTIVE\u001b[0m\n\u001b[32m2026-02-24 01:32:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migrating server e237ead4-6261-44c0-95f1-03bd2fd83f6a\u001b[0m\n\u001b[32m2026-02-24 01:32:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of e237ead4-6261-44c0-95f1-03bd2fd83f6a (test-2) is still in progress\u001b[0m\n\u001b[32m2026-02-24 01:32:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of e237ead4-6261-44c0-95f1-03bd2fd83f6a (test-2) is still in progress\u001b[0m\n\u001b[32m2026-02-24 01:32:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of e237ead4-6261-44c0-95f1-03bd2fd83f6a (test-2) is still in progress\u001b[0m\n\u001b[32m2026-02-24 01:32:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of e237ead4-6261-44c0-95f1-03bd2fd83f6a (test-2) is still in progress\u001b[0m\n\u001b[32m2026-02-24 01:32:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of e237ead4-6261-44c0-95f1-03bd2fd83f6a (test-2) is still in progress\u001b[0m\n\u001b[32m2026-02-24 01:32:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of e237ead4-6261-44c0-95f1-03bd2fd83f6a (test-2) is still in progress\u001b[0m\n\u001b[32m2026-02-24 01:32:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of e237ead4-6261-44c0-95f1-03bd2fd83f6a (test-2) is still in progress\u001b[0m\n\u001b[32m2026-02-24 01:32:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of e237ead4-6261-44c0-95f1-03bd2fd83f6a (test-2) is still in progress\u001b[0m\n\u001b[32m2026-02-24 01:32:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of e237ead4-6261-44c0-95f1-03bd2fd83f6a (test-2) completed with status ACTIVE\u001b[0m\n+ compute_list\n+ osism manage compute list testbed-node-3\n\u001b[32m2026-02-24 01:32:35\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mUnable to get ansible vault password\u001b[0m\n\u001b[32m2026-02-24 01:32:35\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mUnable to get vault secret: [Errno 2] No such file or directory: '/share/ansible_vault_password.key'\u001b[0m\n\u001b[32m2026-02-24 01:32:35\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mDropping encrypted entries\u001b[0m\n+--------------------------------------+--------+----------+\n| ID                                   | Name   | Status   |\n|--------------------------------------+--------+----------|\n| 0b9309c0-5fce-46a5-86f1-8d5360f5077a | test-4 | ACTIVE   |\n| 36006f27-69d5-4fba-95d5-a09ad6e3d2b6 | test-3 | ACTIVE   |\n| 5ca62725-a636-4f3b-bb5e-7b2c661006b6 | test-1 | ACTIVE   |\n| 73a9cbe2-a4c7-4bb9-be1d-0f4606a87824 | test   | ACTIVE   |\n| e237ead4-6261-44c0-95f1-03bd2fd83f6a | test-2 | ACTIVE   |\n+--------------------------------------+--------+----------+\n+ osism manage compute list testbed-node-4\n\u001b[32m2026-02-24 01:32:39\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mUnable to get ansible vault password\u001b[0m\n\u001b[32m2026-02-24 01:32:39\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mUnable to get vault secret: [Errno 2] No such file or directory: '/share/ansible_vault_password.key'\u001b[0m\n\u001b[32m2026-02-24 01:32:39\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mDropping encrypted entries\u001b[0m\n+------+--------+----------+\n| ID   | Name   | Status   |\n|------+--------+----------|\n+------+--------+----------+\n+ osism manage compute list testbed-node-5\n\u001b[32m2026-02-24 01:32:43\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mUnable to get ansible vault password\u001b[0m\n\u001b[32m2026-02-24 01:32:43\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mUnable to get vault secret: [Errno 2] No such file or directory: '/share/ansible_vault_password.key'\u001b[0m\n\u001b[32m2026-02-24 01:32:43\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mDropping encrypted entries\u001b[0m\n+------+--------+----------+\n| ID   | Name   | Status   |\n|------+--------+----------|\n+------+--------+----------+\n+ server_ping\n++ openstack --os-cloud test floating ip list --status ACTIVE -f value -c 'Floating IP Address'\n++ tr -d '\\r'\n+ for address in $(openstack --os-cloud test floating ip list --status ACTIVE -f value -c \"Floating IP Address\" | tr -d '\\r')\n+ ping -c3 192.168.112.193\nPING 192.168.112.193 (192.168.112.193) 56(84) bytes of data.\n64 bytes from 192.168.112.193: icmp_seq=1 ttl=63 time=8.80 ms\n64 bytes from 192.168.112.193: icmp_seq=2 ttl=63 time=2.40 ms\n64 bytes from 192.168.112.193: icmp_seq=3 ttl=63 time=1.92 ms\n\n--- 192.168.112.193 ping statistics ---\n3 packets transmitted, 3 received, 0% packet loss, time 2003ms\nrtt min/avg/max/mdev = 1.918/4.373/8.802/3.137 ms\n+ for address in $(openstack --os-cloud test floating ip list --status ACTIVE -f value -c \"Floating IP Address\" | tr -d '\\r')\n+ ping -c3 192.168.112.184\nPING 192.168.112.184 (192.168.112.184) 56(84) bytes of data.\n64 bytes from 192.168.112.184: icmp_seq=1 ttl=63 time=5.03 ms\n64 bytes from 192.168.112.184: icmp_seq=2 ttl=63 time=2.81 ms\n64 bytes from 192.168.112.184: icmp_seq=3 ttl=63 time=2.00 ms\n\n--- 192.168.112.184 ping statistics ---\n3 packets transmitted, 3 received, 0% packet loss, time 2003ms\nrtt min/avg/max/mdev = 2.002/3.281/5.033/1.281 ms\n+ for address in $(openstack --os-cloud test floating ip list --status ACTIVE -f value -c \"Floating IP Address\" | tr -d '\\r')\n+ ping -c3 192.168.112.182\nPING 192.168.112.182 (192.168.112.182) 56(84) bytes of data.\n64 bytes from 192.168.112.182: icmp_seq=1 ttl=63 time=6.10 ms\n64 bytes from 192.168.112.182: icmp_seq=2 ttl=63 time=1.99 ms\n64 bytes from 192.168.112.182: icmp_seq=3 ttl=63 time=1.81 ms\n\n--- 192.168.112.182 ping statistics ---\n3 packets transmitted, 3 received, 0% packet loss, time 2002ms\nrtt min/avg/max/mdev = 1.807/3.300/6.104/1.983 ms\n+ for address in $(openstack --os-cloud test floating ip list --status ACTIVE -f value -c \"Floating IP Address\" | tr -d '\\r')\n+ ping -c3 192.168.112.141\nPING 192.168.112.141 (192.168.112.141) 56(84) bytes of data.\n64 bytes from 192.168.112.141: icmp_seq=1 ttl=63 time=5.24 ms\n64 bytes from 192.168.112.141: icmp_seq=2 ttl=63 time=1.88 ms\n64 bytes from 192.168.112.141: icmp_seq=3 ttl=63 time=1.67 ms\n\n--- 192.168.112.141 ping statistics ---\n3 packets transmitted, 3 received, 0% packet loss, time 2003ms\nrtt min/avg/max/mdev = 1.668/2.928/5.236/1.634 ms\n+ for address in $(openstack --os-cloud test floating ip list --status ACTIVE -f value -c \"Floating IP Address\" | tr -d '\\r')\n+ ping -c3 192.168.112.192\nPING 192.168.112.192 (192.168.112.192) 56(84) bytes of data.\n64 bytes from 192.168.112.192: icmp_seq=1 ttl=63 time=5.20 ms\n64 bytes from 192.168.112.192: icmp_seq=2 ttl=63 time=2.93 ms\n64 bytes from 192.168.112.192: icmp_seq=3 ttl=63 time=2.17 ms\n\n--- 192.168.112.192 ping statistics ---\n3 packets transmitted, 3 received, 0% packet loss, time 2003ms\nrtt min/avg/max/mdev = 2.171/3.433/5.200/1.286 ms\n+ osism manage compute migrate --yes --target testbed-node-3 testbed-node-5\n\u001b[32m2026-02-24 01:33:00\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mUnable to get ansible vault password\u001b[0m\n\u001b[32m2026-02-24 01:33:00\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mUnable to get vault secret: [Errno 2] No such file or directory: '/share/ansible_vault_password.key'\u001b[0m\n\u001b[32m2026-02-24 01:33:00\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mDropping encrypted entries\u001b[0m\n\u001b[32m2026-02-24 01:33:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mNo migratable instances found on node testbed-node-5\u001b[0m\n+ compute_list\n+ osism manage compute list testbed-node-3\n\u001b[32m2026-02-24 01:33:04\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mUnable to get ansible vault password\u001b[0m\n\u001b[32m2026-02-24 01:33:04\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mUnable to get vault secret: [Errno 2] No such file or directory: '/share/ansible_vault_password.key'\u001b[0m\n\u001b[32m2026-02-24 01:33:04\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mDropping encrypted entries\u001b[0m\n+--------------------------------------+--------+----------+\n| ID                                   | Name   | Status   |\n|--------------------------------------+--------+----------|\n| 0b9309c0-5fce-46a5-86f1-8d5360f5077a | test-4 | ACTIVE   |\n| 36006f27-69d5-4fba-95d5-a09ad6e3d2b6 | test-3 | ACTIVE   |\n| 5ca62725-a636-4f3b-bb5e-7b2c661006b6 | test-1 | ACTIVE   |\n| 73a9cbe2-a4c7-4bb9-be1d-0f4606a87824 | test   | ACTIVE   |\n| e237ead4-6261-44c0-95f1-03bd2fd83f6a | test-2 | ACTIVE   |\n+--------------------------------------+--------+----------+\n+ osism manage compute list testbed-node-4\n\u001b[32m2026-02-24 01:33:08\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mUnable to get ansible vault password\u001b[0m\n\u001b[32m2026-02-24 01:33:08\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mUnable to get vault secret: [Errno 2] No such file or directory: '/share/ansible_vault_password.key'\u001b[0m\n\u001b[32m2026-02-24 01:33:08\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mDropping encrypted entries\u001b[0m\n+------+--------+----------+\n| ID   | Name   | Status   |\n|------+--------+----------|\n+------+--------+----------+\n+ osism manage compute list testbed-node-5\n\u001b[32m2026-02-24 01:33:12\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mUnable to get ansible vault password\u001b[0m\n\u001b[32m2026-02-24 01:33:12\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mUnable to get vault secret: [Errno 2] No such file or directory: '/share/ansible_vault_password.key'\u001b[0m\n\u001b[32m2026-02-24 01:33:12\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mDropping encrypted entries\u001b[0m\n+------+--------+----------+\n| ID   | Name   | Status   |\n|------+--------+----------|\n+------+--------+----------+\n+ server_ping\n++ openstack --os-cloud test floating ip list --status ACTIVE -f value -c 'Floating IP Address'\n++ tr -d '\\r'\n+ for address in $(openstack --os-cloud test floating ip list --status ACTIVE -f value -c \"Floating IP Address\" | tr -d '\\r')\n+ ping -c3 192.168.112.193\nPING 192.168.112.193 (192.168.112.193) 56(84) bytes of data.\n64 bytes from 192.168.112.193: icmp_seq=1 ttl=63 time=12.5 ms\n64 bytes from 192.168.112.193: icmp_seq=2 ttl=63 time=2.48 ms\n64 bytes from 192.168.112.193: icmp_seq=3 ttl=63 time=2.23 ms\n\n--- 192.168.112.193 ping statistics ---\n3 packets transmitted, 3 received, 0% packet loss, time 2003ms\nrtt min/avg/max/mdev = 2.231/5.726/12.466/4.766 ms\n+ for address in $(openstack --os-cloud test floating ip list --status ACTIVE -f value -c \"Floating IP Address\" | tr -d '\\r')\n+ ping -c3 192.168.112.184\nPING 192.168.112.184 (192.168.112.184) 56(84) bytes of data.\n64 bytes from 192.168.112.184: icmp_seq=1 ttl=63 time=5.77 ms\n64 bytes from 192.168.112.184: icmp_seq=2 ttl=63 time=2.71 ms\n64 bytes from 192.168.112.184: icmp_seq=3 ttl=63 time=2.21 ms\n\n--- 192.168.112.184 ping statistics ---\n3 packets transmitted, 3 received, 0% packet loss, time 2002ms\nrtt min/avg/max/mdev = 2.211/3.564/5.770/1.572 ms\n+ for address in $(openstack --os-cloud test floating ip list --status ACTIVE -f value -c \"Floating IP Address\" | tr -d '\\r')\n+ ping -c3 192.168.112.182\nPING 192.168.112.182 (192.168.112.182) 56(84) bytes of data.\n64 bytes from 192.168.112.182: icmp_seq=1 ttl=63 time=5.63 ms\n64 bytes from 192.168.112.182: icmp_seq=2 ttl=63 time=2.95 ms\n64 bytes from 192.168.112.182: icmp_seq=3 ttl=63 time=1.93 ms\n\n--- 192.168.112.182 ping statistics ---\n3 packets transmitted, 3 received, 0% packet loss, time 2003ms\nrtt min/avg/max/mdev = 1.928/3.500/5.626/1.559 ms\n+ for address in $(openstack --os-cloud test floating ip list --status ACTIVE -f value -c \"Floating IP Address\" | tr -d '\\r')\n+ ping -c3 192.168.112.141\nPING 192.168.112.141 (192.168.112.141) 56(84) bytes of data.\n64 bytes from 192.168.112.141: icmp_seq=1 ttl=63 time=5.45 ms\n64 bytes from 192.168.112.141: icmp_seq=2 ttl=63 time=1.91 ms\n64 bytes from 192.168.112.141: icmp_seq=3 ttl=63 time=2.18 ms\n\n--- 192.168.112.141 ping statistics ---\n3 packets transmitted, 3 received, 0% packet loss, time 2002ms\nrtt min/avg/max/mdev = 1.908/3.182/5.454/1.610 ms\n+ for address in $(openstack --os-cloud test floating ip list --status ACTIVE -f value -c \"Floating IP Address\" | tr -d '\\r')\n+ ping -c3 192.168.112.192\nPING 192.168.112.192 (192.168.112.192) 56(84) bytes of data.\n64 bytes from 192.168.112.192: icmp_seq=1 ttl=63 time=6.51 ms\n64 bytes from 192.168.112.192: icmp_seq=2 ttl=63 time=2.55 ms\n64 bytes from 192.168.112.192: icmp_seq=3 ttl=63 time=2.73 ms\n\n--- 192.168.112.192 ping statistics ---\n3 packets transmitted, 3 received, 0% packet loss, time 2003ms\nrtt min/avg/max/mdev = 2.548/3.928/6.505/1.823 ms\n+ osism manage compute migrate --yes --target testbed-node-4 testbed-node-3\n\u001b[32m2026-02-24 01:33:29\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mUnable to get ansible vault password\u001b[0m\n\u001b[32m2026-02-24 01:33:29\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mUnable to get vault secret: [Errno 2] No such file or directory: '/share/ansible_vault_password.key'\u001b[0m\n\u001b[32m2026-02-24 01:33:29\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mDropping encrypted entries\u001b[0m\n\u001b[32m2026-02-24 01:33:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migrating server 0b9309c0-5fce-46a5-86f1-8d5360f5077a\u001b[0m\n\u001b[32m2026-02-24 01:33:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 0b9309c0-5fce-46a5-86f1-8d5360f5077a (test-4) is still in progress\u001b[0m\n\u001b[32m2026-02-24 01:33:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 0b9309c0-5fce-46a5-86f1-8d5360f5077a (test-4) is still in progress\u001b[0m\n\u001b[32m2026-02-24 01:33:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 0b9309c0-5fce-46a5-86f1-8d5360f5077a (test-4) is still in progress\u001b[0m\n\u001b[32m2026-02-24 01:33:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 0b9309c0-5fce-46a5-86f1-8d5360f5077a (test-4) is still in progress\u001b[0m\n\u001b[32m2026-02-24 01:33:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 0b9309c0-5fce-46a5-86f1-8d5360f5077a (test-4) is still in progress\u001b[0m\n\u001b[32m2026-02-24 01:33:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 0b9309c0-5fce-46a5-86f1-8d5360f5077a (test-4) is still in progress\u001b[0m\n\u001b[32m2026-02-24 01:33:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 0b9309c0-5fce-46a5-86f1-8d5360f5077a (test-4) is still in progress\u001b[0m\n\u001b[32m2026-02-24 01:33:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 0b9309c0-5fce-46a5-86f1-8d5360f5077a (test-4) is still in progress\u001b[0m\n\u001b[32m2026-02-24 01:34:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 0b9309c0-5fce-46a5-86f1-8d5360f5077a (test-4) completed with status ACTIVE\u001b[0m\n\u001b[32m2026-02-24 01:34:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migrating server 36006f27-69d5-4fba-95d5-a09ad6e3d2b6\u001b[0m\n\u001b[32m2026-02-24 01:34:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 36006f27-69d5-4fba-95d5-a09ad6e3d2b6 (test-3) is still in progress\u001b[0m\n\u001b[32m2026-02-24 01:34:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 36006f27-69d5-4fba-95d5-a09ad6e3d2b6 (test-3) is still in progress\u001b[0m\n\u001b[32m2026-02-24 01:34:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 36006f27-69d5-4fba-95d5-a09ad6e3d2b6 (test-3) is still in progress\u001b[0m\n\u001b[32m2026-02-24 01:34:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 36006f27-69d5-4fba-95d5-a09ad6e3d2b6 (test-3) is still in progress\u001b[0m\n\u001b[32m2026-02-24 01:34:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 36006f27-69d5-4fba-95d5-a09ad6e3d2b6 (test-3) is still in progress\u001b[0m\n\u001b[32m2026-02-24 01:34:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 36006f27-69d5-4fba-95d5-a09ad6e3d2b6 (test-3) is still in progress\u001b[0m\n\u001b[32m2026-02-24 01:34:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 36006f27-69d5-4fba-95d5-a09ad6e3d2b6 (test-3) is still in progress\u001b[0m\n\u001b[32m2026-02-24 01:34:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 36006f27-69d5-4fba-95d5-a09ad6e3d2b6 (test-3) is still in progress\u001b[0m\n\u001b[32m2026-02-24 01:34:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 36006f27-69d5-4fba-95d5-a09ad6e3d2b6 (test-3) completed with status ACTIVE\u001b[0m\n\u001b[32m2026-02-24 01:34:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migrating server 5ca62725-a636-4f3b-bb5e-7b2c661006b6\u001b[0m\n\u001b[32m2026-02-24 01:34:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 5ca62725-a636-4f3b-bb5e-7b2c661006b6 (test-1) is still in progress\u001b[0m\n\u001b[32m2026-02-24 01:34:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 5ca62725-a636-4f3b-bb5e-7b2c661006b6 (test-1) is still in progress\u001b[0m\n\u001b[32m2026-02-24 01:34:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 5ca62725-a636-4f3b-bb5e-7b2c661006b6 (test-1) is still in progress\u001b[0m\n\u001b[32m2026-02-24 01:34:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 5ca62725-a636-4f3b-bb5e-7b2c661006b6 (test-1) is still in progress\u001b[0m\n\u001b[32m2026-02-24 01:34:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 5ca62725-a636-4f3b-bb5e-7b2c661006b6 (test-1) is still in progress\u001b[0m\n\u001b[32m2026-02-24 01:34:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 5ca62725-a636-4f3b-bb5e-7b2c661006b6 (test-1) is still in progress\u001b[0m\n\u001b[32m2026-02-24 01:34:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 5ca62725-a636-4f3b-bb5e-7b2c661006b6 (test-1) is still in progress\u001b[0m\n\u001b[32m2026-02-24 01:34:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 5ca62725-a636-4f3b-bb5e-7b2c661006b6 (test-1) is still in progress\u001b[0m\n\u001b[32m2026-02-24 01:35:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 5ca62725-a636-4f3b-bb5e-7b2c661006b6 (test-1) completed with status ACTIVE\u001b[0m\n\u001b[32m2026-02-24 01:35:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migrating server 73a9cbe2-a4c7-4bb9-be1d-0f4606a87824\u001b[0m\n\u001b[32m2026-02-24 01:35:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 73a9cbe2-a4c7-4bb9-be1d-0f4606a87824 (test) is still in progress\u001b[0m\n\u001b[32m2026-02-24 01:35:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 73a9cbe2-a4c7-4bb9-be1d-0f4606a87824 (test) is still in progress\u001b[0m\n\u001b[32m2026-02-24 01:35:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 73a9cbe2-a4c7-4bb9-be1d-0f4606a87824 (test) is still in progress\u001b[0m\n\u001b[32m2026-02-24 01:35:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 73a9cbe2-a4c7-4bb9-be1d-0f4606a87824 (test) is still in progress\u001b[0m\n\u001b[32m2026-02-24 01:35:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 73a9cbe2-a4c7-4bb9-be1d-0f4606a87824 (test) is still in progress\u001b[0m\n\u001b[32m2026-02-24 01:35:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 73a9cbe2-a4c7-4bb9-be1d-0f4606a87824 (test) is still in progress\u001b[0m\n\u001b[32m2026-02-24 01:35:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 73a9cbe2-a4c7-4bb9-be1d-0f4606a87824 (test) is still in progress\u001b[0m\n\u001b[32m2026-02-24 01:35:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 73a9cbe2-a4c7-4bb9-be1d-0f4606a87824 (test) is still in progress\u001b[0m\n\u001b[32m2026-02-24 01:35:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 73a9cbe2-a4c7-4bb9-be1d-0f4606a87824 (test) is still in progress\u001b[0m\n\u001b[32m2026-02-24 01:35:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 73a9cbe2-a4c7-4bb9-be1d-0f4606a87824 (test) is still in progress\u001b[0m\n\u001b[32m2026-02-24 01:35:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 73a9cbe2-a4c7-4bb9-be1d-0f4606a87824 (test) completed with status ACTIVE\u001b[0m\n\u001b[32m2026-02-24 01:35:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migrating server e237ead4-6261-44c0-95f1-03bd2fd83f6a\u001b[0m\n\u001b[32m2026-02-24 01:35:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of e237ead4-6261-44c0-95f1-03bd2fd83f6a (test-2) is still in progress\u001b[0m\n\u001b[32m2026-02-24 01:35:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of e237ead4-6261-44c0-95f1-03bd2fd83f6a (test-2) is still in progress\u001b[0m\n\u001b[32m2026-02-24 01:35:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of e237ead4-6261-44c0-95f1-03bd2fd83f6a (test-2) is still in progress\u001b[0m\n\u001b[32m2026-02-24 01:35:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of e237ead4-6261-44c0-95f1-03bd2fd83f6a (test-2) is still in progress\u001b[0m\n\u001b[32m2026-02-24 01:35:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of e237ead4-6261-44c0-95f1-03bd2fd83f6a (test-2) is still in progress\u001b[0m\n\u001b[32m2026-02-24 01:35:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of e237ead4-6261-44c0-95f1-03bd2fd83f6a (test-2) is still in progress\u001b[0m\n\u001b[32m2026-02-24 01:36:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of e237ead4-6261-44c0-95f1-03bd2fd83f6a (test-2) is still in progress\u001b[0m\n\u001b[32m2026-02-24 01:36:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of e237ead4-6261-44c0-95f1-03bd2fd83f6a (test-2) is still in progress\u001b[0m\n\u001b[32m2026-02-24 01:36:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of e237ead4-6261-44c0-95f1-03bd2fd83f6a (test-2) completed with status ACTIVE\u001b[0m\n+ compute_list\n+ osism manage compute list testbed-node-3\n\u001b[32m2026-02-24 01:36:07\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mUnable to get ansible vault password\u001b[0m\n\u001b[32m2026-02-24 01:36:07\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mUnable to get vault secret: [Errno 2] No such file or directory: '/share/ansible_vault_password.key'\u001b[0m\n\u001b[32m2026-02-24 01:36:07\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mDropping encrypted entries\u001b[0m\n+------+--------+----------+\n| ID   | Name   | Status   |\n|------+--------+----------|\n+------+--------+----------+\n+ osism manage compute list testbed-node-4\n\u001b[32m2026-02-24 01:36:11\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mUnable to get ansible vault password\u001b[0m\n\u001b[32m2026-02-24 01:36:11\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mUnable to get vault secret: [Errno 2] No such file or directory: '/share/ansible_vault_password.key'\u001b[0m\n\u001b[32m2026-02-24 01:36:11\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mDropping encrypted entries\u001b[0m\n+--------------------------------------+--------+----------+\n| ID                                   | Name   | Status   |\n|--------------------------------------+--------+----------|\n| 0b9309c0-5fce-46a5-86f1-8d5360f5077a | test-4 | ACTIVE   |\n| 36006f27-69d5-4fba-95d5-a09ad6e3d2b6 | test-3 | ACTIVE   |\n| 5ca62725-a636-4f3b-bb5e-7b2c661006b6 | test-1 | ACTIVE   |\n| 73a9cbe2-a4c7-4bb9-be1d-0f4606a87824 | test   | ACTIVE   |\n| e237ead4-6261-44c0-95f1-03bd2fd83f6a | test-2 | ACTIVE   |\n+--------------------------------------+--------+----------+\n+ osism manage compute list testbed-node-5\n\u001b[32m2026-02-24 01:36:14\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mUnable to get ansible vault password\u001b[0m\n\u001b[32m2026-02-24 01:36:14\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mUnable to get vault secret: [Errno 2] No such file or directory: '/share/ansible_vault_password.key'\u001b[0m\n\u001b[32m2026-02-24 01:36:14\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mDropping encrypted entries\u001b[0m\n+------+--------+----------+\n| ID   | Name   | Status   |\n|------+--------+----------|\n+------+--------+----------+\n+ server_ping\n++ tr -d '\\r'\n++ openstack --os-cloud test floating ip list --status ACTIVE -f value -c 'Floating IP Address'\n+ for address in $(openstack --os-cloud test floating ip list --status ACTIVE -f value -c \"Floating IP Address\" | tr -d '\\r')\n+ ping -c3 192.168.112.193\nPING 192.168.112.193 (192.168.112.193) 56(84) bytes of data.\n64 bytes from 192.168.112.193: icmp_seq=1 ttl=63 time=9.30 ms\n64 bytes from 192.168.112.193: icmp_seq=2 ttl=63 time=2.07 ms\n64 bytes from 192.168.112.193: icmp_seq=3 ttl=63 time=1.91 ms\n\n--- 192.168.112.193 ping statistics ---\n3 packets transmitted, 3 received, 0% packet loss, time 2003ms\nrtt min/avg/max/mdev = 1.912/4.427/9.296/3.443 ms\n+ for address in $(openstack --os-cloud test floating ip list --status ACTIVE -f value -c \"Floating IP Address\" | tr -d '\\r')\n+ ping -c3 192.168.112.184\nPING 192.168.112.184 (192.168.112.184) 56(84) bytes of data.\n64 bytes from 192.168.112.184: icmp_seq=1 ttl=63 time=6.00 ms\n64 bytes from 192.168.112.184: icmp_seq=2 ttl=63 time=2.92 ms\n64 bytes from 192.168.112.184: icmp_seq=3 ttl=63 time=1.82 ms\n\n--- 192.168.112.184 ping statistics ---\n3 packets transmitted, 3 received, 0% packet loss, time 2003ms\nrtt min/avg/max/mdev = 1.822/3.582/6.001/1.768 ms\n+ for address in $(openstack --os-cloud test floating ip list --status ACTIVE -f value -c \"Floating IP Address\" | tr -d '\\r')\n+ ping -c3 192.168.112.182\nPING 192.168.112.182 (192.168.112.182) 56(84) bytes of data.\n64 bytes from 192.168.112.182: icmp_seq=1 ttl=63 time=4.97 ms\n64 bytes from 192.168.112.182: icmp_seq=2 ttl=63 time=2.17 ms\n64 bytes from 192.168.112.182: icmp_seq=3 ttl=63 time=1.76 ms\n\n--- 192.168.112.182 ping statistics ---\n3 packets transmitted, 3 received, 0% packet loss, time 2002ms\nrtt min/avg/max/mdev = 1.762/2.968/4.970/1.425 ms\n+ for address in $(openstack --os-cloud test floating ip list --status ACTIVE -f value -c \"Floating IP Address\" | tr -d '\\r')\n+ ping -c3 192.168.112.141\nPING 192.168.112.141 (192.168.112.141) 56(84) bytes of data.\n64 bytes from 192.168.112.141: icmp_seq=1 ttl=63 time=5.99 ms\n64 bytes from 192.168.112.141: icmp_seq=2 ttl=63 time=2.14 ms\n64 bytes from 192.168.112.141: icmp_seq=3 ttl=63 time=1.76 ms\n\n--- 192.168.112.141 ping statistics ---\n3 packets transmitted, 3 received, 0% packet loss, time 2002ms\nrtt min/avg/max/mdev = 1.759/3.298/5.991/1.910 ms\n+ for address in $(openstack --os-cloud test floating ip list --status ACTIVE -f value -c \"Floating IP Address\" | tr -d '\\r')\n+ ping -c3 192.168.112.192\nPING 192.168.112.192 (192.168.112.192) 56(84) bytes of data.\n64 bytes from 192.168.112.192: icmp_seq=1 ttl=63 time=5.50 ms\n64 bytes from 192.168.112.192: icmp_seq=2 ttl=63 time=1.96 ms\n64 bytes from 192.168.112.192: icmp_seq=3 ttl=63 time=1.34 ms\n\n--- 192.168.112.192 ping statistics ---\n3 packets transmitted, 3 received, 0% packet loss, time 2003ms\nrtt min/avg/max/mdev = 1.335/2.933/5.504/1.835 ms\n+ osism manage compute migrate --yes --target testbed-node-5 testbed-node-4\n\u001b[32m2026-02-24 01:36:31\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mUnable to get ansible vault password\u001b[0m\n\u001b[32m2026-02-24 01:36:31\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mUnable to get vault secret: [Errno 2] No such file or directory: '/share/ansible_vault_password.key'\u001b[0m\n\u001b[32m2026-02-24 01:36:31\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mDropping encrypted entries\u001b[0m\n\u001b[32m2026-02-24 01:36:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migrating server 0b9309c0-5fce-46a5-86f1-8d5360f5077a\u001b[0m\n\u001b[32m2026-02-24 01:36:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 0b9309c0-5fce-46a5-86f1-8d5360f5077a (test-4) is still in progress\u001b[0m\n\u001b[32m2026-02-24 01:36:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 0b9309c0-5fce-46a5-86f1-8d5360f5077a (test-4) is still in progress\u001b[0m\n\u001b[32m2026-02-24 01:36:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 0b9309c0-5fce-46a5-86f1-8d5360f5077a (test-4) is still in progress\u001b[0m\n\u001b[32m2026-02-24 01:36:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 0b9309c0-5fce-46a5-86f1-8d5360f5077a (test-4) is still in progress\u001b[0m\n\u001b[32m2026-02-24 01:36:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 0b9309c0-5fce-46a5-86f1-8d5360f5077a (test-4) is still in progress\u001b[0m\n\u001b[32m2026-02-24 01:36:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 0b9309c0-5fce-46a5-86f1-8d5360f5077a (test-4) is still in progress\u001b[0m\n\u001b[32m2026-02-24 01:36:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 0b9309c0-5fce-46a5-86f1-8d5360f5077a (test-4) is still in progress\u001b[0m\n\u001b[32m2026-02-24 01:37:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 0b9309c0-5fce-46a5-86f1-8d5360f5077a (test-4) is still in progress\u001b[0m\n\u001b[32m2026-02-24 01:37:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 0b9309c0-5fce-46a5-86f1-8d5360f5077a (test-4) is still in progress\u001b[0m\n\u001b[32m2026-02-24 01:37:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 0b9309c0-5fce-46a5-86f1-8d5360f5077a (test-4) is still in progress\u001b[0m\n\u001b[32m2026-02-24 01:37:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 0b9309c0-5fce-46a5-86f1-8d5360f5077a (test-4) is still in progress\u001b[0m\n\u001b[32m2026-02-24 01:37:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 0b9309c0-5fce-46a5-86f1-8d5360f5077a (test-4) completed with status ACTIVE\u001b[0m\n\u001b[32m2026-02-24 01:37:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migrating server 36006f27-69d5-4fba-95d5-a09ad6e3d2b6\u001b[0m\n\u001b[32m2026-02-24 01:37:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 36006f27-69d5-4fba-95d5-a09ad6e3d2b6 (test-3) is still in progress\u001b[0m\n\u001b[32m2026-02-24 01:37:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 36006f27-69d5-4fba-95d5-a09ad6e3d2b6 (test-3) is still in progress\u001b[0m\n\u001b[32m2026-02-24 01:37:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 36006f27-69d5-4fba-95d5-a09ad6e3d2b6 (test-3) is still in progress\u001b[0m\n\u001b[32m2026-02-24 01:37:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 36006f27-69d5-4fba-95d5-a09ad6e3d2b6 (test-3) is still in progress\u001b[0m\n\u001b[32m2026-02-24 01:37:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 36006f27-69d5-4fba-95d5-a09ad6e3d2b6 (test-3) is still in progress\u001b[0m\n\u001b[32m2026-02-24 01:37:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 36006f27-69d5-4fba-95d5-a09ad6e3d2b6 (test-3) is still in progress\u001b[0m\n\u001b[32m2026-02-24 01:37:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 36006f27-69d5-4fba-95d5-a09ad6e3d2b6 (test-3) is still in progress\u001b[0m\n\u001b[32m2026-02-24 01:37:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 36006f27-69d5-4fba-95d5-a09ad6e3d2b6 (test-3) is still in progress\u001b[0m\n\u001b[32m2026-02-24 01:37:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 36006f27-69d5-4fba-95d5-a09ad6e3d2b6 (test-3) completed with status ACTIVE\u001b[0m\n\u001b[32m2026-02-24 01:37:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migrating server 5ca62725-a636-4f3b-bb5e-7b2c661006b6\u001b[0m\n\u001b[32m2026-02-24 01:37:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 5ca62725-a636-4f3b-bb5e-7b2c661006b6 (test-1) is still in progress\u001b[0m\n\u001b[32m2026-02-24 01:37:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 5ca62725-a636-4f3b-bb5e-7b2c661006b6 (test-1) is still in progress\u001b[0m\n\u001b[32m2026-02-24 01:37:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 5ca62725-a636-4f3b-bb5e-7b2c661006b6 (test-1) is still in progress\u001b[0m\n\u001b[32m2026-02-24 01:37:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 5ca62725-a636-4f3b-bb5e-7b2c661006b6 (test-1) is still in progress\u001b[0m\n\u001b[32m2026-02-24 01:37:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 5ca62725-a636-4f3b-bb5e-7b2c661006b6 (test-1) is still in progress\u001b[0m\n\u001b[32m2026-02-24 01:38:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 5ca62725-a636-4f3b-bb5e-7b2c661006b6 (test-1) is still in progress\u001b[0m\n\u001b[32m2026-02-24 01:38:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 5ca62725-a636-4f3b-bb5e-7b2c661006b6 (test-1) is still in progress\u001b[0m\n\u001b[32m2026-02-24 01:38:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 5ca62725-a636-4f3b-bb5e-7b2c661006b6 (test-1) is still in progress\u001b[0m\n\u001b[32m2026-02-24 01:38:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 5ca62725-a636-4f3b-bb5e-7b2c661006b6 (test-1) completed with status ACTIVE\u001b[0m\n\u001b[32m2026-02-24 01:38:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migrating server 73a9cbe2-a4c7-4bb9-be1d-0f4606a87824\u001b[0m\n\u001b[32m2026-02-24 01:38:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 73a9cbe2-a4c7-4bb9-be1d-0f4606a87824 (test) is still in progress\u001b[0m\n\u001b[32m2026-02-24 01:38:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 73a9cbe2-a4c7-4bb9-be1d-0f4606a87824 (test) is still in progress\u001b[0m\n\u001b[32m2026-02-24 01:38:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 73a9cbe2-a4c7-4bb9-be1d-0f4606a87824 (test) is still in progress\u001b[0m\n\u001b[32m2026-02-24 01:38:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 73a9cbe2-a4c7-4bb9-be1d-0f4606a87824 (test) is still in progress\u001b[0m\n\u001b[32m2026-02-24 01:38:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 73a9cbe2-a4c7-4bb9-be1d-0f4606a87824 (test) is still in progress\u001b[0m\n\u001b[32m2026-02-24 01:38:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 73a9cbe2-a4c7-4bb9-be1d-0f4606a87824 (test) is still in progress\u001b[0m\n\u001b[32m2026-02-24 01:38:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 73a9cbe2-a4c7-4bb9-be1d-0f4606a87824 (test) is still in progress\u001b[0m\n\u001b[32m2026-02-24 01:38:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 73a9cbe2-a4c7-4bb9-be1d-0f4606a87824 (test) is still in progress\u001b[0m\n\u001b[32m2026-02-24 01:38:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 73a9cbe2-a4c7-4bb9-be1d-0f4606a87824 (test) is still in progress\u001b[0m\n\u001b[32m2026-02-24 01:38:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 73a9cbe2-a4c7-4bb9-be1d-0f4606a87824 (test) is still in progress\u001b[0m\n\u001b[32m2026-02-24 01:38:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 73a9cbe2-a4c7-4bb9-be1d-0f4606a87824 (test) completed with status ACTIVE\u001b[0m\n\u001b[32m2026-02-24 01:38:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migrating server e237ead4-6261-44c0-95f1-03bd2fd83f6a\u001b[0m\n\u001b[32m2026-02-24 01:38:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of e237ead4-6261-44c0-95f1-03bd2fd83f6a (test-2) is still in progress\u001b[0m\n\u001b[32m2026-02-24 01:38:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of e237ead4-6261-44c0-95f1-03bd2fd83f6a (test-2) is still in progress\u001b[0m\n\u001b[32m2026-02-24 01:38:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of e237ead4-6261-44c0-95f1-03bd2fd83f6a (test-2) is still in progress\u001b[0m\n\u001b[32m2026-02-24 01:38:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of e237ead4-6261-44c0-95f1-03bd2fd83f6a (test-2) is still in progress\u001b[0m\n\u001b[32m2026-02-24 01:39:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of e237ead4-6261-44c0-95f1-03bd2fd83f6a (test-2) is still in progress\u001b[0m\n\u001b[32m2026-02-24 01:39:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of e237ead4-6261-44c0-95f1-03bd2fd83f6a (test-2) is still in progress\u001b[0m\n\u001b[32m2026-02-24 01:39:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of e237ead4-6261-44c0-95f1-03bd2fd83f6a (test-2) is still in progress\u001b[0m\n\u001b[32m2026-02-24 01:39:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of e237ead4-6261-44c0-95f1-03bd2fd83f6a (test-2) is still in progress\u001b[0m\n\u001b[32m2026-02-24 01:39:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of e237ead4-6261-44c0-95f1-03bd2fd83f6a (test-2) completed with status ACTIVE\u001b[0m\n+ compute_list\n+ osism manage compute list testbed-node-3\n\u001b[32m2026-02-24 01:39:12\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mUnable to get ansible vault password\u001b[0m\n\u001b[32m2026-02-24 01:39:12\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mUnable to get vault secret: [Errno 2] No such file or directory: '/share/ansible_vault_password.key'\u001b[0m\n\u001b[32m2026-02-24 01:39:12\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mDropping encrypted entries\u001b[0m\n+------+--------+----------+\n| ID   | Name   | Status   |\n|------+--------+----------|\n+------+--------+----------+\n+ osism manage compute list testbed-node-4\n\u001b[32m2026-02-24 01:39:15\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mUnable to get ansible vault password\u001b[0m\n\u001b[32m2026-02-24 01:39:15\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mUnable to get vault secret: [Errno 2] No such file or directory: '/share/ansible_vault_password.key'\u001b[0m\n\u001b[32m2026-02-24 01:39:15\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mDropping encrypted entries\u001b[0m\n+------+--------+----------+\n| ID   | Name   | Status   |\n|------+--------+----------|\n+------+--------+----------+\n+ osism manage compute list testbed-node-5\n\u001b[32m2026-02-24 01:39:19\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mUnable to get ansible vault password\u001b[0m\n\u001b[32m2026-02-24 01:39:19\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mUnable to get vault secret: [Errno 2] No such file or directory: '/share/ansible_vault_password.key'\u001b[0m\n\u001b[32m2026-02-24 01:39:19\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mDropping encrypted entries\u001b[0m\n+--------------------------------------+--------+----------+\n| ID                                   | Name   | Status   |\n|--------------------------------------+--------+----------|\n| 0b9309c0-5fce-46a5-86f1-8d5360f5077a | test-4 | ACTIVE   |\n| 36006f27-69d5-4fba-95d5-a09ad6e3d2b6 | test-3 | ACTIVE   |\n| 5ca62725-a636-4f3b-bb5e-7b2c661006b6 | test-1 | ACTIVE   |\n| 73a9cbe2-a4c7-4bb9-be1d-0f4606a87824 | test   | ACTIVE   |\n| e237ead4-6261-44c0-95f1-03bd2fd83f6a | test-2 | ACTIVE   |\n+--------------------------------------+--------+----------+\n+ server_ping\n++ openstack --os-cloud test floating ip list --status ACTIVE -f value -c 'Floating IP Address'\n++ tr -d '\\r'\n+ for address in $(openstack --os-cloud test floating ip list --status ACTIVE -f value -c \"Floating IP Address\" | tr -d '\\r')\n+ ping -c3 192.168.112.193\nPING 192.168.112.193 (192.168.112.193) 56(84) bytes of data.\n64 bytes from 192.168.112.193: icmp_seq=1 ttl=63 time=6.05 ms\n64 bytes from 192.168.112.193: icmp_seq=2 ttl=63 time=2.51 ms\n64 bytes from 192.168.112.193: icmp_seq=3 ttl=63 time=1.73 ms\n\n--- 192.168.112.193 ping statistics ---\n3 packets transmitted, 3 received, 0% packet loss, time 2003ms\nrtt min/avg/max/mdev = 1.732/3.428/6.046/1.877 ms\n+ for address in $(openstack --os-cloud test floating ip list --status ACTIVE -f value -c \"Floating IP Address\" | tr -d '\\r')\n+ ping -c3 192.168.112.184\nPING 192.168.112.184 (192.168.112.184) 56(84) bytes of data.\n64 bytes from 192.168.112.184: icmp_seq=1 ttl=63 time=6.48 ms\n64 bytes from 192.168.112.184: icmp_seq=2 ttl=63 time=2.72 ms\n64 bytes from 192.168.112.184: icmp_seq=3 ttl=63 time=1.97 ms\n\n--- 192.168.112.184 ping statistics ---\n3 packets transmitted, 3 received, 0% packet loss, time 2004ms\nrtt min/avg/max/mdev = 1.972/3.721/6.478/1.972 ms\n+ for address in $(openstack --os-cloud test floating ip list --status ACTIVE -f value -c \"Floating IP Address\" | tr -d '\\r')\n+ ping -c3 192.168.112.182\nPING 192.168.112.182 (192.168.112.182) 56(84) bytes of data.\n64 bytes from 192.168.112.182: icmp_seq=1 ttl=63 time=5.39 ms\n64 bytes from 192.168.112.182: icmp_seq=2 ttl=63 time=2.04 ms\n64 bytes from 192.168.112.182: icmp_seq=3 ttl=63 time=1.95 ms\n\n--- 192.168.112.182 ping statistics ---\n3 packets transmitted, 3 received, 0% packet loss, time 2003ms\nrtt min/avg/max/mdev = 1.953/3.127/5.390/1.600 ms\n+ for address in $(openstack --os-cloud test floating ip list --status ACTIVE -f value -c \"Floating IP Address\" | tr -d '\\r')\n+ ping -c3 192.168.112.141\nPING 192.168.112.141 (192.168.112.141) 56(84) bytes of data.\n64 bytes from 192.168.112.141: icmp_seq=1 ttl=63 time=7.32 ms\n64 bytes from 192.168.112.141: icmp_seq=2 ttl=63 time=3.61 ms\n64 bytes from 192.168.112.141: icmp_seq=3 ttl=63 time=2.57 ms\n\n--- 192.168.112.141 ping statistics ---\n3 packets transmitted, 3 received, 0% packet loss, time 2003ms\nrtt min/avg/max/mdev = 2.573/4.503/7.324/2.039 ms\n+ for address in $(openstack --os-cloud test floating ip list --status ACTIVE -f value -c \"Floating IP Address\" | tr -d '\\r')\n+ ping -c3 192.168.112.192\nPING 192.168.112.192 (192.168.112.192) 56(84) bytes of data.\n64 bytes from 192.168.112.192: icmp_seq=1 ttl=63 time=8.24 ms\n64 bytes from 192.168.112.192: icmp_seq=2 ttl=63 time=3.07 ms\n64 bytes from 192.168.112.192: icmp_seq=3 ttl=63 time=1.99 ms\n\n--- 192.168.112.192 ping statistics ---\n3 packets transmitted, 3 received, 0% packet loss, time 2003ms\nrtt min/avg/max/mdev = 1.985/4.433/8.243/2.730 ms",
                            "stdout_lines": [
                                "+ set -e",
                                "+ source /opt/configuration/scripts/include.sh",
                                "++ export INTERACTIVE=false",
                                "++ INTERACTIVE=false",
                                "++ export OSISM_APPLY_RETRY=1",
                                "++ OSISM_APPLY_RETRY=1",
                                "+ source /opt/configuration/scripts/manager-version.sh",
                                "+++ awk '-F: ' '/^manager_version:/ { print $2 }' /opt/configuration/environments/manager/configuration.yml",
                                "",
                                "# CHECK",
                                "",
                                "++ export MANAGER_VERSION=latest",
                                "++ MANAGER_VERSION=latest",
                                "+ echo",
                                "+ echo '# CHECK'",
                                "+ echo",
                                "+ for node in testbed-manager testbed-node-0 testbed-node-1 testbed-node-2",
                                "++ semver latest 5.0.0",
                                "",
                                "## Containers @ testbed-manager",
                                "",
                                "+ [[ -1 -eq -1 ]]",
                                "+ [[ latest != \\l\\a\\t\\e\\s\\t ]]",
                                "+ echo",
                                "+ echo '## Containers @ testbed-manager'",
                                "+ echo",
                                "+ osism container testbed-manager ps",
                                "\u001b[32m2026-02-24 01:22:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mCreating empty known_hosts file: /share/known_hosts\u001b[0m",
                                "CONTAINER ID   IMAGE                                                           COMMAND                  CREATED             STATUS                       PORTS                                                                             NAMES",
                                "dee2ca207554   registry.osism.tech/kolla/prometheus-blackbox-exporter:2024.2   \"dumb-init --single-\u2026\"   12 minutes ago      Up 12 minutes                                                                                                  prometheus_blackbox_exporter",
                                "e96a2bb2a6f6   registry.osism.tech/kolla/prometheus-alertmanager:2024.2        \"dumb-init --single-\u2026\"   12 minutes ago      Up 12 minutes                                                                                                  prometheus_alertmanager",
                                "e7894d6147dc   registry.osism.tech/kolla/prometheus-cadvisor:2024.2            \"dumb-init --single-\u2026\"   13 minutes ago      Up 13 minutes                                                                                                  prometheus_cadvisor",
                                "eb5b37ad51bd   registry.osism.tech/kolla/prometheus-node-exporter:2024.2       \"dumb-init --single-\u2026\"   13 minutes ago      Up 13 minutes                                                                                                  prometheus_node_exporter",
                                "d0d10a5ed6b3   registry.osism.tech/kolla/prometheus-v2-server:2024.2           \"dumb-init --single-\u2026\"   13 minutes ago      Up 13 minutes                                                                                                  prometheus_server",
                                "420b7e422cda   registry.osism.tech/osism/cephclient:reef                       \"/usr/bin/dumb-init \u2026\"   19 minutes ago      Up 19 minutes                                                                                                  cephclient",
                                "0a0ceb50f3d9   registry.osism.tech/kolla/cron:2024.2                           \"dumb-init --single-\u2026\"   32 minutes ago      Up 32 minutes                                                                                                  cron",
                                "b6386485e219   registry.osism.tech/kolla/kolla-toolbox:2024.2                  \"dumb-init --single-\u2026\"   33 minutes ago      Up 32 minutes                                                                                                  kolla_toolbox",
                                "7cfd91f0f86a   registry.osism.tech/kolla/fluentd:2024.2                        \"dumb-init --single-\u2026\"   33 minutes ago      Up 33 minutes                                                                                                  fluentd",
                                "dcad4d685626   phpmyadmin/phpmyadmin:5.2                                       \"/docker-entrypoint.\u2026\"   34 minutes ago      Up 33 minutes (healthy)      80/tcp                                                                            phpmyadmin",
                                "513977b148a1   registry.osism.tech/osism/openstackclient:2024.2                \"/usr/bin/dumb-init \u2026\"   34 minutes ago      Up 34 minutes                                                                                                  openstackclient",
                                "efec17d7241e   registry.osism.tech/osism/homer:v25.10.1                        \"/bin/sh /entrypoint\u2026\"   34 minutes ago      Up 34 minutes (healthy)      8080/tcp                                                                          homer",
                                "0546b63110c2   registry.osism.tech/dockerhub/ubuntu/squid:6.1-23.10_beta       \"entrypoint.sh -f /e\u2026\"   58 minutes ago      Up 57 minutes (healthy)      192.168.16.5:3128->3128/tcp                                                       squid",
                                "d16968ccbd7e   registry.osism.tech/osism/inventory-reconciler:latest           \"/sbin/tini -- /entr\u2026\"   About an hour ago   Up 40 minutes (healthy)                                                                                        manager-inventory_reconciler-1",
                                "125c6d34e4e3   registry.osism.tech/osism/osism-ansible:latest                  \"/entrypoint.sh osis\u2026\"   About an hour ago   Up 41 minutes (healthy)                                                                                        osism-ansible",
                                "7a3c2a4639c8   registry.osism.tech/osism/ceph-ansible:reef                     \"/entrypoint.sh osis\u2026\"   About an hour ago   Up 41 minutes (healthy)                                                                                        ceph-ansible",
                                "04b4496a5f9c   registry.osism.tech/osism/kolla-ansible:2024.2                  \"/entrypoint.sh osis\u2026\"   About an hour ago   Up 41 minutes (healthy)                                                                                        kolla-ansible",
                                "d9037511a0a8   registry.osism.tech/osism/osism-kubernetes:latest               \"/entrypoint.sh osis\u2026\"   About an hour ago   Up 41 minutes (healthy)                                                                                        osism-kubernetes",
                                "3c94059e4b0f   registry.osism.tech/osism/ara-server:1.7.3                      \"sh -c '/wait && /ru\u2026\"   About an hour ago   Up 41 minutes (healthy)      8000/tcp                                                                          manager-ara-server-1",
                                "2de2ca3ef170   registry.osism.tech/dockerhub/library/mariadb:11.8.4            \"docker-entrypoint.s\u2026\"   About an hour ago   Up 41 minutes (healthy)      3306/tcp                                                                          manager-mariadb-1",
                                "13fae06f8b51   registry.osism.tech/osism/osism:latest                          \"/sbin/tini -- osism\u2026\"   About an hour ago   Up 41 minutes (healthy)                                                                                        manager-openstack-1",
                                "515cac366103   registry.osism.tech/osism/osism:latest                          \"/sbin/tini -- osism\u2026\"   About an hour ago   Up 41 minutes (healthy)                                                                                        manager-beat-1",
                                "fadf80797961   registry.osism.tech/osism/osism:latest                          \"/sbin/tini -- osism\u2026\"   About an hour ago   Up 41 minutes (healthy)      192.168.16.5:8000->8000/tcp                                                       manager-api-1",
                                "3c464a2e2250   registry.osism.tech/osism/osism:latest                          \"/sbin/tini -- sleep\u2026\"   About an hour ago   Up 41 minutes (healthy)                                                                                        osismclient",
                                "e3ec25c34f93   registry.osism.tech/dockerhub/library/redis:7.4.7-alpine        \"docker-entrypoint.s\u2026\"   About an hour ago   Up 41 minutes (healthy)      6379/tcp                                                                          manager-redis-1",
                                "86c5161f8563   registry.osism.tech/osism/osism:latest                          \"/sbin/tini -- osism\u2026\"   About an hour ago   Up 41 minutes (healthy)                                                                                        manager-listener-1",
                                "d2b38759b77b   registry.osism.tech/osism/osism:latest                          \"/sbin/tini -- osism\u2026\"   About an hour ago   Up 41 minutes (healthy)                                                                                        manager-flower-1",
                                "4aca83748015   registry.osism.tech/osism/osism-frontend:latest                 \"docker-entrypoint.s\u2026\"   About an hour ago   Up 41 minutes                192.168.16.5:3000->3000/tcp                                                       osism-frontend",
                                "6bb13d8ad793   registry.osism.tech/dockerhub/library/traefik:v3.5.0            \"/entrypoint.sh trae\u2026\"   About an hour ago   Up About an hour (healthy)   192.168.16.5:80->80/tcp, 192.168.16.5:443->443/tcp, 192.168.16.5:8122->8080/tcp   traefik",
                                "",
                                "## Images @ testbed-manager",
                                "",
                                "+ echo",
                                "+ echo '## Images @ testbed-manager'",
                                "+ echo",
                                "+ osism container testbed-manager images",
                                "REPOSITORY                                               TAG              IMAGE ID       CREATED             SIZE",
                                "registry.osism.tech/osism/osism-ansible                  latest           b31da8624ff4   About an hour ago   613MB",
                                "registry.osism.tech/osism/kolla-ansible                  2024.2           c655c028deb8   About an hour ago   610MB",
                                "registry.osism.tech/osism/ceph-ansible                   reef             f45fb23f980e   About an hour ago   559MB",
                                "registry.osism.tech/osism/osism-kubernetes               latest           093bdc689455   About an hour ago   1.22GB",
                                "registry.osism.tech/osism/osism-frontend                 latest           bd2b8e9cfdde   About an hour ago   232MB",
                                "registry.osism.tech/osism/osism                          latest           99863fdf241c   About an hour ago   409MB",
                                "registry.osism.tech/osism/inventory-reconciler           latest           66f25a2df9f3   About an hour ago   335MB",
                                "registry.osism.tech/osism/osism-ansible                  <none>           c6f06501950d   6 hours ago         613MB",
                                "registry.osism.tech/osism/openstackclient                2024.2           3e84edb92a41   21 hours ago        239MB",
                                "registry.osism.tech/osism/cephclient                     reef             ebb02e0df028   22 hours ago        453MB",
                                "registry.osism.tech/kolla/kolla-toolbox                  2024.2           f894a5bc6d0b   24 hours ago        673MB",
                                "registry.osism.tech/kolla/cron                           2024.2           5b921837d137   24 hours ago        271MB",
                                "registry.osism.tech/kolla/fluentd                        2024.2           743bb34b6cee   24 hours ago        584MB",
                                "registry.osism.tech/kolla/prometheus-node-exporter       2024.2           278eaa73d02c   24 hours ago        311MB",
                                "registry.osism.tech/kolla/prometheus-alertmanager        2024.2           23416dcdd72e   24 hours ago        409MB",
                                "registry.osism.tech/kolla/prometheus-blackbox-exporter   2024.2           ffe2e0b2e550   24 hours ago        313MB",
                                "registry.osism.tech/kolla/prometheus-v2-server           2024.2           290ded05e43a   24 hours ago        844MB",
                                "registry.osism.tech/kolla/prometheus-cadvisor            2024.2           4e4cb8e72dfc   24 hours ago        363MB",
                                "registry.osism.tech/dockerhub/library/redis              7.4.7-alpine     e08bd8d5a677   3 weeks ago         41.4MB",
                                "registry.osism.tech/osism/homer                          v25.10.1         ea34b371c716   2 months ago        11.5MB",
                                "registry.osism.tech/dockerhub/library/mariadb            11.8.4           70745dd8f1d0   3 months ago        334MB",
                                "phpmyadmin/phpmyadmin                                    5.2              e66b1f5a8c58   4 months ago        742MB",
                                "registry.osism.tech/osism/ara-server                     1.7.3            d1b687333f2f   6 months ago        275MB",
                                "registry.osism.tech/dockerhub/library/traefik            v3.5.0           11cc59587f6a   7 months ago        226MB",
                                "registry.osism.tech/dockerhub/ubuntu/squid               6.1-23.10_beta   34b6bbbcf74b   20 months ago       146MB",
                                "+ for node in testbed-manager testbed-node-0 testbed-node-1 testbed-node-2",
                                "++ semver latest 5.0.0",
                                "",
                                "## Containers @ testbed-node-0",
                                "",
                                "+ [[ -1 -eq -1 ]]",
                                "+ [[ latest != \\l\\a\\t\\e\\s\\t ]]",
                                "+ echo",
                                "+ echo '## Containers @ testbed-node-0'",
                                "+ echo",
                                "+ osism container testbed-node-0 ps",
                                "CONTAINER ID   IMAGE                                                                COMMAND                  CREATED          STATUS                    PORTS     NAMES",
                                "c446d4dc5ee8   registry.osism.tech/kolla/octavia-worker:2024.2                      \"dumb-init --single-\u2026\"   4 minutes ago    Up 4 minutes (healthy)              octavia_worker",
                                "b5c92ab41dc7   registry.osism.tech/kolla/octavia-housekeeping:2024.2                \"dumb-init --single-\u2026\"   4 minutes ago    Up 4 minutes (healthy)              octavia_housekeeping",
                                "55b4f5aec8dd   registry.osism.tech/kolla/octavia-health-manager:2024.2              \"dumb-init --single-\u2026\"   5 minutes ago    Up 5 minutes (healthy)              octavia_health_manager",
                                "47566925e4d0   registry.osism.tech/kolla/octavia-driver-agent:2024.2                \"dumb-init --single-\u2026\"   5 minutes ago    Up 5 minutes                        octavia_driver_agent",
                                "93adbcd52864   registry.osism.tech/kolla/octavia-api:2024.2                         \"dumb-init --single-\u2026\"   5 minutes ago    Up 5 minutes (healthy)              octavia_api",
                                "fcc8d20bbbb8   registry.osism.tech/kolla/nova-novncproxy:2024.2                     \"dumb-init --single-\u2026\"   8 minutes ago    Up 8 minutes (healthy)              nova_novncproxy",
                                "2e88e7afc50d   registry.osism.tech/kolla/nova-conductor:2024.2                      \"dumb-init --single-\u2026\"   8 minutes ago    Up 8 minutes (healthy)              nova_conductor",
                                "28ee07355de0   registry.osism.tech/kolla/nova-api:2024.2                            \"dumb-init --single-\u2026\"   9 minutes ago    Up 9 minutes (healthy)              nova_api",
                                "2b4f54d58446   registry.osism.tech/kolla/nova-scheduler:2024.2                      \"dumb-init --single-\u2026\"   10 minutes ago   Up 10 minutes (healthy)             nova_scheduler",
                                "e514848de6bb   registry.osism.tech/kolla/grafana:2024.2                             \"dumb-init --single-\u2026\"   10 minutes ago   Up 10 minutes                       grafana",
                                "eeaca894092b   registry.osism.tech/kolla/cinder-backup:2024.2                       \"dumb-init --single-\u2026\"   11 minutes ago   Up 11 minutes (healthy)             cinder_backup",
                                "1702c0db4a81   registry.osism.tech/kolla/cinder-volume:2024.2                       \"dumb-init --single-\u2026\"   11 minutes ago   Up 11 minutes (healthy)             cinder_volume",
                                "15cceafd9c02   registry.osism.tech/kolla/cinder-scheduler:2024.2                    \"dumb-init --single-\u2026\"   12 minutes ago   Up 12 minutes (healthy)             cinder_scheduler",
                                "8dba7f6fd2b6   registry.osism.tech/kolla/glance-api:2024.2                          \"dumb-init --single-\u2026\"   12 minutes ago   Up 12 minutes (healthy)             glance_api",
                                "980ec8d288d0   registry.osism.tech/kolla/cinder-api:2024.2                          \"dumb-init --single-\u2026\"   12 minutes ago   Up 12 minutes (healthy)             cinder_api",
                                "82e10dbf1d99   registry.osism.tech/kolla/prometheus-elasticsearch-exporter:2024.2   \"dumb-init --single-\u2026\"   12 minutes ago   Up 12 minutes                       prometheus_elasticsearch_exporter",
                                "8e3ac52b5f37   registry.osism.tech/kolla/prometheus-cadvisor:2024.2                 \"dumb-init --single-\u2026\"   13 minutes ago   Up 13 minutes                       prometheus_cadvisor",
                                "d3146db51139   registry.osism.tech/kolla/prometheus-memcached-exporter:2024.2       \"dumb-init --single-\u2026\"   13 minutes ago   Up 13 minutes                       prometheus_memcached_exporter",
                                "48bf8f58f2f7   registry.osism.tech/kolla/prometheus-mysqld-exporter:2024.2          \"dumb-init --single-\u2026\"   13 minutes ago   Up 13 minutes                       prometheus_mysqld_exporter",
                                "db99b39df5fa   registry.osism.tech/kolla/prometheus-node-exporter:2024.2            \"dumb-init --single-\u2026\"   13 minutes ago   Up 13 minutes                       prometheus_node_exporter",
                                "6dc45ba06456   registry.osism.tech/kolla/magnum-conductor:2024.2                    \"dumb-init --single-\u2026\"   14 minutes ago   Up 14 minutes (healthy)             magnum_conductor",
                                "46f7079c6b5f   registry.osism.tech/kolla/magnum-api:2024.2                          \"dumb-init --single-\u2026\"   15 minutes ago   Up 15 minutes (healthy)             magnum_api",
                                "331ecd41b313   registry.osism.tech/kolla/neutron-server:2024.2                      \"dumb-init --single-\u2026\"   15 minutes ago   Up 15 minutes (healthy)             neutron_server",
                                "dc28d708b8d5   registry.osism.tech/kolla/designate-worker:2024.2                    \"dumb-init --single-\u2026\"   16 minutes ago   Up 16 minutes (healthy)             designate_worker",
                                "0c627847d85c   registry.osism.tech/kolla/designate-mdns:2024.2                      \"dumb-init --single-\u2026\"   16 minutes ago   Up 16 minutes (healthy)             designate_mdns",
                                "9a1784bcbe3c   registry.osism.tech/kolla/placement-api:2024.2                       \"dumb-init --single-\u2026\"   16 minutes ago   Up 16 minutes (healthy)             placement_api",
                                "2baf5f5c4bde   registry.osism.tech/kolla/designate-producer:2024.2                  \"dumb-init --single-\u2026\"   16 minutes ago   Up 16 minutes (healthy)             designate_producer",
                                "acdad461bfd6   registry.osism.tech/kolla/designate-central:2024.2                   \"dumb-init --single-\u2026\"   17 minutes ago   Up 16 minutes (healthy)             designate_central",
                                "c5a13c72704a   registry.osism.tech/kolla/designate-api:2024.2                       \"dumb-init --single-\u2026\"   17 minutes ago   Up 17 minutes (healthy)             designate_api",
                                "1866693c5939   registry.osism.tech/kolla/barbican-worker:2024.2                     \"dumb-init --single-\u2026\"   17 minutes ago   Up 17 minutes (healthy)             barbican_worker",
                                "d55d06c19e8f   registry.osism.tech/kolla/barbican-keystone-listener:2024.2          \"dumb-init --single-\u2026\"   17 minutes ago   Up 17 minutes (healthy)             barbican_keystone_listener",
                                "3babadf8056e   registry.osism.tech/kolla/designate-backend-bind9:2024.2             \"dumb-init --single-\u2026\"   17 minutes ago   Up 17 minutes (healthy)             designate_backend_bind9",
                                "e5ae1024a5a0   registry.osism.tech/kolla/barbican-api:2024.2                        \"dumb-init --single-\u2026\"   17 minutes ago   Up 17 minutes (healthy)             barbican_api",
                                "edd213c2b417   registry.osism.tech/osism/ceph-daemon:reef                           \"/usr/bin/ceph-mgr -\u2026\"   18 minutes ago   Up 18 minutes                       ceph-mgr-testbed-node-0",
                                "3616d3cf4fc2   registry.osism.tech/kolla/keystone:2024.2                            \"dumb-init --single-\u2026\"   21 minutes ago   Up 21 minutes (healthy)             keystone",
                                "6ee631dff16c   registry.osism.tech/kolla/keystone-fernet:2024.2                     \"dumb-init --single-\u2026\"   21 minutes ago   Up 21 minutes (healthy)             keystone_fernet",
                                "dafeea7513bf   registry.osism.tech/kolla/keystone-ssh:2024.2                        \"dumb-init --single-\u2026\"   21 minutes ago   Up 21 minutes (healthy)             keystone_ssh",
                                "08587638fa20   registry.osism.tech/kolla/horizon:2024.2                             \"dumb-init --single-\u2026\"   21 minutes ago   Up 21 minutes (healthy)             horizon",
                                "0f4762768487   registry.osism.tech/kolla/mariadb-server:2024.2                      \"dumb-init -- kolla_\u2026\"   23 minutes ago   Up 23 minutes (healthy)             mariadb",
                                "341fc807413d   registry.osism.tech/kolla/opensearch-dashboards:2024.2               \"dumb-init --single-\u2026\"   24 minutes ago   Up 24 minutes (healthy)             opensearch_dashboards",
                                "9e484d669985   registry.osism.tech/kolla/opensearch:2024.2                          \"dumb-init --single-\u2026\"   25 minutes ago   Up 25 minutes (healthy)             opensearch",
                                "4aad693a4e67   registry.osism.tech/osism/ceph-daemon:reef                           \"/usr/bin/ceph-crash\"    25 minutes ago   Up 25 minutes                       ceph-crash-testbed-node-0",
                                "12a27e2732ed   registry.osism.tech/kolla/keepalived:2024.2                          \"dumb-init --single-\u2026\"   26 minutes ago   Up 26 minutes                       keepalived",
                                "feeddafe9200   registry.osism.tech/kolla/proxysql:2024.2                            \"dumb-init --single-\u2026\"   26 minutes ago   Up 26 minutes (healthy)             proxysql",
                                "ae60ab2ca74f   registry.osism.tech/kolla/haproxy:2024.2                             \"dumb-init --single-\u2026\"   26 minutes ago   Up 26 minutes (healthy)             haproxy",
                                "0bf73eb1bc2a   registry.osism.tech/kolla/ovn-northd:2024.2                          \"dumb-init --single-\u2026\"   30 minutes ago   Up 30 minutes                       ovn_northd",
                                "9c61d16c643d   registry.osism.tech/kolla/ovn-sb-db-server:2024.2                    \"dumb-init --single-\u2026\"   30 minutes ago   Up 30 minutes                       ovn_sb_db",
                                "17111b2e006e   registry.osism.tech/kolla/ovn-nb-db-server:2024.2                    \"dumb-init --single-\u2026\"   30 minutes ago   Up 30 minutes                       ovn_nb_db",
                                "f43c50bae847   registry.osism.tech/kolla/ovn-controller:2024.2                      \"dumb-init --single-\u2026\"   31 minutes ago   Up 31 minutes                       ovn_controller",
                                "1cb145726d67   registry.osism.tech/osism/ceph-daemon:reef                           \"/usr/bin/ceph-mon -\u2026\"   31 minutes ago   Up 31 minutes                       ceph-mon-testbed-node-0",
                                "99bb95cd6961   registry.osism.tech/kolla/rabbitmq:2024.2                            \"dumb-init --single-\u2026\"   31 minutes ago   Up 31 minutes (healthy)             rabbitmq",
                                "77f4d5303968   registry.osism.tech/kolla/openvswitch-vswitchd:2024.2                \"dumb-init --single-\u2026\"   32 minutes ago   Up 31 minutes (healthy)             openvswitch_vswitchd",
                                "a9f9d1ebe587   registry.osism.tech/kolla/redis-sentinel:2024.2                      \"dumb-init --single-\u2026\"   32 minutes ago   Up 32 minutes (healthy)             redis_sentinel",
                                "14dbe26fb9bf   registry.osism.tech/kolla/openvswitch-db-server:2024.2               \"dumb-init --single-\u2026\"   32 minutes ago   Up 32 minutes (healthy)             openvswitch_db",
                                "e40fd89e362d   registry.osism.tech/kolla/redis:2024.2                               \"dumb-init --single-\u2026\"   32 minutes ago   Up 32 minutes (healthy)             redis",
                                "96f285897c5e   registry.osism.tech/kolla/memcached:2024.2                           \"dumb-init --single-\u2026\"   32 minutes ago   Up 32 minutes (healthy)             memcached",
                                "2236df7422f3   registry.osism.tech/kolla/cron:2024.2                                \"dumb-init --single-\u2026\"   33 minutes ago   Up 33 minutes                       cron",
                                "879bca2009ad   registry.osism.tech/kolla/kolla-toolbox:2024.2                       \"dumb-init --single-\u2026\"   33 minutes ago   Up 33 minutes                       kolla_toolbox",
                                "7101ed2ee4e0   registry.osism.tech/kolla/fluentd:2024.2                             \"dumb-init --single-\u2026\"   34 minutes ago   Up 34 minutes                       fluentd",
                                "",
                                "## Images @ testbed-node-0",
                                "",
                                "+ echo",
                                "+ echo '## Images @ testbed-node-0'",
                                "+ echo",
                                "+ osism container testbed-node-0 images",
                                "REPOSITORY                                                    TAG       IMAGE ID       CREATED        SIZE",
                                "registry.osism.tech/osism/ceph-daemon                         reef      ff59b29334dd   22 hours ago   1.27GB",
                                "registry.osism.tech/kolla/memcached                           2024.2    3652d3a7de56   24 hours ago   272MB",
                                "registry.osism.tech/kolla/proxysql                            2024.2    dbe28c2fd6c3   24 hours ago   418MB",
                                "registry.osism.tech/kolla/kolla-toolbox                       2024.2    f894a5bc6d0b   24 hours ago   673MB",
                                "registry.osism.tech/kolla/haproxy                             2024.2    b1fd4c643197   24 hours ago   279MB",
                                "registry.osism.tech/kolla/cron                                2024.2    5b921837d137   24 hours ago   271MB",
                                "registry.osism.tech/kolla/rabbitmq                            2024.2    540f6409fbf5   24 hours ago   328MB",
                                "registry.osism.tech/kolla/keepalived                          2024.2    13679e653294   24 hours ago   282MB",
                                "registry.osism.tech/kolla/grafana                             2024.2    37fbf97e7f6f   24 hours ago   1.02GB",
                                "registry.osism.tech/kolla/fluentd                             2024.2    743bb34b6cee   24 hours ago   584MB",
                                "registry.osism.tech/kolla/opensearch-dashboards               2024.2    c159d79f20f0   24 hours ago   1.53GB",
                                "registry.osism.tech/kolla/opensearch                          2024.2    80d69bb5f098   24 hours ago   1.56GB",
                                "registry.osism.tech/kolla/openvswitch-db-server               2024.2    7e08fe973c8e   24 hours ago   284MB",
                                "registry.osism.tech/kolla/openvswitch-vswitchd                2024.2    4d41d184aa52   24 hours ago   284MB",
                                "registry.osism.tech/kolla/redis                               2024.2    0fa4d8b23966   24 hours ago   278MB",
                                "registry.osism.tech/kolla/redis-sentinel                      2024.2    2052a297db29   24 hours ago   278MB",
                                "registry.osism.tech/kolla/horizon                             2024.2    4d92e9384e36   24 hours ago   1.15GB",
                                "registry.osism.tech/kolla/mariadb-server                      2024.2    5dbd7b932e63   24 hours ago   457MB",
                                "registry.osism.tech/kolla/prometheus-node-exporter            2024.2    278eaa73d02c   24 hours ago   311MB",
                                "registry.osism.tech/kolla/prometheus-elasticsearch-exporter   2024.2    13638ae36e57   24 hours ago   297MB",
                                "registry.osism.tech/kolla/prometheus-memcached-exporter       2024.2    8ff5a6b3cd16   24 hours ago   304MB",
                                "registry.osism.tech/kolla/prometheus-mysqld-exporter          2024.2    d9cb73d898bb   24 hours ago   306MB",
                                "registry.osism.tech/kolla/prometheus-cadvisor                 2024.2    4e4cb8e72dfc   24 hours ago   363MB",
                                "registry.osism.tech/kolla/ovn-nb-db-server                    2024.2    51d6ef378bbd   24 hours ago   846MB",
                                "registry.osism.tech/kolla/ovn-sb-db-server                    2024.2    25bb84f2289b   24 hours ago   846MB",
                                "registry.osism.tech/kolla/ovn-northd                          2024.2    47dae0f63bd5   24 hours ago   846MB",
                                "registry.osism.tech/kolla/ovn-controller                      2024.2    342e8da8ac3a   24 hours ago   846MB",
                                "registry.osism.tech/kolla/neutron-server                      2024.2    189daad12150   24 hours ago   1.17GB",
                                "registry.osism.tech/kolla/octavia-api                         2024.2    624955c2c115   24 hours ago   1.06GB",
                                "registry.osism.tech/kolla/octavia-housekeeping                2024.2    9b80c75ebf88   24 hours ago   1.03GB",
                                "registry.osism.tech/kolla/octavia-driver-agent                2024.2    3f4a0924d9ed   24 hours ago   1.06GB",
                                "registry.osism.tech/kolla/octavia-health-manager              2024.2    741914d697fa   24 hours ago   1.03GB",
                                "registry.osism.tech/kolla/octavia-worker                      2024.2    45d576cfc14a   24 hours ago   1.03GB",
                                "registry.osism.tech/kolla/skyline-apiserver                   2024.2    1f7f7078acd6   24 hours ago   995MB",
                                "registry.osism.tech/kolla/skyline-console                     2024.2    d6dc835952f4   24 hours ago   1.05GB",
                                "registry.osism.tech/kolla/cinder-backup                       2024.2    db0ba08b76e7   24 hours ago   1.42GB",
                                "registry.osism.tech/kolla/cinder-volume                       2024.2    aad4ba476f1c   24 hours ago   1.72GB",
                                "registry.osism.tech/kolla/cinder-scheduler                    2024.2    3102035cbd3f   24 hours ago   1.41GB",
                                "registry.osism.tech/kolla/cinder-api                          2024.2    ffa30f3933ed   24 hours ago   1.41GB",
                                "registry.osism.tech/kolla/placement-api                       2024.2    e4b48b9cb79e   24 hours ago   981MB",
                                "registry.osism.tech/kolla/designate-producer                  2024.2    e3b7812cd6ff   24 hours ago   990MB",
                                "registry.osism.tech/kolla/designate-central                   2024.2    8ab2ea11494a   24 hours ago   989MB",
                                "registry.osism.tech/kolla/designate-worker                    2024.2    13bd17ae9e62   24 hours ago   994MB",
                                "registry.osism.tech/kolla/designate-api                       2024.2    2e44006b8d8c   24 hours ago   990MB",
                                "registry.osism.tech/kolla/designate-backend-bind9             2024.2    7b60bbae1ff0   24 hours ago   994MB",
                                "registry.osism.tech/kolla/designate-mdns                      2024.2    a28e142a95bf   24 hours ago   990MB",
                                "registry.osism.tech/kolla/barbican-worker                     2024.2    0eec7e0f7b33   24 hours ago   996MB",
                                "registry.osism.tech/kolla/barbican-keystone-listener          2024.2    8c429b2daef4   24 hours ago   996MB",
                                "registry.osism.tech/kolla/barbican-api                        2024.2    9d4b161f78ee   24 hours ago   996MB",
                                "registry.osism.tech/kolla/keystone-fernet                     2024.2    3770c4e50a62   24 hours ago   1.04GB",
                                "registry.osism.tech/kolla/keystone-ssh                        2024.2    a75dcc713ad6   24 hours ago   1.05GB",
                                "registry.osism.tech/kolla/keystone                            2024.2    f3afb1708fa9   24 hours ago   1.07GB",
                                "registry.osism.tech/kolla/nova-conductor                      2024.2    7e0dfe6dfac8   24 hours ago   1.22GB",
                                "registry.osism.tech/kolla/nova-api                            2024.2    b141835a8ce3   24 hours ago   1.22GB",
                                "registry.osism.tech/kolla/nova-novncproxy                     2024.2    0e230dfe273c   24 hours ago   1.37GB",
                                "registry.osism.tech/kolla/nova-scheduler                      2024.2    1d5f182c3a75   24 hours ago   1.22GB",
                                "registry.osism.tech/kolla/aodh-listener                       2024.2    1d3d06d0cc72   24 hours ago   979MB",
                                "registry.osism.tech/kolla/aodh-api                            2024.2    d7ef40f551b0   24 hours ago   979MB",
                                "registry.osism.tech/kolla/aodh-notifier                       2024.2    826ad4a5837c   24 hours ago   979MB",
                                "registry.osism.tech/kolla/aodh-evaluator                      2024.2    9fa396903596   24 hours ago   979MB",
                                "registry.osism.tech/kolla/ceilometer-notification             2024.2    2d0736f4b996   24 hours ago   981MB",
                                "registry.osism.tech/kolla/ceilometer-central                  2024.2    112e904da6ac   24 hours ago   982MB",
                                "registry.osism.tech/kolla/magnum-api                          2024.2    5f8755d26e30   24 hours ago   1.13GB",
                                "registry.osism.tech/kolla/magnum-conductor                    2024.2    ad0e7877ae1a   24 hours ago   1.25GB",
                                "registry.osism.tech/kolla/glance-api                          2024.2    4e1be079f392   24 hours ago   1.1GB",
                                "+ for node in testbed-manager testbed-node-0 testbed-node-1 testbed-node-2",
                                "++ semver latest 5.0.0",
                                "",
                                "## Containers @ testbed-node-1",
                                "",
                                "+ [[ -1 -eq -1 ]]",
                                "+ [[ latest != \\l\\a\\t\\e\\s\\t ]]",
                                "+ echo",
                                "+ echo '## Containers @ testbed-node-1'",
                                "+ echo",
                                "+ osism container testbed-node-1 ps",
                                "CONTAINER ID   IMAGE                                                                COMMAND                  CREATED          STATUS                    PORTS     NAMES",
                                "443c3105e0a7   registry.osism.tech/kolla/octavia-worker:2024.2                      \"dumb-init --single-\u2026\"   4 minutes ago    Up 4 minutes (healthy)              octavia_worker",
                                "20d02ff22874   registry.osism.tech/kolla/octavia-housekeeping:2024.2                \"dumb-init --single-\u2026\"   4 minutes ago    Up 4 minutes (healthy)              octavia_housekeeping",
                                "b063938944b8   registry.osism.tech/kolla/octavia-health-manager:2024.2              \"dumb-init --single-\u2026\"   5 minutes ago    Up 5 minutes (healthy)              octavia_health_manager",
                                "715117e226ad   registry.osism.tech/kolla/octavia-driver-agent:2024.2                \"dumb-init --single-\u2026\"   5 minutes ago    Up 5 minutes                        octavia_driver_agent",
                                "1fe566592e1b   registry.osism.tech/kolla/octavia-api:2024.2                         \"dumb-init --single-\u2026\"   5 minutes ago    Up 5 minutes (healthy)              octavia_api",
                                "6f2a47747067   registry.osism.tech/kolla/nova-novncproxy:2024.2                     \"dumb-init --single-\u2026\"   8 minutes ago    Up 8 minutes (healthy)              nova_novncproxy",
                                "db660a20e692   registry.osism.tech/kolla/nova-conductor:2024.2                      \"dumb-init --single-\u2026\"   8 minutes ago    Up 8 minutes (healthy)              nova_conductor",
                                "240b47367be8   registry.osism.tech/kolla/grafana:2024.2                             \"dumb-init --single-\u2026\"   10 minutes ago   Up 10 minutes                       grafana",
                                "34998db525ca   registry.osism.tech/kolla/nova-api:2024.2                            \"dumb-init --single-\u2026\"   10 minutes ago   Up 10 minutes (healthy)             nova_api",
                                "1de9015b70ff   registry.osism.tech/kolla/nova-scheduler:2024.2                      \"dumb-init --single-\u2026\"   10 minutes ago   Up 10 minutes (healthy)             nova_scheduler",
                                "6b89e77fcb73   registry.osism.tech/kolla/cinder-backup:2024.2                       \"dumb-init --single-\u2026\"   11 minutes ago   Up 11 minutes (healthy)             cinder_backup",
                                "83a200208cd2   registry.osism.tech/kolla/cinder-volume:2024.2                       \"dumb-init --single-\u2026\"   11 minutes ago   Up 11 minutes (healthy)             cinder_volume",
                                "87ba4474b65d   registry.osism.tech/kolla/glance-api:2024.2                          \"dumb-init --single-\u2026\"   12 minutes ago   Up 12 minutes (healthy)             glance_api",
                                "419a24ce1d7c   registry.osism.tech/kolla/cinder-scheduler:2024.2                    \"dumb-init --single-\u2026\"   12 minutes ago   Up 12 minutes (healthy)             cinder_scheduler",
                                "8641a24ebad2   registry.osism.tech/kolla/cinder-api:2024.2                          \"dumb-init --single-\u2026\"   12 minutes ago   Up 12 minutes (healthy)             cinder_api",
                                "982a8d255f2d   registry.osism.tech/kolla/prometheus-elasticsearch-exporter:2024.2   \"dumb-init --single-\u2026\"   13 minutes ago   Up 13 minutes                       prometheus_elasticsearch_exporter",
                                "7202121cd714   registry.osism.tech/kolla/prometheus-cadvisor:2024.2                 \"dumb-init --single-\u2026\"   13 minutes ago   Up 13 minutes                       prometheus_cadvisor",
                                "9b82bdefbd04   registry.osism.tech/kolla/prometheus-memcached-exporter:2024.2       \"dumb-init --single-\u2026\"   13 minutes ago   Up 13 minutes                       prometheus_memcached_exporter",
                                "7f03efd478d8   registry.osism.tech/kolla/prometheus-mysqld-exporter:2024.2          \"dumb-init --single-\u2026\"   13 minutes ago   Up 13 minutes                       prometheus_mysqld_exporter",
                                "5cf1ef52b333   registry.osism.tech/kolla/prometheus-node-exporter:2024.2            \"dumb-init --single-\u2026\"   13 minutes ago   Up 13 minutes                       prometheus_node_exporter",
                                "d05b983ca5cd   registry.osism.tech/kolla/magnum-conductor:2024.2                    \"dumb-init --single-\u2026\"   15 minutes ago   Up 15 minutes (healthy)             magnum_conductor",
                                "2bc61d574472   registry.osism.tech/kolla/magnum-api:2024.2                          \"dumb-init --single-\u2026\"   15 minutes ago   Up 15 minutes (healthy)             magnum_api",
                                "a12bdf928dd3   registry.osism.tech/kolla/neutron-server:2024.2                      \"dumb-init --single-\u2026\"   15 minutes ago   Up 15 minutes (healthy)             neutron_server",
                                "696c642518c4   registry.osism.tech/kolla/designate-worker:2024.2                    \"dumb-init --single-\u2026\"   16 minutes ago   Up 16 minutes (healthy)             designate_worker",
                                "3b21b349fbde   registry.osism.tech/kolla/placement-api:2024.2                       \"dumb-init --single-\u2026\"   16 minutes ago   Up 16 minutes (healthy)             placement_api",
                                "274437c23da9   registry.osism.tech/kolla/designate-mdns:2024.2                      \"dumb-init --single-\u2026\"   16 minutes ago   Up 16 minutes (healthy)             designate_mdns",
                                "76ab2f193443   registry.osism.tech/kolla/designate-producer:2024.2                  \"dumb-init --single-\u2026\"   16 minutes ago   Up 16 minutes (healthy)             designate_producer",
                                "5c5d7a49f4bd   registry.osism.tech/kolla/designate-central:2024.2                   \"dumb-init --single-\u2026\"   17 minutes ago   Up 17 minutes (healthy)             designate_central",
                                "d77a29bb1e26   registry.osism.tech/kolla/designate-api:2024.2                       \"dumb-init --single-\u2026\"   17 minutes ago   Up 17 minutes (healthy)             designate_api",
                                "fc3dd7be7471   registry.osism.tech/kolla/barbican-worker:2024.2                     \"dumb-init --single-\u2026\"   17 minutes ago   Up 17 minutes (healthy)             barbican_worker",
                                "336a8e22c323   registry.osism.tech/kolla/barbican-keystone-listener:2024.2          \"dumb-init --single-\u2026\"   17 minutes ago   Up 17 minutes (healthy)             barbican_keystone_listener",
                                "b3eca58eb0ff   registry.osism.tech/kolla/designate-backend-bind9:2024.2             \"dumb-init --single-\u2026\"   17 minutes ago   Up 17 minutes (healthy)             designate_backend_bind9",
                                "b19004b05880   registry.osism.tech/kolla/barbican-api:2024.2                        \"dumb-init --single-\u2026\"   17 minutes ago   Up 17 minutes (healthy)             barbican_api",
                                "3534c242e3d3   registry.osism.tech/osism/ceph-daemon:reef                           \"/usr/bin/ceph-mgr -\u2026\"   18 minutes ago   Up 18 minutes                       ceph-mgr-testbed-node-1",
                                "537e4ace13ae   registry.osism.tech/kolla/keystone:2024.2                            \"dumb-init --single-\u2026\"   21 minutes ago   Up 21 minutes (healthy)             keystone",
                                "3e420c4d8d47   registry.osism.tech/kolla/horizon:2024.2                             \"dumb-init --single-\u2026\"   21 minutes ago   Up 21 minutes (healthy)             horizon",
                                "3f909ec0fd73   registry.osism.tech/kolla/keystone-fernet:2024.2                     \"dumb-init --single-\u2026\"   21 minutes ago   Up 21 minutes (healthy)             keystone_fernet",
                                "bed5bcc8f080   registry.osism.tech/kolla/keystone-ssh:2024.2                        \"dumb-init --single-\u2026\"   21 minutes ago   Up 21 minutes (healthy)             keystone_ssh",
                                "eb6cc91c9aeb   registry.osism.tech/kolla/opensearch-dashboards:2024.2               \"dumb-init --single-\u2026\"   23 minutes ago   Up 23 minutes (healthy)             opensearch_dashboards",
                                "84651137feeb   registry.osism.tech/kolla/mariadb-server:2024.2                      \"dumb-init -- kolla_\u2026\"   24 minutes ago   Up 24 minutes (healthy)             mariadb",
                                "479f29fcf95d   registry.osism.tech/kolla/opensearch:2024.2                          \"dumb-init --single-\u2026\"   24 minutes ago   Up 24 minutes (healthy)             opensearch",
                                "2d3469b3e3ff   registry.osism.tech/osism/ceph-daemon:reef                           \"/usr/bin/ceph-crash\"    26 minutes ago   Up 26 minutes                       ceph-crash-testbed-node-1",
                                "c19fe3d26618   registry.osism.tech/kolla/keepalived:2024.2                          \"dumb-init --single-\u2026\"   26 minutes ago   Up 26 minutes                       keepalived",
                                "6204d745fe4b   registry.osism.tech/kolla/proxysql:2024.2                            \"dumb-init --single-\u2026\"   26 minutes ago   Up 26 minutes (healthy)             proxysql",
                                "80e9102d54d3   registry.osism.tech/kolla/haproxy:2024.2                             \"dumb-init --single-\u2026\"   27 minutes ago   Up 27 minutes (healthy)             haproxy",
                                "784e37f7ab24   registry.osism.tech/kolla/ovn-northd:2024.2                          \"dumb-init --single-\u2026\"   30 minutes ago   Up 29 minutes                       ovn_northd",
                                "23c769291118   registry.osism.tech/kolla/ovn-sb-db-server:2024.2                    \"dumb-init --single-\u2026\"   30 minutes ago   Up 29 minutes                       ovn_sb_db",
                                "575c84596223   registry.osism.tech/kolla/ovn-nb-db-server:2024.2                    \"dumb-init --single-\u2026\"   30 minutes ago   Up 29 minutes                       ovn_nb_db",
                                "82884e1e4587   registry.osism.tech/kolla/ovn-controller:2024.2                      \"dumb-init --single-\u2026\"   30 minutes ago   Up 30 minutes                       ovn_controller",
                                "7264d68f426f   registry.osism.tech/kolla/rabbitmq:2024.2                            \"dumb-init --single-\u2026\"   30 minutes ago   Up 30 minutes (healthy)             rabbitmq",
                                "65c197c778ba   registry.osism.tech/osism/ceph-daemon:reef                           \"/usr/bin/ceph-mon -\u2026\"   31 minutes ago   Up 31 minutes                       ceph-mon-testbed-node-1",
                                "b010803b2bf5   registry.osism.tech/kolla/openvswitch-vswitchd:2024.2                \"dumb-init --single-\u2026\"   32 minutes ago   Up 31 minutes (healthy)             openvswitch_vswitchd",
                                "e7259e9ebf53   registry.osism.tech/kolla/redis-sentinel:2024.2                      \"dumb-init --single-\u2026\"   32 minutes ago   Up 32 minutes (healthy)             redis_sentinel",
                                "d7292eec7f67   registry.osism.tech/kolla/openvswitch-db-server:2024.2               \"dumb-init --single-\u2026\"   32 minutes ago   Up 32 minutes (healthy)             openvswitch_db",
                                "7d769c827d50   registry.osism.tech/kolla/redis:2024.2                               \"dumb-init --single-\u2026\"   32 minutes ago   Up 32 minutes (healthy)             redis",
                                "7f70aeb5717e   registry.osism.tech/kolla/memcached:2024.2                           \"dumb-init --single-\u2026\"   32 minutes ago   Up 32 minutes (healthy)             memcached",
                                "1d9ed2d48c5a   registry.osism.tech/kolla/cron:2024.2                                \"dumb-init --single-\u2026\"   33 minutes ago   Up 33 minutes                       cron",
                                "e2a61650650f   registry.osism.tech/kolla/kolla-toolbox:2024.2                       \"dumb-init --single-\u2026\"   33 minutes ago   Up 33 minutes                       kolla_toolbox",
                                "baedf1310f6d   registry.osism.tech/kolla/fluentd:2024.2                             \"dumb-init --single-\u2026\"   33 minutes ago   Up 33 minutes                       fluentd",
                                "",
                                "## Images @ testbed-node-1",
                                "",
                                "+ echo",
                                "+ echo '## Images @ testbed-node-1'",
                                "+ echo",
                                "+ osism container testbed-node-1 images",
                                "REPOSITORY                                                    TAG       IMAGE ID       CREATED        SIZE",
                                "registry.osism.tech/osism/ceph-daemon                         reef      ff59b29334dd   22 hours ago   1.27GB",
                                "registry.osism.tech/kolla/memcached                           2024.2    3652d3a7de56   24 hours ago   272MB",
                                "registry.osism.tech/kolla/proxysql                            2024.2    dbe28c2fd6c3   24 hours ago   418MB",
                                "registry.osism.tech/kolla/kolla-toolbox                       2024.2    f894a5bc6d0b   24 hours ago   673MB",
                                "registry.osism.tech/kolla/haproxy                             2024.2    b1fd4c643197   24 hours ago   279MB",
                                "registry.osism.tech/kolla/cron                                2024.2    5b921837d137   24 hours ago   271MB",
                                "registry.osism.tech/kolla/rabbitmq                            2024.2    540f6409fbf5   24 hours ago   328MB",
                                "registry.osism.tech/kolla/keepalived                          2024.2    13679e653294   24 hours ago   282MB",
                                "registry.osism.tech/kolla/grafana                             2024.2    37fbf97e7f6f   24 hours ago   1.02GB",
                                "registry.osism.tech/kolla/fluentd                             2024.2    743bb34b6cee   24 hours ago   584MB",
                                "registry.osism.tech/kolla/opensearch-dashboards               2024.2    c159d79f20f0   24 hours ago   1.53GB",
                                "registry.osism.tech/kolla/opensearch                          2024.2    80d69bb5f098   24 hours ago   1.56GB",
                                "registry.osism.tech/kolla/openvswitch-vswitchd                2024.2    4d41d184aa52   24 hours ago   284MB",
                                "registry.osism.tech/kolla/openvswitch-db-server               2024.2    7e08fe973c8e   24 hours ago   284MB",
                                "registry.osism.tech/kolla/redis                               2024.2    0fa4d8b23966   24 hours ago   278MB",
                                "registry.osism.tech/kolla/redis-sentinel                      2024.2    2052a297db29   24 hours ago   278MB",
                                "registry.osism.tech/kolla/horizon                             2024.2    4d92e9384e36   24 hours ago   1.15GB",
                                "registry.osism.tech/kolla/mariadb-server                      2024.2    5dbd7b932e63   24 hours ago   457MB",
                                "registry.osism.tech/kolla/prometheus-node-exporter            2024.2    278eaa73d02c   24 hours ago   311MB",
                                "registry.osism.tech/kolla/prometheus-elasticsearch-exporter   2024.2    13638ae36e57   24 hours ago   297MB",
                                "registry.osism.tech/kolla/prometheus-memcached-exporter       2024.2    8ff5a6b3cd16   24 hours ago   304MB",
                                "registry.osism.tech/kolla/prometheus-mysqld-exporter          2024.2    d9cb73d898bb   24 hours ago   306MB",
                                "registry.osism.tech/kolla/prometheus-cadvisor                 2024.2    4e4cb8e72dfc   24 hours ago   363MB",
                                "registry.osism.tech/kolla/ovn-nb-db-server                    2024.2    51d6ef378bbd   24 hours ago   846MB",
                                "registry.osism.tech/kolla/ovn-sb-db-server                    2024.2    25bb84f2289b   24 hours ago   846MB",
                                "registry.osism.tech/kolla/ovn-controller                      2024.2    342e8da8ac3a   24 hours ago   846MB",
                                "registry.osism.tech/kolla/ovn-northd                          2024.2    47dae0f63bd5   24 hours ago   846MB",
                                "registry.osism.tech/kolla/neutron-server                      2024.2    189daad12150   24 hours ago   1.17GB",
                                "registry.osism.tech/kolla/octavia-api                         2024.2    624955c2c115   24 hours ago   1.06GB",
                                "registry.osism.tech/kolla/octavia-housekeeping                2024.2    9b80c75ebf88   24 hours ago   1.03GB",
                                "registry.osism.tech/kolla/octavia-driver-agent                2024.2    3f4a0924d9ed   24 hours ago   1.06GB",
                                "registry.osism.tech/kolla/octavia-health-manager              2024.2    741914d697fa   24 hours ago   1.03GB",
                                "registry.osism.tech/kolla/octavia-worker                      2024.2    45d576cfc14a   24 hours ago   1.03GB",
                                "registry.osism.tech/kolla/cinder-backup                       2024.2    db0ba08b76e7   24 hours ago   1.42GB",
                                "registry.osism.tech/kolla/cinder-volume                       2024.2    aad4ba476f1c   24 hours ago   1.72GB",
                                "registry.osism.tech/kolla/cinder-scheduler                    2024.2    3102035cbd3f   24 hours ago   1.41GB",
                                "registry.osism.tech/kolla/cinder-api                          2024.2    ffa30f3933ed   24 hours ago   1.41GB",
                                "registry.osism.tech/kolla/placement-api                       2024.2    e4b48b9cb79e   24 hours ago   981MB",
                                "registry.osism.tech/kolla/designate-producer                  2024.2    e3b7812cd6ff   24 hours ago   990MB",
                                "registry.osism.tech/kolla/designate-central                   2024.2    8ab2ea11494a   24 hours ago   989MB",
                                "registry.osism.tech/kolla/designate-worker                    2024.2    13bd17ae9e62   24 hours ago   994MB",
                                "registry.osism.tech/kolla/designate-api                       2024.2    2e44006b8d8c   24 hours ago   990MB",
                                "registry.osism.tech/kolla/designate-backend-bind9             2024.2    7b60bbae1ff0   24 hours ago   994MB",
                                "registry.osism.tech/kolla/designate-mdns                      2024.2    a28e142a95bf   24 hours ago   990MB",
                                "registry.osism.tech/kolla/barbican-worker                     2024.2    0eec7e0f7b33   24 hours ago   996MB",
                                "registry.osism.tech/kolla/barbican-keystone-listener          2024.2    8c429b2daef4   24 hours ago   996MB",
                                "registry.osism.tech/kolla/barbican-api                        2024.2    9d4b161f78ee   24 hours ago   996MB",
                                "registry.osism.tech/kolla/keystone-fernet                     2024.2    3770c4e50a62   24 hours ago   1.04GB",
                                "registry.osism.tech/kolla/keystone-ssh                        2024.2    a75dcc713ad6   24 hours ago   1.05GB",
                                "registry.osism.tech/kolla/keystone                            2024.2    f3afb1708fa9   24 hours ago   1.07GB",
                                "registry.osism.tech/kolla/nova-conductor                      2024.2    7e0dfe6dfac8   24 hours ago   1.22GB",
                                "registry.osism.tech/kolla/nova-api                            2024.2    b141835a8ce3   24 hours ago   1.22GB",
                                "registry.osism.tech/kolla/nova-novncproxy                     2024.2    0e230dfe273c   24 hours ago   1.37GB",
                                "registry.osism.tech/kolla/nova-scheduler                      2024.2    1d5f182c3a75   24 hours ago   1.22GB",
                                "registry.osism.tech/kolla/magnum-api                          2024.2    5f8755d26e30   24 hours ago   1.13GB",
                                "registry.osism.tech/kolla/magnum-conductor                    2024.2    ad0e7877ae1a   24 hours ago   1.25GB",
                                "registry.osism.tech/kolla/glance-api                          2024.2    4e1be079f392   24 hours ago   1.1GB",
                                "+ for node in testbed-manager testbed-node-0 testbed-node-1 testbed-node-2",
                                "++ semver latest 5.0.0",
                                "",
                                "## Containers @ testbed-node-2",
                                "",
                                "+ [[ -1 -eq -1 ]]",
                                "+ [[ latest != \\l\\a\\t\\e\\s\\t ]]",
                                "+ echo",
                                "+ echo '## Containers @ testbed-node-2'",
                                "+ echo",
                                "+ osism container testbed-node-2 ps",
                                "CONTAINER ID   IMAGE                                                                COMMAND                  CREATED          STATUS                    PORTS     NAMES",
                                "204b517fdebb   registry.osism.tech/kolla/octavia-worker:2024.2                      \"dumb-init --single-\u2026\"   4 minutes ago    Up 4 minutes (healthy)              octavia_worker",
                                "36a8f6d80767   registry.osism.tech/kolla/octavia-housekeeping:2024.2                \"dumb-init --single-\u2026\"   5 minutes ago    Up 5 minutes (healthy)              octavia_housekeeping",
                                "a0f8119d1238   registry.osism.tech/kolla/octavia-health-manager:2024.2              \"dumb-init --single-\u2026\"   5 minutes ago    Up 5 minutes (healthy)              octavia_health_manager",
                                "90709743ce68   registry.osism.tech/kolla/octavia-driver-agent:2024.2                \"dumb-init --single-\u2026\"   5 minutes ago    Up 5 minutes                        octavia_driver_agent",
                                "a3dccaadc734   registry.osism.tech/kolla/octavia-api:2024.2                         \"dumb-init --single-\u2026\"   5 minutes ago    Up 5 minutes (healthy)              octavia_api",
                                "f2dcf01d1373   registry.osism.tech/kolla/nova-novncproxy:2024.2                     \"dumb-init --single-\u2026\"   8 minutes ago    Up 8 minutes (healthy)              nova_novncproxy",
                                "811663d32f5d   registry.osism.tech/kolla/nova-conductor:2024.2                      \"dumb-init --single-\u2026\"   8 minutes ago    Up 8 minutes (healthy)              nova_conductor",
                                "c271d10de826   registry.osism.tech/kolla/grafana:2024.2                             \"dumb-init --single-\u2026\"   10 minutes ago   Up 10 minutes                       grafana",
                                "58dceffbac3d   registry.osism.tech/kolla/nova-api:2024.2                            \"dumb-init --single-\u2026\"   10 minutes ago   Up 10 minutes (healthy)             nova_api",
                                "acb91a3b3e52   registry.osism.tech/kolla/nova-scheduler:2024.2                      \"dumb-init --single-\u2026\"   10 minutes ago   Up 10 minutes (healthy)             nova_scheduler",
                                "66c3acb647f6   registry.osism.tech/kolla/cinder-backup:2024.2                       \"dumb-init --single-\u2026\"   11 minutes ago   Up 11 minutes (healthy)             cinder_backup",
                                "8ae7f28555cd   registry.osism.tech/kolla/cinder-volume:2024.2                       \"dumb-init --single-\u2026\"   11 minutes ago   Up 11 minutes (healthy)             cinder_volume",
                                "9bf88c0ea1d0   registry.osism.tech/kolla/glance-api:2024.2                          \"dumb-init --single-\u2026\"   12 minutes ago   Up 12 minutes (healthy)             glance_api",
                                "e5a3eec726f2   registry.osism.tech/kolla/cinder-scheduler:2024.2                    \"dumb-init --single-\u2026\"   12 minutes ago   Up 12 minutes (healthy)             cinder_scheduler",
                                "a36c539b639c   registry.osism.tech/kolla/cinder-api:2024.2                          \"dumb-init --single-\u2026\"   12 minutes ago   Up 12 minutes (healthy)             cinder_api",
                                "827264021b9d   registry.osism.tech/kolla/prometheus-elasticsearch-exporter:2024.2   \"dumb-init --single-\u2026\"   13 minutes ago   Up 13 minutes                       prometheus_elasticsearch_exporter",
                                "c0038ca67690   registry.osism.tech/kolla/prometheus-cadvisor:2024.2                 \"dumb-init --single-\u2026\"   13 minutes ago   Up 13 minutes                       prometheus_cadvisor",
                                "1941c2999ee6   registry.osism.tech/kolla/prometheus-memcached-exporter:2024.2       \"dumb-init --single-\u2026\"   13 minutes ago   Up 13 minutes                       prometheus_memcached_exporter",
                                "bac938ae89ca   registry.osism.tech/kolla/prometheus-mysqld-exporter:2024.2          \"dumb-init --single-\u2026\"   13 minutes ago   Up 13 minutes                       prometheus_mysqld_exporter",
                                "031f59732dc3   registry.osism.tech/kolla/prometheus-node-exporter:2024.2            \"dumb-init --single-\u2026\"   14 minutes ago   Up 14 minutes                       prometheus_node_exporter",
                                "9e744e3eaacd   registry.osism.tech/kolla/magnum-conductor:2024.2                    \"dumb-init --single-\u2026\"   15 minutes ago   Up 15 minutes (healthy)             magnum_conductor",
                                "301fea6b9b29   registry.osism.tech/kolla/magnum-api:2024.2                          \"dumb-init --single-\u2026\"   15 minutes ago   Up 15 minutes (healthy)             magnum_api",
                                "39d94878860b   registry.osism.tech/kolla/neutron-server:2024.2                      \"dumb-init --single-\u2026\"   15 minutes ago   Up 15 minutes (healthy)             neutron_server",
                                "980b2f7e410d   registry.osism.tech/kolla/designate-worker:2024.2                    \"dumb-init --single-\u2026\"   16 minutes ago   Up 16 minutes (healthy)             designate_worker",
                                "6ccc7bd0eff7   registry.osism.tech/kolla/designate-mdns:2024.2                      \"dumb-init --single-\u2026\"   16 minutes ago   Up 16 minutes (healthy)             designate_mdns",
                                "7acc7ed7f354   registry.osism.tech/kolla/placement-api:2024.2                       \"dumb-init --single-\u2026\"   16 minutes ago   Up 16 minutes (healthy)             placement_api",
                                "19ff8721f4df   registry.osism.tech/kolla/designate-producer:2024.2                  \"dumb-init --single-\u2026\"   17 minutes ago   Up 16 minutes (healthy)             designate_producer",
                                "d3435344fc9d   registry.osism.tech/kolla/designate-central:2024.2                   \"dumb-init --single-\u2026\"   17 minutes ago   Up 17 minutes (healthy)             designate_central",
                                "e99370646bc6   registry.osism.tech/kolla/designate-api:2024.2                       \"dumb-init --single-\u2026\"   17 minutes ago   Up 17 minutes (healthy)             designate_api",
                                "26829716792d   registry.osism.tech/kolla/barbican-worker:2024.2                     \"dumb-init --single-\u2026\"   17 minutes ago   Up 17 minutes (healthy)             barbican_worker",
                                "d93c9f61dfeb   registry.osism.tech/kolla/barbican-keystone-listener:2024.2          \"dumb-init --single-\u2026\"   17 minutes ago   Up 17 minutes (healthy)             barbican_keystone_listener",
                                "6d90d9f2763b   registry.osism.tech/kolla/designate-backend-bind9:2024.2             \"dumb-init --single-\u2026\"   17 minutes ago   Up 17 minutes (healthy)             designate_backend_bind9",
                                "b727b0e64d9f   registry.osism.tech/kolla/barbican-api:2024.2                        \"dumb-init --single-\u2026\"   17 minutes ago   Up 17 minutes (healthy)             barbican_api",
                                "ce6817b37e84   registry.osism.tech/osism/ceph-daemon:reef                           \"/usr/bin/ceph-mgr -\u2026\"   18 minutes ago   Up 18 minutes                       ceph-mgr-testbed-node-2",
                                "1dfda4644c27   registry.osism.tech/kolla/keystone:2024.2                            \"dumb-init --single-\u2026\"   21 minutes ago   Up 21 minutes (healthy)             keystone",
                                "03bc6395d015   registry.osism.tech/kolla/horizon:2024.2                             \"dumb-init --single-\u2026\"   21 minutes ago   Up 21 minutes (healthy)             horizon",
                                "2ea35d17ff6d   registry.osism.tech/kolla/keystone-fernet:2024.2                     \"dumb-init --single-\u2026\"   21 minutes ago   Up 21 minutes (healthy)             keystone_fernet",
                                "b1526b1936bc   registry.osism.tech/kolla/keystone-ssh:2024.2                        \"dumb-init --single-\u2026\"   21 minutes ago   Up 21 minutes (healthy)             keystone_ssh",
                                "d954826f0c18   registry.osism.tech/kolla/opensearch-dashboards:2024.2               \"dumb-init --single-\u2026\"   23 minutes ago   Up 23 minutes (healthy)             opensearch_dashboards",
                                "d2f01bbf78dc   registry.osism.tech/kolla/mariadb-server:2024.2                      \"dumb-init -- kolla_\u2026\"   24 minutes ago   Up 24 minutes (healthy)             mariadb",
                                "8ae3b0ecb003   registry.osism.tech/kolla/opensearch:2024.2                          \"dumb-init --single-\u2026\"   25 minutes ago   Up 24 minutes (healthy)             opensearch",
                                "41019e3dcdce   registry.osism.tech/osism/ceph-daemon:reef                           \"/usr/bin/ceph-crash\"    26 minutes ago   Up 26 minutes                       ceph-crash-testbed-node-2",
                                "05144260ecef   registry.osism.tech/kolla/keepalived:2024.2                          \"dumb-init --single-\u2026\"   26 minutes ago   Up 26 minutes                       keepalived",
                                "e61290ed4ec8   registry.osism.tech/kolla/proxysql:2024.2                            \"dumb-init --single-\u2026\"   26 minutes ago   Up 26 minutes (healthy)             proxysql",
                                "6bc306d583d0   registry.osism.tech/kolla/haproxy:2024.2                             \"dumb-init --single-\u2026\"   27 minutes ago   Up 27 minutes (healthy)             haproxy",
                                "109abbc59211   registry.osism.tech/kolla/ovn-northd:2024.2                          \"dumb-init --single-\u2026\"   30 minutes ago   Up 29 minutes                       ovn_northd",
                                "1bb7fc0990bc   registry.osism.tech/kolla/ovn-sb-db-server:2024.2                    \"dumb-init --single-\u2026\"   30 minutes ago   Up 29 minutes                       ovn_sb_db",
                                "7bf6dc9448f0   registry.osism.tech/kolla/ovn-nb-db-server:2024.2                    \"dumb-init --single-\u2026\"   30 minutes ago   Up 30 minutes                       ovn_nb_db",
                                "9670df24e6f8   registry.osism.tech/kolla/rabbitmq:2024.2                            \"dumb-init --single-\u2026\"   30 minutes ago   Up 30 minutes (healthy)             rabbitmq",
                                "c2f6ebfa804e   registry.osism.tech/kolla/ovn-controller:2024.2                      \"dumb-init --single-\u2026\"   30 minutes ago   Up 30 minutes                       ovn_controller",
                                "839d7a58ab74   registry.osism.tech/osism/ceph-daemon:reef                           \"/usr/bin/ceph-mon -\u2026\"   31 minutes ago   Up 31 minutes                       ceph-mon-testbed-node-2",
                                "006f681ca5fa   registry.osism.tech/kolla/openvswitch-vswitchd:2024.2                \"dumb-init --single-\u2026\"   32 minutes ago   Up 32 minutes (healthy)             openvswitch_vswitchd",
                                "1e6b05f5a799   registry.osism.tech/kolla/redis-sentinel:2024.2                      \"dumb-init --single-\u2026\"   32 minutes ago   Up 32 minutes (healthy)             redis_sentinel",
                                "5a83712dd552   registry.osism.tech/kolla/openvswitch-db-server:2024.2               \"dumb-init --single-\u2026\"   32 minutes ago   Up 32 minutes (healthy)             openvswitch_db",
                                "2957a3cb2019   registry.osism.tech/kolla/redis:2024.2                               \"dumb-init --single-\u2026\"   32 minutes ago   Up 32 minutes (healthy)             redis",
                                "695203b12a33   registry.osism.tech/kolla/memcached:2024.2                           \"dumb-init --single-\u2026\"   32 minutes ago   Up 32 minutes (healthy)             memcached",
                                "e86e254e6196   registry.osism.tech/kolla/cron:2024.2                                \"dumb-init --single-\u2026\"   33 minutes ago   Up 33 minutes                       cron",
                                "5a47ec47399b   registry.osism.tech/kolla/kolla-toolbox:2024.2                       \"dumb-init --single-\u2026\"   33 minutes ago   Up 33 minutes                       kolla_toolbox",
                                "7a0d1026089c   registry.osism.tech/kolla/fluentd:2024.2                             \"dumb-init --single-\u2026\"   33 minutes ago   Up 33 minutes                       fluentd",
                                "",
                                "## Images @ testbed-node-2",
                                "",
                                "+ echo",
                                "+ echo '## Images @ testbed-node-2'",
                                "+ echo",
                                "+ osism container testbed-node-2 images",
                                "REPOSITORY                                                    TAG       IMAGE ID       CREATED        SIZE",
                                "registry.osism.tech/osism/ceph-daemon                         reef      ff59b29334dd   22 hours ago   1.27GB",
                                "registry.osism.tech/kolla/memcached                           2024.2    3652d3a7de56   24 hours ago   272MB",
                                "registry.osism.tech/kolla/proxysql                            2024.2    dbe28c2fd6c3   24 hours ago   418MB",
                                "registry.osism.tech/kolla/kolla-toolbox                       2024.2    f894a5bc6d0b   24 hours ago   673MB",
                                "registry.osism.tech/kolla/haproxy                             2024.2    b1fd4c643197   24 hours ago   279MB",
                                "registry.osism.tech/kolla/cron                                2024.2    5b921837d137   24 hours ago   271MB",
                                "registry.osism.tech/kolla/rabbitmq                            2024.2    540f6409fbf5   24 hours ago   328MB",
                                "registry.osism.tech/kolla/keepalived                          2024.2    13679e653294   24 hours ago   282MB",
                                "registry.osism.tech/kolla/grafana                             2024.2    37fbf97e7f6f   24 hours ago   1.02GB",
                                "registry.osism.tech/kolla/fluentd                             2024.2    743bb34b6cee   24 hours ago   584MB",
                                "registry.osism.tech/kolla/opensearch-dashboards               2024.2    c159d79f20f0   24 hours ago   1.53GB",
                                "registry.osism.tech/kolla/opensearch                          2024.2    80d69bb5f098   24 hours ago   1.56GB",
                                "registry.osism.tech/kolla/openvswitch-db-server               2024.2    7e08fe973c8e   24 hours ago   284MB",
                                "registry.osism.tech/kolla/openvswitch-vswitchd                2024.2    4d41d184aa52   24 hours ago   284MB",
                                "registry.osism.tech/kolla/redis                               2024.2    0fa4d8b23966   24 hours ago   278MB",
                                "registry.osism.tech/kolla/redis-sentinel                      2024.2    2052a297db29   24 hours ago   278MB",
                                "registry.osism.tech/kolla/horizon                             2024.2    4d92e9384e36   24 hours ago   1.15GB",
                                "registry.osism.tech/kolla/mariadb-server                      2024.2    5dbd7b932e63   24 hours ago   457MB",
                                "registry.osism.tech/kolla/prometheus-node-exporter            2024.2    278eaa73d02c   24 hours ago   311MB",
                                "registry.osism.tech/kolla/prometheus-elasticsearch-exporter   2024.2    13638ae36e57   24 hours ago   297MB",
                                "registry.osism.tech/kolla/prometheus-memcached-exporter       2024.2    8ff5a6b3cd16   24 hours ago   304MB",
                                "registry.osism.tech/kolla/prometheus-mysqld-exporter          2024.2    d9cb73d898bb   24 hours ago   306MB",
                                "registry.osism.tech/kolla/prometheus-cadvisor                 2024.2    4e4cb8e72dfc   24 hours ago   363MB",
                                "registry.osism.tech/kolla/ovn-nb-db-server                    2024.2    51d6ef378bbd   24 hours ago   846MB",
                                "registry.osism.tech/kolla/ovn-sb-db-server                    2024.2    25bb84f2289b   24 hours ago   846MB",
                                "registry.osism.tech/kolla/ovn-northd                          2024.2    47dae0f63bd5   24 hours ago   846MB",
                                "registry.osism.tech/kolla/ovn-controller                      2024.2    342e8da8ac3a   24 hours ago   846MB",
                                "registry.osism.tech/kolla/neutron-server                      2024.2    189daad12150   24 hours ago   1.17GB",
                                "registry.osism.tech/kolla/octavia-api                         2024.2    624955c2c115   24 hours ago   1.06GB",
                                "registry.osism.tech/kolla/octavia-housekeeping                2024.2    9b80c75ebf88   24 hours ago   1.03GB",
                                "registry.osism.tech/kolla/octavia-driver-agent                2024.2    3f4a0924d9ed   24 hours ago   1.06GB",
                                "registry.osism.tech/kolla/octavia-health-manager              2024.2    741914d697fa   24 hours ago   1.03GB",
                                "registry.osism.tech/kolla/octavia-worker                      2024.2    45d576cfc14a   24 hours ago   1.03GB",
                                "registry.osism.tech/kolla/cinder-backup                       2024.2    db0ba08b76e7   24 hours ago   1.42GB",
                                "registry.osism.tech/kolla/cinder-volume                       2024.2    aad4ba476f1c   24 hours ago   1.72GB",
                                "registry.osism.tech/kolla/cinder-scheduler                    2024.2    3102035cbd3f   24 hours ago   1.41GB",
                                "registry.osism.tech/kolla/cinder-api                          2024.2    ffa30f3933ed   24 hours ago   1.41GB",
                                "registry.osism.tech/kolla/placement-api                       2024.2    e4b48b9cb79e   24 hours ago   981MB",
                                "registry.osism.tech/kolla/designate-producer                  2024.2    e3b7812cd6ff   24 hours ago   990MB",
                                "registry.osism.tech/kolla/designate-central                   2024.2    8ab2ea11494a   24 hours ago   989MB",
                                "registry.osism.tech/kolla/designate-worker                    2024.2    13bd17ae9e62   24 hours ago   994MB",
                                "registry.osism.tech/kolla/designate-api                       2024.2    2e44006b8d8c   24 hours ago   990MB",
                                "registry.osism.tech/kolla/designate-backend-bind9             2024.2    7b60bbae1ff0   24 hours ago   994MB",
                                "registry.osism.tech/kolla/designate-mdns                      2024.2    a28e142a95bf   24 hours ago   990MB",
                                "registry.osism.tech/kolla/barbican-worker                     2024.2    0eec7e0f7b33   24 hours ago   996MB",
                                "registry.osism.tech/kolla/barbican-keystone-listener          2024.2    8c429b2daef4   24 hours ago   996MB",
                                "registry.osism.tech/kolla/barbican-api                        2024.2    9d4b161f78ee   24 hours ago   996MB",
                                "registry.osism.tech/kolla/keystone-fernet                     2024.2    3770c4e50a62   24 hours ago   1.04GB",
                                "registry.osism.tech/kolla/keystone-ssh                        2024.2    a75dcc713ad6   24 hours ago   1.05GB",
                                "registry.osism.tech/kolla/keystone                            2024.2    f3afb1708fa9   24 hours ago   1.07GB",
                                "registry.osism.tech/kolla/nova-conductor                      2024.2    7e0dfe6dfac8   24 hours ago   1.22GB",
                                "registry.osism.tech/kolla/nova-api                            2024.2    b141835a8ce3   24 hours ago   1.22GB",
                                "registry.osism.tech/kolla/nova-novncproxy                     2024.2    0e230dfe273c   24 hours ago   1.37GB",
                                "registry.osism.tech/kolla/nova-scheduler                      2024.2    1d5f182c3a75   24 hours ago   1.22GB",
                                "registry.osism.tech/kolla/magnum-api                          2024.2    5f8755d26e30   24 hours ago   1.13GB",
                                "registry.osism.tech/kolla/magnum-conductor                    2024.2    ad0e7877ae1a   24 hours ago   1.25GB",
                                "registry.osism.tech/kolla/glance-api                          2024.2    4e1be079f392   24 hours ago   1.1GB",
                                "+ sh -c /opt/configuration/scripts/check-services.sh",
                                "+ set -e",
                                "+ source /opt/manager-vars.sh",
                                "++ export NUMBER_OF_NODES=6",
                                "++ NUMBER_OF_NODES=6",
                                "++ export CEPH_VERSION=reef",
                                "++ CEPH_VERSION=reef",
                                "++ export CONFIGURATION_VERSION=main",
                                "++ CONFIGURATION_VERSION=main",
                                "++ export MANAGER_VERSION=latest",
                                "++ MANAGER_VERSION=latest",
                                "++ export OPENSTACK_VERSION=2024.2",
                                "++ OPENSTACK_VERSION=2024.2",
                                "++ export ARA=false",
                                "++ ARA=false",
                                "++ export DEPLOY_MODE=manager",
                                "++ DEPLOY_MODE=manager",
                                "++ export TEMPEST=true",
                                "++ TEMPEST=true",
                                "++ export IS_ZUUL=true",
                                "++ IS_ZUUL=true",
                                "++ export MANAGER_PUBLIC_IP_ADDRESS=81.163.192.195",
                                "++ MANAGER_PUBLIC_IP_ADDRESS=81.163.192.195",
                                "++ export EXTERNAL_API=false",
                                "++ EXTERNAL_API=false",
                                "++ export IMAGE_USER=ubuntu",
                                "++ IMAGE_USER=ubuntu",
                                "++ export IMAGE_NODE_USER=ubuntu",
                                "++ IMAGE_NODE_USER=ubuntu",
                                "++ export CEPH_STACK=ceph-ansible",
                                "++ CEPH_STACK=ceph-ansible",
                                "+ [[ ceph-ansible == \\c\\e\\p\\h\\-\\a\\n\\s\\i\\b\\l\\e ]]",
                                "+ sh -c /opt/configuration/scripts/check/100-ceph-with-ansible.sh",
                                "+ set -e",
                                "+ source /opt/configuration/scripts/include.sh",
                                "++ export INTERACTIVE=false",
                                "++ INTERACTIVE=false",
                                "++ export OSISM_APPLY_RETRY=1",
                                "++ OSISM_APPLY_RETRY=1",
                                "+ source /opt/configuration/scripts/manager-version.sh",
                                "+++ awk '-F: ' '/^manager_version:/ { print $2 }' /opt/configuration/environments/manager/configuration.yml",
                                "",
                                "# Ceph status",
                                "",
                                "++ export MANAGER_VERSION=latest",
                                "++ MANAGER_VERSION=latest",
                                "+ echo",
                                "+ echo '# Ceph status'",
                                "+ echo",
                                "+ ceph -s",
                                "  cluster:",
                                "    id:     11111111-1111-1111-1111-111111111111",
                                "    health: HEALTH_OK",
                                " ",
                                "  services:",
                                "    mon: 3 daemons, quorum testbed-node-0,testbed-node-1,testbed-node-2 (age 31m)",
                                "    mgr: testbed-node-0(active, since 17m), standbys: testbed-node-1, testbed-node-2",
                                "    mds: 1/1 daemons up, 2 standby",
                                "    osd: 6 osds: 6 up (since 27m), 6 in (since 27m)",
                                "    rgw: 3 daemons active (3 hosts, 1 zones)",
                                " ",
                                "  data:",
                                "    volumes: 1/1 healthy",
                                "    pools:   14 pools, 401 pgs",
                                "    objects: 554 objects, 2.2 GiB",
                                "    usage:   7.1 GiB used, 113 GiB / 120 GiB avail",
                                "    pgs:     401 active+clean",
                                " ",
                                "",
                                "# Ceph versions",
                                "",
                                "+ echo",
                                "+ echo '# Ceph versions'",
                                "+ echo",
                                "+ ceph versions",
                                "{",
                                "    \"mon\": {",
                                "        \"ceph version 18.2.7 (6b0e988052ec84cf2d4a54ff9bbbc5e720b621ad) reef (stable)\": 3",
                                "    },",
                                "    \"mgr\": {",
                                "        \"ceph version 18.2.7 (6b0e988052ec84cf2d4a54ff9bbbc5e720b621ad) reef (stable)\": 3",
                                "    },",
                                "    \"osd\": {",
                                "        \"ceph version 18.2.7 (6b0e988052ec84cf2d4a54ff9bbbc5e720b621ad) reef (stable)\": 6",
                                "    },",
                                "    \"mds\": {",
                                "        \"ceph version 18.2.7 (6b0e988052ec84cf2d4a54ff9bbbc5e720b621ad) reef (stable)\": 3",
                                "    },",
                                "    \"rgw\": {",
                                "        \"ceph version 18.2.7 (6b0e988052ec84cf2d4a54ff9bbbc5e720b621ad) reef (stable)\": 3",
                                "    },",
                                "    \"overall\": {",
                                "        \"ceph version 18.2.7 (6b0e988052ec84cf2d4a54ff9bbbc5e720b621ad) reef (stable)\": 18",
                                "    }",
                                "}",
                                "",
                                "# Ceph OSD tree",
                                "",
                                "+ echo",
                                "+ echo '# Ceph OSD tree'",
                                "+ echo",
                                "+ ceph osd df tree",
                                "ID  CLASS  WEIGHT   REWEIGHT  SIZE     RAW USE  DATA     OMAP     META     AVAIL    %USE  VAR   PGS  STATUS  TYPE NAME              ",
                                "-1         0.11691         -  120 GiB  7.1 GiB  6.7 GiB    6 KiB  412 MiB  113 GiB  5.90  1.00    -          root default           ",
                                "-5         0.03897         -   40 GiB  2.4 GiB  2.2 GiB    2 KiB  143 MiB   38 GiB  5.91  1.00    -              host testbed-node-3",
                                " 0    hdd  0.01949   1.00000   20 GiB  1.1 GiB  1.0 GiB    1 KiB   74 MiB   19 GiB  5.57  0.94  190      up          osd.0          ",
                                " 4    hdd  0.01949   1.00000   20 GiB  1.3 GiB  1.2 GiB    1 KiB   70 MiB   19 GiB  6.26  1.06  202      up          osd.4          ",
                                "-7         0.03897         -   40 GiB  2.4 GiB  2.2 GiB    2 KiB  143 MiB   38 GiB  5.91  1.00    -              host testbed-node-4",
                                " 1    hdd  0.01949   1.00000   20 GiB  1.5 GiB  1.4 GiB    1 KiB   74 MiB   19 GiB  7.41  1.26  197      up          osd.1          ",
                                " 5    hdd  0.01949   1.00000   20 GiB  904 MiB  835 MiB    1 KiB   70 MiB   19 GiB  4.42  0.75  191      up          osd.5          ",
                                "-3         0.03897         -   40 GiB  2.3 GiB  2.2 GiB    2 KiB  126 MiB   38 GiB  5.87  0.99    -              host testbed-node-5",
                                " 2    hdd  0.01949   1.00000   20 GiB  984 MiB  914 MiB    1 KiB   70 MiB   19 GiB  4.81  0.82  196      up          osd.2          ",
                                " 3    hdd  0.01949   1.00000   20 GiB  1.4 GiB  1.3 GiB    1 KiB   56 MiB   19 GiB  6.93  1.17  194      up          osd.3          ",
                                "                       TOTAL  120 GiB  7.1 GiB  6.7 GiB  9.3 KiB  412 MiB  113 GiB  5.90                                            ",
                                "MIN/MAX VAR: 0.75/1.26  STDDEV: 1.08",
                                "",
                                "# Ceph monitor status",
                                "",
                                "+ echo",
                                "+ echo '# Ceph monitor status'",
                                "+ echo",
                                "+ ceph mon stat",
                                "e1: 3 mons at {testbed-node-0=[v2:192.168.16.10:3300/0,v1:192.168.16.10:6789/0],testbed-node-1=[v2:192.168.16.11:3300/0,v1:192.168.16.11:6789/0],testbed-node-2=[v2:192.168.16.12:3300/0,v1:192.168.16.12:6789/0]} removed_ranks: {} disallowed_leaders: {}, election epoch 6, leader 0 testbed-node-0, quorum 0,1,2 testbed-node-0,testbed-node-1,testbed-node-2",
                                "",
                                "# Ceph quorum status",
                                "",
                                "+ echo",
                                "+ echo '# Ceph quorum status'",
                                "+ echo",
                                "+ ceph quorum_status",
                                "+ jq",
                                "{",
                                "  \"election_epoch\": 6,",
                                "  \"quorum\": [",
                                "    0,",
                                "    1,",
                                "    2",
                                "  ],",
                                "  \"quorum_names\": [",
                                "    \"testbed-node-0\",",
                                "    \"testbed-node-1\",",
                                "    \"testbed-node-2\"",
                                "  ],",
                                "  \"quorum_leader_name\": \"testbed-node-0\",",
                                "  \"quorum_age\": 1877,",
                                "  \"features\": {",
                                "    \"quorum_con\": \"4540138322906710015\",",
                                "    \"quorum_mon\": [",
                                "      \"kraken\",",
                                "      \"luminous\",",
                                "      \"mimic\",",
                                "      \"osdmap-prune\",",
                                "      \"nautilus\",",
                                "      \"octopus\",",
                                "      \"pacific\",",
                                "      \"elector-pinging\",",
                                "      \"quincy\",",
                                "      \"reef\"",
                                "    ]",
                                "  },",
                                "  \"monmap\": {",
                                "    \"epoch\": 1,",
                                "    \"fsid\": \"11111111-1111-1111-1111-111111111111\",",
                                "    \"modified\": \"2026-02-24T00:51:03.217233Z\",",
                                "    \"created\": \"2026-02-24T00:51:03.217233Z\",",
                                "    \"min_mon_release\": 18,",
                                "    \"min_mon_release_name\": \"reef\",",
                                "    \"election_strategy\": 1,",
                                "    \"disallowed_leaders: \": \"\",",
                                "    \"stretch_mode\": false,",
                                "    \"tiebreaker_mon\": \"\",",
                                "    \"removed_ranks: \": \"\",",
                                "    \"features\": {",
                                "      \"persistent\": [",
                                "        \"kraken\",",
                                "        \"luminous\",",
                                "        \"mimic\",",
                                "        \"osdmap-prune\",",
                                "        \"nautilus\",",
                                "        \"octopus\",",
                                "        \"pacific\",",
                                "        \"elector-pinging\",",
                                "        \"quincy\",",
                                "        \"reef\"",
                                "      ],",
                                "      \"optional\": []",
                                "    },",
                                "    \"mons\": [",
                                "      {",
                                "        \"rank\": 0,",
                                "        \"name\": \"testbed-node-0\",",
                                "        \"public_addrs\": {",
                                "          \"addrvec\": [",
                                "            {",
                                "              \"type\": \"v2\",",
                                "              \"addr\": \"192.168.16.10:3300\",",
                                "              \"nonce\": 0",
                                "            },",
                                "            {",
                                "              \"type\": \"v1\",",
                                "              \"addr\": \"192.168.16.10:6789\",",
                                "              \"nonce\": 0",
                                "            }",
                                "          ]",
                                "        },",
                                "        \"addr\": \"192.168.16.10:6789/0\",",
                                "        \"public_addr\": \"192.168.16.10:6789/0\",",
                                "        \"priority\": 0,",
                                "        \"weight\": 0,",
                                "        \"crush_location\": \"{}\"",
                                "      },",
                                "      {",
                                "        \"rank\": 1,",
                                "        \"name\": \"testbed-node-1\",",
                                "        \"public_addrs\": {",
                                "          \"addrvec\": [",
                                "            {",
                                "              \"type\": \"v2\",",
                                "              \"addr\": \"192.168.16.11:3300\",",
                                "              \"nonce\": 0",
                                "            },",
                                "            {",
                                "              \"type\": \"v1\",",
                                "              \"addr\": \"192.168.16.11:6789\",",
                                "              \"nonce\": 0",
                                "            }",
                                "          ]",
                                "        },",
                                "        \"addr\": \"192.168.16.11:6789/0\",",
                                "        \"public_addr\": \"192.168.16.11:6789/0\",",
                                "        \"priority\": 0,",
                                "        \"weight\": 0,",
                                "        \"crush_location\": \"{}\"",
                                "      },",
                                "      {",
                                "        \"rank\": 2,",
                                "        \"name\": \"testbed-node-2\",",
                                "        \"public_addrs\": {",
                                "          \"addrvec\": [",
                                "            {",
                                "              \"type\": \"v2\",",
                                "              \"addr\": \"192.168.16.12:3300\",",
                                "              \"nonce\": 0",
                                "            },",
                                "            {",
                                "              \"type\": \"v1\",",
                                "              \"addr\": \"192.168.16.12:6789\",",
                                "              \"nonce\": 0",
                                "            }",
                                "          ]",
                                "        },",
                                "        \"addr\": \"192.168.16.12:6789/0\",",
                                "        \"public_addr\": \"192.168.16.12:6789/0\",",
                                "        \"priority\": 0,",
                                "        \"weight\": 0,",
                                "        \"crush_location\": \"{}\"",
                                "      }",
                                "    ]",
                                "  }",
                                "}",
                                "",
                                "# Ceph free space status",
                                "",
                                "+ echo",
                                "+ echo '# Ceph free space status'",
                                "+ echo",
                                "+ ceph df",
                                "--- RAW STORAGE ---",
                                "CLASS     SIZE    AVAIL     USED  RAW USED  %RAW USED",
                                "hdd    120 GiB  113 GiB  7.1 GiB   7.1 GiB       5.90",
                                "TOTAL  120 GiB  113 GiB  7.1 GiB   7.1 GiB       5.90",
                                " ",
                                "--- POOLS ---",
                                "POOL                       ID  PGS   STORED  OBJECTS     USED  %USED  MAX AVAIL",
                                ".mgr                        1    1  577 KiB        2  1.1 MiB      0     53 GiB",
                                "cephfs_data                 2   32      0 B        0      0 B      0     35 GiB",
                                "cephfs_metadata             3   16  4.4 KiB       22   96 KiB      0     35 GiB",
                                "default.rgw.buckets.data    4   32      0 B        0      0 B      0     35 GiB",
                                "default.rgw.buckets.index   5   32      0 B        0      0 B      0     35 GiB",
                                "default.rgw.control         6   32      0 B        8      0 B      0     35 GiB",
                                "default.rgw.log             7   32  3.6 KiB      209  408 KiB      0     35 GiB",
                                "default.rgw.meta            8   32      0 B        0      0 B      0     35 GiB",
                                ".rgw.root                   9   32  2.6 KiB        6   48 KiB      0     53 GiB",
                                "backups                    10   32     19 B        2   12 KiB      0     35 GiB",
                                "volumes                    11   32     19 B        2   12 KiB      0     35 GiB",
                                "images                     12   32  2.2 GiB      299  6.7 GiB   5.96     35 GiB",
                                "metrics                    13   32     19 B        2   12 KiB      0     35 GiB",
                                "vms                        14   32     19 B        2   12 KiB      0     35 GiB",
                                "++ semver latest 5.0.0",
                                "+ [[ -1 -eq -1 ]]",
                                "+ [[ latest != \\l\\a\\t\\e\\s\\t ]]",
                                "+ [[ ! -e /etc/redhat-release ]]",
                                "+ osism apply facts",
                                "\u001b[32m2026-02-24 01:22:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mPrepare task for execution of facts.\u001b[0m",
                                "\u001b[32m2026-02-24 01:22:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 87d1577d-7457-4d6b-809f-ff0d65828ac8 (facts) was prepared for execution.\u001b[0m",
                                "\u001b[32m2026-02-24 01:22:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task 87d1577d-7457-4d6b-809f-ff0d65828ac8 (facts) has been started and output is visible here.\u001b[0m",
                                "",
                                "PLAY [Apply role facts] ********************************************************",
                                "",
                                "TASK [osism.commons.facts : Create custom facts directory] *********************",
                                "Tuesday 24 February 2026  01:22:52 +0000 (0:00:00.284)       0:00:00.284 ****** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [osism.commons.facts : Copy fact files] ***********************************",
                                "Tuesday 24 February 2026  01:22:53 +0000 (0:00:01.573)       0:00:01.857 ****** ",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "PLAY [Gather facts for all hosts] **********************************************",
                                "",
                                "TASK [Gathers facts about hosts] ***********************************************",
                                "Tuesday 24 February 2026  01:22:55 +0000 (0:00:01.635)       0:00:03.493 ****** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "PLAY [Gather facts for all hosts if using --limit] *****************************",
                                "",
                                "TASK [Gather facts for all hosts] **********************************************",
                                "Tuesday 24 February 2026  01:23:00 +0000 (0:00:05.428)       0:00:08.921 ****** ",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;32mtestbed-manager\u001b[0m            : \u001b[0;32mok=2   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;32mtestbed-node-0\u001b[0m             : \u001b[0;32mok=2   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;32mtestbed-node-1\u001b[0m             : \u001b[0;32mok=2   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;32mtestbed-node-2\u001b[0m             : \u001b[0;32mok=2   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;32mtestbed-node-3\u001b[0m             : \u001b[0;32mok=2   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;32mtestbed-node-4\u001b[0m             : \u001b[0;32mok=2   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;32mtestbed-node-5\u001b[0m             : \u001b[0;32mok=2   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Tuesday 24 February 2026  01:23:01 +0000 (0:00:00.650)       0:00:09.571 ****** ",
                                "=============================================================================== ",
                                "Gathers facts about hosts ----------------------------------------------- 5.43s",
                                "osism.commons.facts : Copy fact files ----------------------------------- 1.64s",
                                "osism.commons.facts : Create custom facts directory --------------------- 1.57s",
                                "Gather facts for all hosts ---------------------------------------------- 0.65s",
                                "+ osism validate ceph-mons",
                                "",
                                "PLAY [Ceph validate mons] ******************************************************",
                                "",
                                "TASK [Get timestamp for report file] *******************************************",
                                "Tuesday 24 February 2026  01:23:20 +0000 (0:00:00.517)       0:00:00.517 ****** ",
                                "\u001b[0;32mok: [testbed-node-0 -> testbed-manager(192.168.16.5)]\u001b[0m",
                                "",
                                "TASK [Create report output directory] ******************************************",
                                "Tuesday 24 February 2026  01:23:21 +0000 (0:00:00.994)       0:00:01.511 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0 -> testbed-manager(192.168.16.5)]\u001b[0m",
                                "",
                                "TASK [Define report vars] ******************************************************",
                                "Tuesday 24 February 2026  01:23:22 +0000 (0:00:01.123)       0:00:02.635 ****** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [Prepare test data for container existance test] **************************",
                                "Tuesday 24 February 2026  01:23:22 +0000 (0:00:00.129)       0:00:02.764 ****** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [Get container info] ******************************************************",
                                "Tuesday 24 February 2026  01:23:22 +0000 (0:00:00.336)       0:00:03.101 ****** ",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [Set test result to failed if container is missing] ***********************",
                                "Tuesday 24 February 2026  01:23:24 +0000 (0:00:01.113)       0:00:04.215 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [Set test result to passed if container is existing] **********************",
                                "Tuesday 24 February 2026  01:23:24 +0000 (0:00:00.347)       0:00:04.562 ****** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [Prepare test data] *******************************************************",
                                "Tuesday 24 February 2026  01:23:24 +0000 (0:00:00.559)       0:00:05.122 ****** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [Set test result to failed if ceph-mon is not running] ********************",
                                "Tuesday 24 February 2026  01:23:25 +0000 (0:00:00.366)       0:00:05.488 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [Set test result to passed if ceph-mon is running] ************************",
                                "Tuesday 24 February 2026  01:23:25 +0000 (0:00:00.335)       0:00:05.823 ****** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [Aggregate test results step one] *****************************************",
                                "Tuesday 24 February 2026  01:23:26 +0000 (0:00:00.545)       0:00:06.369 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [Aggregate test results step two] *****************************************",
                                "Tuesday 24 February 2026  01:23:26 +0000 (0:00:00.281)       0:00:06.651 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [Aggregate test results step three] ***************************************",
                                "Tuesday 24 February 2026  01:23:26 +0000 (0:00:00.266)       0:00:06.918 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [Flush handlers] **********************************************************",
                                "Tuesday 24 February 2026  01:23:27 +0000 (0:00:00.264)       0:00:07.182 ****** ",
                                "",
                                "TASK [Flush handlers] **********************************************************",
                                "Tuesday 24 February 2026  01:23:27 +0000 (0:00:00.077)       0:00:07.260 ****** ",
                                "",
                                "TASK [Flush handlers] **********************************************************",
                                "Tuesday 24 February 2026  01:23:27 +0000 (0:00:00.073)       0:00:07.333 ****** ",
                                "",
                                "TASK [Print report file information] *******************************************",
                                "Tuesday 24 February 2026  01:23:27 +0000 (0:00:00.074)       0:00:07.408 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [Fail due to missing containers] ******************************************",
                                "Tuesday 24 February 2026  01:23:27 +0000 (0:00:00.264)       0:00:07.672 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [Prepare quorum test vars] ************************************************",
                                "Tuesday 24 February 2026  01:23:27 +0000 (0:00:00.271)       0:00:07.944 ****** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [Get monmap info from one mon container] **********************************",
                                "Tuesday 24 February 2026  01:23:27 +0000 (0:00:00.134)       0:00:08.079 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [Set quorum test data] ****************************************************",
                                "Tuesday 24 February 2026  01:23:29 +0000 (0:00:01.709)       0:00:09.788 ****** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [Fail quorum test if not all monitors are in quorum] **********************",
                                "Tuesday 24 February 2026  01:23:30 +0000 (0:00:00.585)       0:00:10.374 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [Pass quorum test if all monitors are in quorum] **************************",
                                "Tuesday 24 February 2026  01:23:30 +0000 (0:00:00.130)       0:00:10.505 ****** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [Set fsid test vars] ******************************************************",
                                "Tuesday 24 February 2026  01:23:30 +0000 (0:00:00.426)       0:00:10.932 ****** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [Fail Cluster FSID test if FSID does not match configuration] *************",
                                "Tuesday 24 February 2026  01:23:31 +0000 (0:00:00.334)       0:00:11.267 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [Pass Cluster FSID test if it matches configuration] **********************",
                                "Tuesday 24 February 2026  01:23:31 +0000 (0:00:00.123)       0:00:11.390 ****** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [Prepare status test vars] ************************************************",
                                "Tuesday 24 February 2026  01:23:31 +0000 (0:00:00.128)       0:00:11.519 ****** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [Gather status data] ******************************************************",
                                "Tuesday 24 February 2026  01:23:31 +0000 (0:00:00.128)       0:00:11.648 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [Set health test data] ****************************************************",
                                "Tuesday 24 February 2026  01:23:32 +0000 (0:00:01.508)       0:00:13.157 ****** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [Fail cluster-health if health is not acceptable] *************************",
                                "Tuesday 24 February 2026  01:23:33 +0000 (0:00:00.380)       0:00:13.537 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [Pass cluster-health if health is acceptable] *****************************",
                                "Tuesday 24 February 2026  01:23:33 +0000 (0:00:00.154)       0:00:13.692 ****** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [Fail cluster-health if health is not acceptable (strict)] ****************",
                                "Tuesday 24 February 2026  01:23:33 +0000 (0:00:00.161)       0:00:13.854 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [Pass cluster-health if status is OK (strict)] ****************************",
                                "Tuesday 24 February 2026  01:23:34 +0000 (0:00:00.379)       0:00:14.233 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [Set validation result to passed if no test failed] ***********************",
                                "Tuesday 24 February 2026  01:23:34 +0000 (0:00:00.142)       0:00:14.376 ****** ",
                                "\u001b[0;32mok: [testbed-node-0 -> testbed-manager(192.168.16.5)]\u001b[0m",
                                "",
                                "TASK [Set validation result to failed if a test failed] ************************",
                                "Tuesday 24 February 2026  01:23:34 +0000 (0:00:00.283)       0:00:14.659 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [Aggregate test results step one] *****************************************",
                                "Tuesday 24 February 2026  01:23:34 +0000 (0:00:00.268)       0:00:14.928 ****** ",
                                "\u001b[0;32mok: [testbed-node-0 -> testbed-manager(192.168.16.5)]\u001b[0m",
                                "",
                                "TASK [Aggregate test results step two] *****************************************",
                                "Tuesday 24 February 2026  01:23:36 +0000 (0:00:01.921)       0:00:16.850 ****** ",
                                "\u001b[0;32mok: [testbed-node-0 -> testbed-manager(192.168.16.5)]\u001b[0m",
                                "",
                                "TASK [Aggregate test results step three] ***************************************",
                                "Tuesday 24 February 2026  01:23:36 +0000 (0:00:00.259)       0:00:17.109 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0 -> testbed-manager(192.168.16.5)]\u001b[0m",
                                "",
                                "TASK [Flush handlers] **********************************************************",
                                "Tuesday 24 February 2026  01:23:37 +0000 (0:00:00.312)       0:00:17.422 ****** ",
                                "",
                                "TASK [Flush handlers] **********************************************************",
                                "Tuesday 24 February 2026  01:23:37 +0000 (0:00:00.077)       0:00:17.499 ****** ",
                                "",
                                "TASK [Flush handlers] **********************************************************",
                                "Tuesday 24 February 2026  01:23:37 +0000 (0:00:00.072)       0:00:17.572 ****** ",
                                "",
                                "RUNNING HANDLER [Write report file] ********************************************",
                                "Tuesday 24 February 2026  01:23:37 +0000 (0:00:00.079)       0:00:17.651 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0 -> testbed-manager(192.168.16.5)]\u001b[0m",
                                "",
                                "TASK [Print report file information] *******************************************",
                                "Tuesday 24 February 2026  01:23:39 +0000 (0:00:01.717)       0:00:19.369 ****** ",
                                "\u001b[0;32mok: [testbed-node-0 -> testbed-manager(192.168.16.5)] => {\u001b[0m",
                                "\u001b[0;32m    \"msg\": [\u001b[0m",
                                "\u001b[0;32m        \"Validator run completed.\",\u001b[0m",
                                "\u001b[0;32m        \"You can find the report file here:\",\u001b[0m",
                                "\u001b[0;32m        \"/opt/reports/validator/ceph-mons-validator-2026-02-24T01:23:21+00:00-report.json\",\u001b[0m",
                                "\u001b[0;32m        \"on the following host:\",\u001b[0m",
                                "\u001b[0;32m        \"testbed-manager\"\u001b[0m",
                                "\u001b[0;32m    ]\u001b[0m",
                                "\u001b[0;32m}\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=24  \u001b[0m \u001b[0;33mchanged=5   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=13  \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;32mtestbed-node-1\u001b[0m             : \u001b[0;32mok=5   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;32mtestbed-node-2\u001b[0m             : \u001b[0;32mok=5   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Tuesday 24 February 2026  01:23:40 +0000 (0:00:01.070)       0:00:20.440 ****** ",
                                "=============================================================================== ",
                                "Aggregate test results step one ----------------------------------------- 1.92s",
                                "Write report file ------------------------------------------------------- 1.72s",
                                "Get monmap info from one mon container ---------------------------------- 1.71s",
                                "Gather status data ------------------------------------------------------ 1.51s",
                                "Create report output directory ------------------------------------------ 1.12s",
                                "Get container info ------------------------------------------------------ 1.11s",
                                "Print report file information ------------------------------------------- 1.07s",
                                "Get timestamp for report file ------------------------------------------- 0.99s",
                                "Set quorum test data ---------------------------------------------------- 0.59s",
                                "Set test result to passed if container is existing ---------------------- 0.56s",
                                "Set test result to passed if ceph-mon is running ------------------------ 0.55s",
                                "Pass quorum test if all monitors are in quorum -------------------------- 0.43s",
                                "Set health test data ---------------------------------------------------- 0.38s",
                                "Fail cluster-health if health is not acceptable (strict) ---------------- 0.38s",
                                "Prepare test data ------------------------------------------------------- 0.37s",
                                "Set test result to failed if container is missing ----------------------- 0.35s",
                                "Prepare test data for container existance test -------------------------- 0.34s",
                                "Set test result to failed if ceph-mon is not running -------------------- 0.34s",
                                "Set fsid test vars ------------------------------------------------------ 0.33s",
                                "Aggregate test results step three --------------------------------------- 0.31s",
                                "+ osism validate ceph-mgrs",
                                "",
                                "PLAY [Ceph validate mgrs] ******************************************************",
                                "",
                                "TASK [Get timestamp for report file] *******************************************",
                                "Tuesday 24 February 2026  01:23:58 +0000 (0:00:00.537)       0:00:00.537 ****** ",
                                "\u001b[0;32mok: [testbed-node-0 -> testbed-manager(192.168.16.5)]\u001b[0m",
                                "",
                                "TASK [Create report output directory] ******************************************",
                                "Tuesday 24 February 2026  01:23:59 +0000 (0:00:01.035)       0:00:01.573 ****** ",
                                "\u001b[0;32mok: [testbed-node-0 -> testbed-manager(192.168.16.5)]\u001b[0m",
                                "",
                                "TASK [Define report vars] ******************************************************",
                                "Tuesday 24 February 2026  01:24:01 +0000 (0:00:01.228)       0:00:02.801 ****** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [Prepare test data for container existance test] **************************",
                                "Tuesday 24 February 2026  01:24:01 +0000 (0:00:00.148)       0:00:02.950 ****** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [Get container info] ******************************************************",
                                "Tuesday 24 February 2026  01:24:01 +0000 (0:00:00.352)       0:00:03.303 ****** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [Set test result to failed if container is missing] ***********************",
                                "Tuesday 24 February 2026  01:24:02 +0000 (0:00:01.087)       0:00:04.391 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [Set test result to passed if container is existing] **********************",
                                "Tuesday 24 February 2026  01:24:03 +0000 (0:00:00.328)       0:00:04.719 ****** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [Prepare test data] *******************************************************",
                                "Tuesday 24 February 2026  01:24:03 +0000 (0:00:00.597)       0:00:05.317 ****** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [Set test result to failed if ceph-mgr is not running] ********************",
                                "Tuesday 24 February 2026  01:24:04 +0000 (0:00:00.337)       0:00:05.654 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [Set test result to passed if ceph-mgr is running] ************************",
                                "Tuesday 24 February 2026  01:24:04 +0000 (0:00:00.330)       0:00:05.985 ****** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [Aggregate test results step one] *****************************************",
                                "Tuesday 24 February 2026  01:24:04 +0000 (0:00:00.552)       0:00:06.537 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [Aggregate test results step two] *****************************************",
                                "Tuesday 24 February 2026  01:24:05 +0000 (0:00:00.306)       0:00:06.843 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [Aggregate test results step three] ***************************************",
                                "Tuesday 24 February 2026  01:24:05 +0000 (0:00:00.278)       0:00:07.122 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [Flush handlers] **********************************************************",
                                "Tuesday 24 February 2026  01:24:05 +0000 (0:00:00.278)       0:00:07.400 ****** ",
                                "",
                                "TASK [Flush handlers] **********************************************************",
                                "Tuesday 24 February 2026  01:24:05 +0000 (0:00:00.071)       0:00:07.472 ****** ",
                                "",
                                "TASK [Flush handlers] **********************************************************",
                                "Tuesday 24 February 2026  01:24:05 +0000 (0:00:00.072)       0:00:07.545 ****** ",
                                "",
                                "TASK [Print report file information] *******************************************",
                                "Tuesday 24 February 2026  01:24:06 +0000 (0:00:00.087)       0:00:07.632 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [Fail due to missing containers] ******************************************",
                                "Tuesday 24 February 2026  01:24:06 +0000 (0:00:00.274)       0:00:07.906 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [Define mgr module test vars] *********************************************",
                                "Tuesday 24 February 2026  01:24:06 +0000 (0:00:00.255)       0:00:08.162 ****** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [Gather list of mgr modules] **********************************************",
                                "Tuesday 24 February 2026  01:24:06 +0000 (0:00:00.121)       0:00:08.283 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [Parse mgr module list from json] *****************************************",
                                "Tuesday 24 February 2026  01:24:08 +0000 (0:00:02.118)       0:00:10.402 ****** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [Extract list of enabled mgr modules] *************************************",
                                "Tuesday 24 February 2026  01:24:09 +0000 (0:00:00.514)       0:00:10.917 ****** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [Fail test if mgr modules are disabled that should be enabled] ************",
                                "Tuesday 24 February 2026  01:24:09 +0000 (0:00:00.372)       0:00:11.289 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [Pass test if required mgr modules are enabled] ***************************",
                                "Tuesday 24 February 2026  01:24:09 +0000 (0:00:00.155)       0:00:11.444 ****** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [Set validation result to passed if no test failed] ***********************",
                                "Tuesday 24 February 2026  01:24:10 +0000 (0:00:00.181)       0:00:11.626 ****** ",
                                "\u001b[0;32mok: [testbed-node-0 -> testbed-manager(192.168.16.5)]\u001b[0m",
                                "",
                                "TASK [Set validation result to failed if a test failed] ************************",
                                "Tuesday 24 February 2026  01:24:10 +0000 (0:00:00.284)       0:00:11.910 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "",
                                "TASK [Aggregate test results step one] *****************************************",
                                "Tuesday 24 February 2026  01:24:10 +0000 (0:00:00.323)       0:00:12.233 ****** ",
                                "\u001b[0;32mok: [testbed-node-0 -> testbed-manager(192.168.16.5)]\u001b[0m",
                                "",
                                "TASK [Aggregate test results step two] *****************************************",
                                "Tuesday 24 February 2026  01:24:12 +0000 (0:00:01.460)       0:00:13.694 ****** ",
                                "\u001b[0;32mok: [testbed-node-0 -> testbed-manager(192.168.16.5)]\u001b[0m",
                                "",
                                "TASK [Aggregate test results step three] ***************************************",
                                "Tuesday 24 February 2026  01:24:12 +0000 (0:00:00.311)       0:00:14.005 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0 -> testbed-manager(192.168.16.5)]\u001b[0m",
                                "",
                                "TASK [Flush handlers] **********************************************************",
                                "Tuesday 24 February 2026  01:24:12 +0000 (0:00:00.292)       0:00:14.298 ****** ",
                                "",
                                "TASK [Flush handlers] **********************************************************",
                                "Tuesday 24 February 2026  01:24:12 +0000 (0:00:00.089)       0:00:14.387 ****** ",
                                "",
                                "TASK [Flush handlers] **********************************************************",
                                "Tuesday 24 February 2026  01:24:12 +0000 (0:00:00.073)       0:00:14.461 ****** ",
                                "",
                                "RUNNING HANDLER [Write report file] ********************************************",
                                "Tuesday 24 February 2026  01:24:13 +0000 (0:00:00.372)       0:00:14.834 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-0 -> testbed-manager(192.168.16.5)]\u001b[0m",
                                "",
                                "TASK [Print report file information] *******************************************",
                                "Tuesday 24 February 2026  01:24:14 +0000 (0:00:01.575)       0:00:16.409 ****** ",
                                "\u001b[0;32mok: [testbed-node-0 -> testbed-manager(192.168.16.5)] => {\u001b[0m",
                                "\u001b[0;32m    \"msg\": [\u001b[0m",
                                "\u001b[0;32m        \"Validator run completed.\",\u001b[0m",
                                "\u001b[0;32m        \"You can find the report file here:\",\u001b[0m",
                                "\u001b[0;32m        \"/opt/reports/validator/ceph-mgrs-validator-2026-02-24T01:23:59+00:00-report.json\",\u001b[0m",
                                "\u001b[0;32m        \"on the following host:\",\u001b[0m",
                                "\u001b[0;32m        \"testbed-manager\"\u001b[0m",
                                "\u001b[0;32m    ]\u001b[0m",
                                "\u001b[0;32m}\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=19  \u001b[0m \u001b[0;33mchanged=3   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=9   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;32mtestbed-node-1\u001b[0m             : \u001b[0;32mok=5   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;32mtestbed-node-2\u001b[0m             : \u001b[0;32mok=5   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Tuesday 24 February 2026  01:24:15 +0000 (0:00:00.485)       0:00:16.895 ****** ",
                                "=============================================================================== ",
                                "Gather list of mgr modules ---------------------------------------------- 2.12s",
                                "Write report file ------------------------------------------------------- 1.58s",
                                "Aggregate test results step one ----------------------------------------- 1.46s",
                                "Create report output directory ------------------------------------------ 1.23s",
                                "Get container info ------------------------------------------------------ 1.09s",
                                "Get timestamp for report file ------------------------------------------- 1.04s",
                                "Set test result to passed if container is existing ---------------------- 0.60s",
                                "Set test result to passed if ceph-mgr is running ------------------------ 0.55s",
                                "Flush handlers ---------------------------------------------------------- 0.54s",
                                "Parse mgr module list from json ----------------------------------------- 0.51s",
                                "Print report file information ------------------------------------------- 0.49s",
                                "Extract list of enabled mgr modules ------------------------------------- 0.37s",
                                "Prepare test data for container existance test -------------------------- 0.35s",
                                "Prepare test data ------------------------------------------------------- 0.34s",
                                "Set test result to failed if ceph-mgr is not running -------------------- 0.33s",
                                "Set test result to failed if container is missing ----------------------- 0.33s",
                                "Set validation result to failed if a test failed ------------------------ 0.32s",
                                "Aggregate test results step two ----------------------------------------- 0.31s",
                                "Aggregate test results step one ----------------------------------------- 0.31s",
                                "Aggregate test results step three --------------------------------------- 0.29s",
                                "+ osism validate ceph-osds",
                                "",
                                "PLAY [Ceph validate OSDs] ******************************************************",
                                "",
                                "TASK [Get timestamp for report file] *******************************************",
                                "Tuesday 24 February 2026  01:24:33 +0000 (0:00:00.439)       0:00:00.439 ****** ",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-manager(192.168.16.5)]\u001b[0m",
                                "",
                                "TASK [Get extra vars for Ceph configuration] ***********************************",
                                "Tuesday 24 February 2026  01:24:34 +0000 (0:00:00.766)       0:00:01.205 ****** ",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-manager(192.168.16.5)]\u001b[0m",
                                "",
                                "TASK [Create report output directory] ******************************************",
                                "Tuesday 24 February 2026  01:24:34 +0000 (0:00:00.457)       0:00:01.663 ****** ",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-manager(192.168.16.5)]\u001b[0m",
                                "",
                                "TASK [Define report vars] ******************************************************",
                                "Tuesday 24 February 2026  01:24:35 +0000 (0:00:00.733)       0:00:02.397 ****** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Define OSD test variables] ***********************************************",
                                "Tuesday 24 February 2026  01:24:35 +0000 (0:00:00.154)       0:00:02.551 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Calculate OSD devices for each host] *************************************",
                                "Tuesday 24 February 2026  01:24:35 +0000 (0:00:00.130)       0:00:02.681 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Define OSD test variables] ***********************************************",
                                "Tuesday 24 February 2026  01:24:36 +0000 (0:00:00.336)       0:00:03.018 ****** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Calculate OSD devices for each host] *************************************",
                                "Tuesday 24 February 2026  01:24:36 +0000 (0:00:00.171)       0:00:03.190 ****** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Calculate total number of OSDs in cluster] *******************************",
                                "Tuesday 24 February 2026  01:24:36 +0000 (0:00:00.412)       0:00:03.602 ****** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Prepare test data] *******************************************************",
                                "Tuesday 24 February 2026  01:24:37 +0000 (0:00:00.979)       0:00:04.582 ****** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Get list of ceph-osd containers on host] *********************************",
                                "Tuesday 24 February 2026  01:24:38 +0000 (0:00:00.343)       0:00:04.925 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'id': 'a507e39101423fe0593a98563102285dee1b9fd5e4e5d6a1d563e1f610290f5c', 'image': 'registry.osism.tech/kolla/nova-compute:2024.2', 'name': '/nova_compute', 'state': 'running', 'status': 'Up 7 minutes (healthy)'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'id': '15f1497f24ccac119e571c3ca4ed4919a800516c5394fe9fa749128c8c905fea', 'image': 'registry.osism.tech/kolla/nova-libvirt:2024.2', 'name': '/nova_libvirt', 'state': 'running', 'status': 'Up 7 minutes (healthy)'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'id': 'e2e543e63c1632d677daa97c5cbf37d6ea88a08477e7bc52d6aee09aadb8e6db', 'image': 'registry.osism.tech/kolla/nova-ssh:2024.2', 'name': '/nova_ssh', 'state': 'running', 'status': 'Up 8 minutes (healthy)'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'id': '4311ed30834d73e01f23c87613b03c0e1bd426d330434ac4d2c4d49561ff357d', 'image': 'registry.osism.tech/kolla/prometheus-libvirt-exporter:2024.2', 'name': '/prometheus_libvirt_exporter', 'state': 'running', 'status': 'Up 13 minutes'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'id': '8522e1cfe839b6a8a827f4a855f7c976c62d9037c614a70a5d252a6fb85efab0', 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2024.2', 'name': '/prometheus_cadvisor', 'state': 'running', 'status': 'Up 13 minutes'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'id': 'ef9cefa9065cb8fc86482d2449bb51824e0ef00be0eab0e05343b55eb0f777aa', 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2024.2', 'name': '/prometheus_node_exporter', 'state': 'running', 'status': 'Up 14 minutes'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'id': '6f45def46706eada0c2405777dcccf6426aed54715f85d10cd9f8e7453251571', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'name': '/neutron_ovn_metadata_agent', 'state': 'running', 'status': 'Up 15 minutes (healthy)'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'id': '4312c9fd5889f17b2cb033af2d2a0b48880ec04b3615d5a47f26b51a8883f40f', 'image': 'registry.osism.tech/osism/ceph-daemon:reef', 'name': '/ceph-rgw-default-testbed-node-3-rgw0', 'state': 'running', 'status': 'Up 24 minutes'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'id': '9c4e10611bcd267b7394da6deb0a1c6ee24c1bef4a599ad52f10da4eec0d178b', 'image': 'registry.osism.tech/osism/ceph-daemon:reef', 'name': '/ceph-mds-testbed-node-3', 'state': 'running', 'status': 'Up 25 minutes'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'id': '89ba6407e0e4d76b96f76ae5868d3f4282281fb62ff7b462fc5a6e487127dd67', 'image': 'registry.osism.tech/osism/ceph-daemon:reef', 'name': '/ceph-crash-testbed-node-3', 'state': 'running', 'status': 'Up 26 minutes'}) \u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item={'id': '922f562c3e4e19031c8fbcd3924f765bc113be0d1c47548542160560d2fb29e3', 'image': 'registry.osism.tech/osism/ceph-daemon:reef', 'name': '/ceph-osd-4', 'state': 'running', 'status': 'Up 27 minutes'})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-3] => (item={'id': 'c03474f234450bddc2a0c9477ef1b7f0f4d0652f44427646b62c4249df7b2471', 'image': 'registry.osism.tech/osism/ceph-daemon:reef', 'name': '/ceph-osd-0', 'state': 'running', 'status': 'Up 27 minutes'})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'id': 'bf8e81e60e6eab0f3a4d0768532240dda904f417d047f3554ef601548d2b5a69', 'image': 'registry.osism.tech/kolla/ovn-controller:2024.2', 'name': '/ovn_controller', 'state': 'running', 'status': 'Up 31 minutes'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'id': '49db629377e40e37c32d4b4436dc13c468c1a978eca01e92607906d576581f3b', 'image': 'registry.osism.tech/kolla/openvswitch-vswitchd:2024.2', 'name': '/openvswitch_vswitchd', 'state': 'running', 'status': 'Up 32 minutes (healthy)'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'id': '996e6fbdfb74a64d1997bbb0740db739ff64078a51f72ec4a4ea33d7d8777b6c', 'image': 'registry.osism.tech/kolla/openvswitch-db-server:2024.2', 'name': '/openvswitch_db', 'state': 'running', 'status': 'Up 32 minutes (healthy)'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'id': '4801bd0b79b2061b5696fea9da6800affdedc6589db051c15f03f6d6b7503835', 'image': 'registry.osism.tech/kolla/cron:2024.2', 'name': '/cron', 'state': 'running', 'status': 'Up 33 minutes'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'id': '995e066d70d5885bcf887baeb9bf278c6f94e406c388c550b49f2e9f0ba5cb57', 'image': 'registry.osism.tech/kolla/kolla-toolbox:2024.2', 'name': '/kolla_toolbox', 'state': 'running', 'status': 'Up 33 minutes'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'id': '67f9710f4059105a2661569001496eb30077fafb374c8d5c28faaf4d5b9c3bc8', 'image': 'registry.osism.tech/kolla/fluentd:2024.2', 'name': '/fluentd', 'state': 'running', 'status': 'Up 34 minutes'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'id': '51e654a6545de29f8704424e2295db8cfe2186daf0112420d21978a218a4c137', 'image': 'registry.osism.tech/kolla/nova-compute:2024.2', 'name': '/nova_compute', 'state': 'running', 'status': 'Up 7 minutes (healthy)'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'id': 'c4c18371cdda7ff45dbf717e5d8105a59a8a8ba34b2476726087b395be68028f', 'image': 'registry.osism.tech/kolla/nova-libvirt:2024.2', 'name': '/nova_libvirt', 'state': 'running', 'status': 'Up 7 minutes (healthy)'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'id': '1ae7fc12531decf870cbc9bd51d380c679b4c44d3793b563313a86448871568e', 'image': 'registry.osism.tech/kolla/nova-ssh:2024.2', 'name': '/nova_ssh', 'state': 'running', 'status': 'Up 8 minutes (healthy)'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'id': '56a6fbeeda2fd1543c68e86dd827b240bd56da9da3fe640562a3582929753412', 'image': 'registry.osism.tech/kolla/prometheus-libvirt-exporter:2024.2', 'name': '/prometheus_libvirt_exporter', 'state': 'running', 'status': 'Up 13 minutes'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'id': '3d69e1cda5c13feb1b014dbdef99f1fcd9bbbccedfde2828586da0e597984609', 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2024.2', 'name': '/prometheus_cadvisor', 'state': 'running', 'status': 'Up 13 minutes'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'id': '000daa7f8fc0b537a11690f4d5b496a3209b7ca9097e55b7f512a96dc85df651', 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2024.2', 'name': '/prometheus_node_exporter', 'state': 'running', 'status': 'Up 14 minutes'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'id': 'd80759bb7fde2e88bb7ccff304e84416534f89ac07b37bfd4a6a6a86edd73a4a', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'name': '/neutron_ovn_metadata_agent', 'state': 'running', 'status': 'Up 15 minutes (healthy)'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'id': '28c367c08ce47d0c2e99bf9b57930ba34a58a39e7314a02f5966c7c168e6d688', 'image': 'registry.osism.tech/osism/ceph-daemon:reef', 'name': '/ceph-rgw-default-testbed-node-4-rgw0', 'state': 'running', 'status': 'Up 24 minutes'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'id': 'b7bd5696edbb4aab30ecda8e91d4fed8bb8b984bf0564e7f9ad3599fb86d3e11', 'image': 'registry.osism.tech/osism/ceph-daemon:reef', 'name': '/ceph-mds-testbed-node-4', 'state': 'running', 'status': 'Up 25 minutes'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'id': '1a44500c43c872bae8490a0656dccfd2f55b0e5cf94ab759e0063df75c1dbaa9', 'image': 'registry.osism.tech/osism/ceph-daemon:reef', 'name': '/ceph-crash-testbed-node-4', 'state': 'running', 'status': 'Up 26 minutes'}) \u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item={'id': '550dd349a35c2047ccf0a20cce01b054274e5147959a93ac2764c6642ef9d171', 'image': 'registry.osism.tech/osism/ceph-daemon:reef', 'name': '/ceph-osd-1', 'state': 'running', 'status': 'Up 27 minutes'})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4] => (item={'id': '914dde566f6c2bd3ee6bdd875e0979ef10f6f9583775e528ae1419f54cf48d77', 'image': 'registry.osism.tech/osism/ceph-daemon:reef', 'name': '/ceph-osd-5', 'state': 'running', 'status': 'Up 27 minutes'})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'id': '2257a99720c614cd19bf2235233dbac7df752935f30cfab15734ba0094276beb', 'image': 'registry.osism.tech/kolla/ovn-controller:2024.2', 'name': '/ovn_controller', 'state': 'running', 'status': 'Up 31 minutes'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'id': '1b73c76cbb63c8895690c0326076d5e803c9db53c64ee63c5394617d9fa58f78', 'image': 'registry.osism.tech/kolla/openvswitch-vswitchd:2024.2', 'name': '/openvswitch_vswitchd', 'state': 'running', 'status': 'Up 32 minutes (healthy)'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'id': '023e8a49ea84a98603face731bc8acae8c39ce5b679dccf7d971701da87f78d0', 'image': 'registry.osism.tech/kolla/openvswitch-db-server:2024.2', 'name': '/openvswitch_db', 'state': 'running', 'status': 'Up 32 minutes (healthy)'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'id': '3de1fb6f795eb0615d2c83b052d202be3f8549da3010f5851552db141df00893', 'image': 'registry.osism.tech/kolla/cron:2024.2', 'name': '/cron', 'state': 'running', 'status': 'Up 33 minutes'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'id': '528140ec3692082ea8f19d80356850ae605f1c70428f61b3f0d7e3cd3b9c4987', 'image': 'registry.osism.tech/kolla/kolla-toolbox:2024.2', 'name': '/kolla_toolbox', 'state': 'running', 'status': 'Up 33 minutes'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'id': 'e22fda3dfc45e78d5c2ead3e48f54abd1a03f359e55b1266341f7ea655e81457', 'image': 'registry.osism.tech/kolla/fluentd:2024.2', 'name': '/fluentd', 'state': 'running', 'status': 'Up 34 minutes'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'id': '9bcf861975e2d45d6534e072b676ef8f1326e2b533be31015f681afdd8de4ec8', 'image': 'registry.osism.tech/kolla/nova-compute:2024.2', 'name': '/nova_compute', 'state': 'running', 'status': 'Up 7 minutes (healthy)'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'id': '767dfe89db6361c0e18e57ebb9f64becb7d61836dd5c888c9b1d41f751ba69c8', 'image': 'registry.osism.tech/kolla/nova-libvirt:2024.2', 'name': '/nova_libvirt', 'state': 'running', 'status': 'Up 7 minutes (healthy)'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'id': '670d1815146d1f4cffdb4ff528e7b32e8177f2740870d7edb31a73035ead5b6c', 'image': 'registry.osism.tech/kolla/nova-ssh:2024.2', 'name': '/nova_ssh', 'state': 'running', 'status': 'Up 8 minutes (healthy)'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'id': '54d8627050388ea00f23210a5c17ba02818a23388e551196c8a36d0cce5e581c', 'image': 'registry.osism.tech/kolla/prometheus-libvirt-exporter:2024.2', 'name': '/prometheus_libvirt_exporter', 'state': 'running', 'status': 'Up 13 minutes'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'id': '492ed32d7a2d24a33b9974da752fb5738700b0dafdcbf9f84e5d3ac8621b44a3', 'image': 'registry.osism.tech/kolla/prometheus-cadvisor:2024.2', 'name': '/prometheus_cadvisor', 'state': 'running', 'status': 'Up 13 minutes'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'id': '71982626b1d88f167ce4c025cccdd5b3fa714c50224c77cc9ac8510773040646', 'image': 'registry.osism.tech/kolla/prometheus-node-exporter:2024.2', 'name': '/prometheus_node_exporter', 'state': 'running', 'status': 'Up 14 minutes'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'id': '69bd630a540ff812d7b1d0cdac1f044a77c92f56c71e3f38cb704883a3ff27b5', 'image': 'registry.osism.tech/kolla/neutron-metadata-agent:2024.2', 'name': '/neutron_ovn_metadata_agent', 'state': 'running', 'status': 'Up 15 minutes (healthy)'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'id': '70034c20eb20d8d23f2a3a5361d234f36ac40a53065542d70db5dd74d0160900', 'image': 'registry.osism.tech/osism/ceph-daemon:reef', 'name': '/ceph-rgw-default-testbed-node-5-rgw0', 'state': 'running', 'status': 'Up 24 minutes'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'id': '81ca12344dd172735a5039a9a3401bb6ec32982f379782396844d65d3af4495d', 'image': 'registry.osism.tech/osism/ceph-daemon:reef', 'name': '/ceph-mds-testbed-node-5', 'state': 'running', 'status': 'Up 25 minutes'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'id': '83cad903f15d3dc2fc3fd10d328830b4be77538bafc32efd7a4af313d66b73a6', 'image': 'registry.osism.tech/osism/ceph-daemon:reef', 'name': '/ceph-crash-testbed-node-5', 'state': 'running', 'status': 'Up 26 minutes'}) \u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item={'id': 'df66acd5ab8adda5830fc22aabf350b2c0de0776b4d53792f54ba39d4a53c4d2', 'image': 'registry.osism.tech/osism/ceph-daemon:reef', 'name': '/ceph-osd-2', 'state': 'running', 'status': 'Up 27 minutes'})\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5] => (item={'id': '1f23e7a3653884d283e6addfab83116f386c1e6a00e715e33c4bf8532d43e904', 'image': 'registry.osism.tech/osism/ceph-daemon:reef', 'name': '/ceph-osd-3', 'state': 'running', 'status': 'Up 27 minutes'})\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'id': 'fcaa772f75e20cec07ace286cfc1e6510a1f29a14f2b41fdf171038597901555', 'image': 'registry.osism.tech/kolla/ovn-controller:2024.2', 'name': '/ovn_controller', 'state': 'running', 'status': 'Up 31 minutes'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'id': 'ec81ef1b068903742aabc86658c38a8ddaec66bfe292ce40d4003dce62ca6137', 'image': 'registry.osism.tech/kolla/openvswitch-vswitchd:2024.2', 'name': '/openvswitch_vswitchd', 'state': 'running', 'status': 'Up 32 minutes (healthy)'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'id': '8bbdc8d1907348d07d4f3ce135a14d7288deef5386d718df5a4f0eccc2818f2f', 'image': 'registry.osism.tech/kolla/openvswitch-db-server:2024.2', 'name': '/openvswitch_db', 'state': 'running', 'status': 'Up 32 minutes (healthy)'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'id': '2f661e4839110de787c96265a8d88854e3c7cb0704897771bce4501fed471a65', 'image': 'registry.osism.tech/kolla/cron:2024.2', 'name': '/cron', 'state': 'running', 'status': 'Up 33 minutes'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'id': 'b9edb24895fb63536299193e977a3fb921cad196014b280edb3a908c2ce88bfc', 'image': 'registry.osism.tech/kolla/kolla-toolbox:2024.2', 'name': '/kolla_toolbox', 'state': 'running', 'status': 'Up 33 minutes'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'id': 'c3d85964fcbea496d2ab6b2cbe7e4b9729bc65ece7016929d5cd7f688c6c08de', 'image': 'registry.osism.tech/kolla/fluentd:2024.2', 'name': '/fluentd', 'state': 'running', 'status': 'Up 34 minutes'}) \u001b[0m",
                                "",
                                "TASK [Get count of ceph-osd containers on host] ********************************",
                                "Tuesday 24 February 2026  01:24:38 +0000 (0:00:00.522)       0:00:05.447 ****** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Set test result to failed when count of containers is wrong] *************",
                                "Tuesday 24 February 2026  01:24:38 +0000 (0:00:00.315)       0:00:05.763 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Set test result to passed if count matches] ******************************",
                                "Tuesday 24 February 2026  01:24:39 +0000 (0:00:00.534)       0:00:06.297 ****** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Prepare test data] *******************************************************",
                                "Tuesday 24 February 2026  01:24:39 +0000 (0:00:00.337)       0:00:06.635 ****** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Get list of ceph-osd containers that are not running] ********************",
                                "Tuesday 24 February 2026  01:24:40 +0000 (0:00:00.332)       0:00:06.967 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'name': 'ceph-osd-4', 'osd_id': '4', 'state': 'running'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3] => (item={'name': 'ceph-osd-0', 'osd_id': '0', 'state': 'running'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'name': 'ceph-osd-1', 'osd_id': '1', 'state': 'running'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4] => (item={'name': 'ceph-osd-5', 'osd_id': '5', 'state': 'running'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'name': 'ceph-osd-2', 'osd_id': '2', 'state': 'running'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5] => (item={'name': 'ceph-osd-3', 'osd_id': '3', 'state': 'running'}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Get count of ceph-osd containers that are not running] *******************",
                                "Tuesday 24 February 2026  01:24:40 +0000 (0:00:00.369)       0:00:07.337 ****** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Set test result to failed if an OSD is not running] **********************",
                                "Tuesday 24 February 2026  01:24:41 +0000 (0:00:00.577)       0:00:07.915 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Set test result to failed if an OSD is not running] **********************",
                                "Tuesday 24 February 2026  01:24:41 +0000 (0:00:00.327)       0:00:08.242 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Set test result to passed if all containers are running] *****************",
                                "Tuesday 24 February 2026  01:24:41 +0000 (0:00:00.315)       0:00:08.558 ****** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Aggregate test results step one] *****************************************",
                                "Tuesday 24 February 2026  01:24:42 +0000 (0:00:00.421)       0:00:08.980 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Aggregate test results step two] *****************************************",
                                "Tuesday 24 February 2026  01:24:42 +0000 (0:00:00.777)       0:00:09.757 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Aggregate test results step three] ***************************************",
                                "Tuesday 24 February 2026  01:24:43 +0000 (0:00:00.289)       0:00:10.047 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Flush handlers] **********************************************************",
                                "Tuesday 24 February 2026  01:24:43 +0000 (0:00:00.321)       0:00:10.368 ****** ",
                                "",
                                "TASK [Flush handlers] **********************************************************",
                                "Tuesday 24 February 2026  01:24:43 +0000 (0:00:00.094)       0:00:10.462 ****** ",
                                "",
                                "TASK [Flush handlers] **********************************************************",
                                "Tuesday 24 February 2026  01:24:43 +0000 (0:00:00.072)       0:00:10.535 ****** ",
                                "",
                                "TASK [Print report file information] *******************************************",
                                "Tuesday 24 February 2026  01:24:43 +0000 (0:00:00.074)       0:00:10.609 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Fail early due to containers not running] ********************************",
                                "Tuesday 24 February 2026  01:24:44 +0000 (0:00:00.308)       0:00:10.918 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Prepare test data] *******************************************************",
                                "Tuesday 24 February 2026  01:24:44 +0000 (0:00:00.273)       0:00:11.191 ****** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Set _mon_hostname fact] **************************************************",
                                "Tuesday 24 February 2026  01:24:44 +0000 (0:00:00.317)       0:00:11.508 ****** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Get ceph osd tree] *******************************************************",
                                "Tuesday 24 February 2026  01:24:45 +0000 (0:00:00.978)       0:00:12.487 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-3 -> testbed-node-0(192.168.16.10)]\u001b[0m",
                                "",
                                "TASK [Parse osd tree from JSON] ************************************************",
                                "Tuesday 24 February 2026  01:24:47 +0000 (0:00:01.850)       0:00:14.337 ****** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Get OSDs that are not up or in] ******************************************",
                                "Tuesday 24 February 2026  01:24:47 +0000 (0:00:00.152)       0:00:14.490 ****** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Fail test if OSDs are not up or in] **************************************",
                                "Tuesday 24 February 2026  01:24:48 +0000 (0:00:00.373)       0:00:14.863 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Pass test if OSDs are all up and in] *************************************",
                                "Tuesday 24 February 2026  01:24:48 +0000 (0:00:00.139)       0:00:15.003 ****** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Prepare test data] *******************************************************",
                                "Tuesday 24 February 2026  01:24:48 +0000 (0:00:00.167)       0:00:15.171 ****** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [List ceph LVM volumes and collect data] **********************************",
                                "Tuesday 24 February 2026  01:24:48 +0000 (0:00:00.336)       0:00:15.507 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-3]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-4]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Parse LVM data as JSON] **************************************************",
                                "Tuesday 24 February 2026  01:24:51 +0000 (0:00:02.906)       0:00:18.414 ****** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Get unencrypted and encrypted OSDs] **************************************",
                                "Tuesday 24 February 2026  01:24:51 +0000 (0:00:00.356)       0:00:18.770 ****** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Fail if count of encrypted OSDs does not match] **************************",
                                "Tuesday 24 February 2026  01:24:52 +0000 (0:00:00.569)       0:00:19.339 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Pass if count of encrypted OSDs equals count of OSDs] ********************",
                                "Tuesday 24 February 2026  01:24:52 +0000 (0:00:00.357)       0:00:19.697 ****** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Fail if count of unencrypted OSDs does not match] ************************",
                                "Tuesday 24 February 2026  01:24:53 +0000 (0:00:00.665)       0:00:20.363 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Pass if count of unencrypted OSDs equals count of OSDs] ******************",
                                "Tuesday 24 February 2026  01:24:53 +0000 (0:00:00.327)       0:00:20.690 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Prepare test data] *******************************************************",
                                "Tuesday 24 February 2026  01:24:54 +0000 (0:00:00.368)       0:00:21.059 ****** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Get CRUSH node data of each OSD host and root node childs] ***************",
                                "Tuesday 24 February 2026  01:24:54 +0000 (0:00:00.591)       0:00:21.650 ****** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Calculate sub test expression results] ***********************************",
                                "Tuesday 24 February 2026  01:24:55 +0000 (0:00:00.963)       0:00:22.614 ****** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Fail test if any sub test failed] ****************************************",
                                "Tuesday 24 February 2026  01:24:56 +0000 (0:00:00.361)       0:00:22.976 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-4]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Pass test if no sub test failed] *****************************************",
                                "Tuesday 24 February 2026  01:24:56 +0000 (0:00:00.369)       0:00:23.345 ****** ",
                                "\u001b[0;32mok: [testbed-node-3]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-4]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-5]\u001b[0m",
                                "",
                                "TASK [Set validation result to passed if no test failed] ***********************",
                                "Tuesday 24 February 2026  01:24:57 +0000 (0:00:00.577)       0:00:23.923 ****** ",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-manager(192.168.16.5)]\u001b[0m",
                                "",
                                "TASK [Set validation result to failed if a test failed] ************************",
                                "Tuesday 24 February 2026  01:24:57 +0000 (0:00:00.330)       0:00:24.253 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-3]\u001b[0m",
                                "",
                                "TASK [Aggregate test results step one] *****************************************",
                                "Tuesday 24 February 2026  01:24:57 +0000 (0:00:00.293)       0:00:24.547 ****** ",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-manager(192.168.16.5)]\u001b[0m",
                                "",
                                "TASK [Aggregate test results step two] *****************************************",
                                "Tuesday 24 February 2026  01:24:59 +0000 (0:00:01.883)       0:00:26.430 ****** ",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-manager(192.168.16.5)]\u001b[0m",
                                "",
                                "TASK [Aggregate test results step three] ***************************************",
                                "Tuesday 24 February 2026  01:24:59 +0000 (0:00:00.325)       0:00:26.756 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-3 -> testbed-manager(192.168.16.5)]\u001b[0m",
                                "",
                                "TASK [Flush handlers] **********************************************************",
                                "Tuesday 24 February 2026  01:25:00 +0000 (0:00:00.308)       0:00:27.065 ****** ",
                                "",
                                "TASK [Flush handlers] **********************************************************",
                                "Tuesday 24 February 2026  01:25:00 +0000 (0:00:00.089)       0:00:27.154 ****** ",
                                "",
                                "TASK [Flush handlers] **********************************************************",
                                "Tuesday 24 February 2026  01:25:00 +0000 (0:00:00.090)       0:00:27.244 ****** ",
                                "",
                                "RUNNING HANDLER [Write report file] ********************************************",
                                "Tuesday 24 February 2026  01:25:00 +0000 (0:00:00.098)       0:00:27.342 ****** ",
                                "\u001b[0;33mchanged: [testbed-node-3 -> testbed-manager(192.168.16.5)]\u001b[0m",
                                "",
                                "TASK [Print report file information] *******************************************",
                                "Tuesday 24 February 2026  01:25:02 +0000 (0:00:02.074)       0:00:29.417 ****** ",
                                "\u001b[0;32mok: [testbed-node-3 -> testbed-manager(192.168.16.5)] => {\u001b[0m",
                                "\u001b[0;32m    \"msg\": [\u001b[0m",
                                "\u001b[0;32m        \"Validator run completed.\",\u001b[0m",
                                "\u001b[0;32m        \"You can find the report file here:\",\u001b[0m",
                                "\u001b[0;32m        \"/opt/reports/validator/ceph-osds-validator-2026-02-24T01:24:34+00:00-report.json\",\u001b[0m",
                                "\u001b[0;32m        \"on the following host:\",\u001b[0m",
                                "\u001b[0;32m        \"testbed-manager\"\u001b[0m",
                                "\u001b[0;32m    ]\u001b[0m",
                                "\u001b[0;32m}\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;33mtestbed-node-3\u001b[0m             : \u001b[0;32mok=35  \u001b[0m \u001b[0;33mchanged=4   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=17  \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-4\u001b[0m             : \u001b[0;32mok=18  \u001b[0m \u001b[0;33mchanged=1   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=9   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;33mtestbed-node-5\u001b[0m             : \u001b[0;32mok=18  \u001b[0m \u001b[0;33mchanged=1   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=9   \u001b[0m rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Tuesday 24 February 2026  01:25:03 +0000 (0:00:00.745)       0:00:30.162 ****** ",
                                "=============================================================================== ",
                                "List ceph LVM volumes and collect data ---------------------------------- 2.91s",
                                "Write report file ------------------------------------------------------- 2.07s",
                                "Aggregate test results step one ----------------------------------------- 1.88s",
                                "Get ceph osd tree ------------------------------------------------------- 1.85s",
                                "Calculate total number of OSDs in cluster ------------------------------- 0.98s",
                                "Set _mon_hostname fact -------------------------------------------------- 0.98s",
                                "Get CRUSH node data of each OSD host and root node childs --------------- 0.96s",
                                "Aggregate test results step one ----------------------------------------- 0.78s",
                                "Get timestamp for report file ------------------------------------------- 0.77s",
                                "Print report file information ------------------------------------------- 0.75s",
                                "Create report output directory ------------------------------------------ 0.73s",
                                "Pass if count of encrypted OSDs equals count of OSDs -------------------- 0.67s",
                                "Prepare test data ------------------------------------------------------- 0.59s",
                                "Get count of ceph-osd containers that are not running ------------------- 0.58s",
                                "Pass test if no sub test failed ----------------------------------------- 0.58s",
                                "Get unencrypted and encrypted OSDs -------------------------------------- 0.57s",
                                "Set test result to failed when count of containers is wrong ------------- 0.53s",
                                "Get list of ceph-osd containers on host --------------------------------- 0.52s",
                                "Get extra vars for Ceph configuration ----------------------------------- 0.46s",
                                "Set test result to passed if all containers are running ----------------- 0.42s",
                                "+ sh -c /opt/configuration/scripts/check/200-infrastructure.sh",
                                "+ set -e",
                                "+ source /opt/manager-vars.sh",
                                "++ export NUMBER_OF_NODES=6",
                                "++ NUMBER_OF_NODES=6",
                                "++ export CEPH_VERSION=reef",
                                "++ CEPH_VERSION=reef",
                                "++ export CONFIGURATION_VERSION=main",
                                "++ CONFIGURATION_VERSION=main",
                                "++ export MANAGER_VERSION=latest",
                                "++ MANAGER_VERSION=latest",
                                "++ export OPENSTACK_VERSION=2024.2",
                                "++ OPENSTACK_VERSION=2024.2",
                                "++ export ARA=false",
                                "++ ARA=false",
                                "++ export DEPLOY_MODE=manager",
                                "++ DEPLOY_MODE=manager",
                                "++ export TEMPEST=true",
                                "++ TEMPEST=true",
                                "++ export IS_ZUUL=true",
                                "++ IS_ZUUL=true",
                                "++ export MANAGER_PUBLIC_IP_ADDRESS=81.163.192.195",
                                "++ MANAGER_PUBLIC_IP_ADDRESS=81.163.192.195",
                                "++ export EXTERNAL_API=false",
                                "++ EXTERNAL_API=false",
                                "++ export IMAGE_USER=ubuntu",
                                "++ IMAGE_USER=ubuntu",
                                "++ export IMAGE_NODE_USER=ubuntu",
                                "++ IMAGE_NODE_USER=ubuntu",
                                "++ export CEPH_STACK=ceph-ansible",
                                "++ CEPH_STACK=ceph-ansible",
                                "+ [[ -e /etc/redhat-release ]]",
                                "+ source /etc/os-release",
                                "++ PRETTY_NAME='Ubuntu 24.04.4 LTS'",
                                "++ NAME=Ubuntu",
                                "++ VERSION_ID=24.04",
                                "++ VERSION='24.04.4 LTS (Noble Numbat)'",
                                "++ VERSION_CODENAME=noble",
                                "++ ID=ubuntu",
                                "++ ID_LIKE=debian",
                                "++ HOME_URL=https://www.ubuntu.com/",
                                "++ SUPPORT_URL=https://help.ubuntu.com/",
                                "++ BUG_REPORT_URL=https://bugs.launchpad.net/ubuntu/",
                                "++ PRIVACY_POLICY_URL=https://www.ubuntu.com/legal/terms-and-policies/privacy-policy",
                                "++ UBUNTU_CODENAME=noble",
                                "++ LOGO=ubuntu-logo",
                                "+ [[ ubuntu == \\u\\b\\u\\n\\t\\u ]]",
                                "+ packages='libmonitoring-plugin-perl libwww-perl libjson-perl monitoring-plugins-basic mysql-client'",
                                "+ dpkg -s libmonitoring-plugin-perl libwww-perl libjson-perl monitoring-plugins-basic mysql-client",
                                "+ sudo apt-get install -y libmonitoring-plugin-perl libwww-perl libjson-perl monitoring-plugins-basic mysql-client",
                                "",
                                "# Status of Elasticsearch",
                                "",
                                "+ pushd /opt/configuration/contrib",
                                "+ echo",
                                "+ echo '# Status of Elasticsearch'",
                                "+ echo",
                                "+ bash nagios-plugins/check_elasticsearch -H api-int.testbed.osism.xyz -s",
                                "OK - elasticsearch (kolla_logging) is running. status: green; timed_out: false; number_of_nodes: 3; number_of_data_nodes: 3; active_primary_shards: 9; active_shards: 22; relocating_shards: 0; initializing_shards: 0; delayed_unassigned_shards: 0; unassigned_shards: 0  | 'active_primary'=9 'active'=22 'relocating'=0 'init'=0 'delay_unass'=0 'unass'=0",
                                "",
                                "# Status of MariaDB",
                                "",
                                "+ echo",
                                "+ echo '# Status of MariaDB'",
                                "+ echo",
                                "++ semver latest 10.0.0-0",
                                "+ [[ -1 -ge 0 ]]",
                                "+ [[ latest == \\l\\a\\t\\e\\s\\t ]]",
                                "+ osism status database",
                                "\u001b[32m2026-02-24 01:25:32\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mUnable to get ansible vault password\u001b[0m",
                                "\u001b[32m2026-02-24 01:25:32\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mUnable to get vault secret: [Errno 2] No such file or directory: '/share/ansible_vault_password.key'\u001b[0m",
                                "\u001b[32m2026-02-24 01:25:32\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mDropping encrypted entries\u001b[0m",
                                "\u001b[32m2026-02-24 01:25:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mConnecting to MariaDB at 192.168.16.9 as root_shard_0...\u001b[0m",
                                "\u001b[32m2026-02-24 01:25:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mCluster Status: Primary\u001b[0m",
                                "\u001b[32m2026-02-24 01:25:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mConnected: ON\u001b[0m",
                                "\u001b[32m2026-02-24 01:25:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mReady: ON\u001b[0m",
                                "\u001b[32m2026-02-24 01:25:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mCluster Size: 3\u001b[0m",
                                "\u001b[32m2026-02-24 01:25:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLocal State: Synced\u001b[0m",
                                "\u001b[32m2026-02-24 01:25:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mCluster State UUID: ce650e1c-111b-11f1-87f2-a37335181952\u001b[0m",
                                "\u001b[32m2026-02-24 01:25:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mCluster Members: 192.168.16.10:3306,192.168.16.11:3306,192.168.16.12:3306\u001b[0m",
                                "\u001b[32m2026-02-24 01:25:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mGalera Version: 26.4.25(r7387a566)\u001b[0m",
                                "\u001b[32m2026-02-24 01:25:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLocal Node UUID: 05d530b6-111c-11f1-88be-d640b4f94096\u001b[0m",
                                "\u001b[32m2026-02-24 01:25:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mFlow Control Paused: 0.00%\u001b[0m",
                                "\u001b[32m2026-02-24 01:25:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mRecv Queue Avg: 0\u001b[0m",
                                "\u001b[32m2026-02-24 01:25:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mSend Queue Avg: 0.000144238\u001b[0m",
                                "\u001b[32m2026-02-24 01:25:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTransactions: 4696 local commits, 6874 replicated, 120 received\u001b[0m",
                                "\u001b[32m2026-02-24 01:25:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mConflicts: 0 cert failures, 0 bf aborts\u001b[0m",
                                "\u001b[32m2026-02-24 01:25:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mMariaDB Uptime: 26 minutes, 25 seconds\u001b[0m",
                                "\u001b[32m2026-02-24 01:25:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mThreads: 125 connected, 1 running\u001b[0m",
                                "\u001b[32m2026-02-24 01:25:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mQueries: 147247 total, 0 slow\u001b[0m",
                                "\u001b[32m2026-02-24 01:25:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mAborted Connects: 148\u001b[0m",
                                "\u001b[32m2026-02-24 01:25:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mMariaDB Galera Cluster validation PASSED\u001b[0m",
                                "",
                                "# Status of Prometheus",
                                "",
                                "+ echo",
                                "+ echo '# Status of Prometheus'",
                                "+ echo",
                                "+ curl -s https://api-int.testbed.osism.xyz:9091/-/healthy",
                                "Unauthorized",
                                "+ curl -s https://api-int.testbed.osism.xyz:9091/-/ready",
                                "Unauthorized",
                                "",
                                "# Status of RabbitMQ",
                                "",
                                "+ echo",
                                "+ echo '# Status of RabbitMQ'",
                                "+ echo",
                                "++ semver latest 10.0.0-0",
                                "+ [[ -1 -ge 0 ]]",
                                "+ [[ latest == \\l\\a\\t\\e\\s\\t ]]",
                                "+ osism status messaging",
                                "\u001b[32m2026-02-24 01:25:57\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mUnable to get ansible vault password\u001b[0m",
                                "\u001b[32m2026-02-24 01:25:57\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mUnable to get vault secret: [Errno 2] No such file or directory: '/share/ansible_vault_password.key'\u001b[0m",
                                "\u001b[32m2026-02-24 01:25:57\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mDropping encrypted entries\u001b[0m",
                                "\u001b[32m2026-02-24 01:25:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-0] Connecting to RabbitMQ Management API at 192.168.16.10:15672 as openstack...\u001b[0m",
                                "\u001b[32m2026-02-24 01:25:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-0] RabbitMQ Version: 3.13.7\u001b[0m",
                                "\u001b[32m2026-02-24 01:25:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-0] Erlang Version: 26.2.5.15\u001b[0m",
                                "\u001b[32m2026-02-24 01:25:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-0] Cluster Name: rabbit@testbed-node-0\u001b[0m",
                                "\u001b[32m2026-02-24 01:25:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-0] Cluster Size: 3\u001b[0m",
                                "\u001b[32m2026-02-24 01:25:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-0] Nodes: rabbit@testbed-node-0, rabbit@testbed-node-1, rabbit@testbed-node-2\u001b[0m",
                                "\u001b[32m2026-02-24 01:25:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-0] Running Nodes: rabbit@testbed-node-0, rabbit@testbed-node-1, rabbit@testbed-node-2\u001b[0m",
                                "\u001b[32m2026-02-24 01:25:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-0] Partitions: None (healthy)\u001b[0m",
                                "\u001b[32m2026-02-24 01:25:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-0] Connections: 194, Channels: 193, Queues: 173\u001b[0m",
                                "\u001b[32m2026-02-24 01:25:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-0] Messages: 218 total, 218 ready, 0 unacked\u001b[0m",
                                "\u001b[32m2026-02-24 01:25:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-0] Message Rates: 6.6/s publish, 6.0/s deliver\u001b[0m",
                                "\u001b[32m2026-02-24 01:25:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-0] Disk Free: 58.2 GB (limit: 0.0 GB)\u001b[0m",
                                "\u001b[32m2026-02-24 01:25:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-0] Memory Used: 0.18 GB (limit: 12.54 GB)\u001b[0m",
                                "\u001b[32m2026-02-24 01:25:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-0] File Descriptors: 113/1024\u001b[0m",
                                "\u001b[32m2026-02-24 01:25:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-0] Sockets: 67/832\u001b[0m",
                                "\u001b[32m2026-02-24 01:25:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-1] Connecting to RabbitMQ Management API at 192.168.16.11:15672 as openstack...\u001b[0m",
                                "\u001b[32m2026-02-24 01:25:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-1] RabbitMQ Version: 3.13.7\u001b[0m",
                                "\u001b[32m2026-02-24 01:25:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-1] Erlang Version: 26.2.5.15\u001b[0m",
                                "\u001b[32m2026-02-24 01:25:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-1] Cluster Name: rabbit@testbed-node-1\u001b[0m",
                                "\u001b[32m2026-02-24 01:25:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-1] Cluster Size: 3\u001b[0m",
                                "\u001b[32m2026-02-24 01:25:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-1] Nodes: rabbit@testbed-node-0, rabbit@testbed-node-1, rabbit@testbed-node-2\u001b[0m",
                                "\u001b[32m2026-02-24 01:25:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-1] Running Nodes: rabbit@testbed-node-0, rabbit@testbed-node-1, rabbit@testbed-node-2\u001b[0m",
                                "\u001b[32m2026-02-24 01:25:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-1] Partitions: None (healthy)\u001b[0m",
                                "\u001b[32m2026-02-24 01:25:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-1] Connections: 194, Channels: 193, Queues: 173\u001b[0m",
                                "\u001b[32m2026-02-24 01:25:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-1] Messages: 218 total, 218 ready, 0 unacked\u001b[0m",
                                "\u001b[32m2026-02-24 01:25:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-1] Message Rates: 6.6/s publish, 6.0/s deliver\u001b[0m",
                                "\u001b[32m2026-02-24 01:25:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-1] Disk Free: 58.5 GB (limit: 0.0 GB)\u001b[0m",
                                "\u001b[32m2026-02-24 01:25:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-1] Memory Used: 0.18 GB (limit: 12.54 GB)\u001b[0m",
                                "\u001b[32m2026-02-24 01:25:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-1] File Descriptors: 119/1024\u001b[0m",
                                "\u001b[32m2026-02-24 01:25:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-1] Sockets: 71/832\u001b[0m",
                                "\u001b[32m2026-02-24 01:25:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-2] Connecting to RabbitMQ Management API at 192.168.16.12:15672 as openstack...\u001b[0m",
                                "\u001b[32m2026-02-24 01:25:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-2] RabbitMQ Version: 3.13.7\u001b[0m",
                                "\u001b[32m2026-02-24 01:25:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-2] Erlang Version: 26.2.5.15\u001b[0m",
                                "\u001b[32m2026-02-24 01:25:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-2] Cluster Name: rabbit@testbed-node-2\u001b[0m",
                                "\u001b[32m2026-02-24 01:25:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-2] Cluster Size: 3\u001b[0m",
                                "\u001b[32m2026-02-24 01:25:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-2] Nodes: rabbit@testbed-node-0, rabbit@testbed-node-1, rabbit@testbed-node-2\u001b[0m",
                                "\u001b[32m2026-02-24 01:25:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-2] Running Nodes: rabbit@testbed-node-0, rabbit@testbed-node-1, rabbit@testbed-node-2\u001b[0m",
                                "\u001b[32m2026-02-24 01:25:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-2] Partitions: None (healthy)\u001b[0m",
                                "\u001b[32m2026-02-24 01:25:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-2] Connections: 194, Channels: 193, Queues: 173\u001b[0m",
                                "\u001b[32m2026-02-24 01:25:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-2] Messages: 218 total, 218 ready, 0 unacked\u001b[0m",
                                "\u001b[32m2026-02-24 01:25:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-2] Message Rates: 6.6/s publish, 6.0/s deliver\u001b[0m",
                                "\u001b[32m2026-02-24 01:25:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-2] Disk Free: 58.5 GB (limit: 0.0 GB)\u001b[0m",
                                "\u001b[32m2026-02-24 01:25:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-2] Memory Used: 0.17 GB (limit: 12.54 GB)\u001b[0m",
                                "\u001b[32m2026-02-24 01:25:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-2] File Descriptors: 102/1024\u001b[0m",
                                "\u001b[32m2026-02-24 01:25:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1m[testbed-node-2] Sockets: 56/832\u001b[0m",
                                "\u001b[32m2026-02-24 01:25:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mRabbitMQ Cluster validation PASSED\u001b[0m",
                                "",
                                "# Status of Redis",
                                "",
                                "+ echo",
                                "+ echo '# Status of Redis'",
                                "+ echo",
                                "+ /usr/lib/nagios/plugins/check_tcp -H 192.168.16.10 -p 6379 -A -E -s 'AUTH QHNA1SZRlOKzLADhUd5ZDgpHfQe6dNfr3bwEdY24\\r\\nPING\\r\\nINFO replication\\r\\nQUIT\\r\\n' -e PONG -e role:master -e slave0:ip=192.168.16.1 -e,port=6379 -j",
                                "TCP OK - 0.002 second response time on 192.168.16.10 port 6379|time=0.002036s;;;0.000000;10.000000",
                                "",
                                "# Create backup of MariaDB database",
                                "",
                                "+ popd",
                                "+ echo",
                                "+ echo '# Create backup of MariaDB database'",
                                "+ echo",
                                "+ osism apply mariadb_backup -e mariadb_backup_type=full",
                                "\u001b[32m2026-02-24 01:26:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mPrepare task for execution of mariadb_backup.\u001b[0m",
                                "\u001b[32m2026-02-24 01:26:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 00dcac0d-65ca-4e85-892e-f92ee847cc55 (mariadb_backup) was prepared for execution.\u001b[0m",
                                "\u001b[32m2026-02-24 01:26:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task 00dcac0d-65ca-4e85-892e-f92ee847cc55 (mariadb_backup) has been started and output is visible here.\u001b[0m",
                                "",
                                "PLAY [Group hosts based on configuration] **************************************",
                                "",
                                "TASK [Group hosts based on Kolla action] ***************************************",
                                "Tuesday 24 February 2026  01:26:05 +0000 (0:00:00.210)       0:00:00.210 ****** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [Group hosts based on enabled services] ***********************************",
                                "Tuesday 24 February 2026  01:26:05 +0000 (0:00:00.362)       0:00:00.572 ****** ",
                                "\u001b[0;32mok: [testbed-node-0] => (item=enable_mariadb_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1] => (item=enable_mariadb_True)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2] => (item=enable_mariadb_True)\u001b[0m",
                                "",
                                "PLAY [Apply role mariadb] ******************************************************",
                                "",
                                "TASK [mariadb : Group MariaDB hosts based on shards] ***************************",
                                "Tuesday 24 February 2026  01:26:06 +0000 (0:00:00.627)       0:00:01.200 ****** ",
                                "\u001b[0;32mok: [testbed-node-0] => (item=testbed-node-0)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0] => (item=testbed-node-1)\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-0] => (item=testbed-node-2)\u001b[0m",
                                "",
                                "TASK [mariadb : include_tasks] *************************************************",
                                "Tuesday 24 February 2026  01:26:06 +0000 (0:00:00.443)       0:00:01.643 ****** ",
                                "\u001b[0;36mincluded: /ansible/roles/mariadb/tasks/backup.yml for testbed-node-0, testbed-node-1, testbed-node-2\u001b[0m",
                                "",
                                "TASK [mariadb : Get MariaDB container facts] ***********************************",
                                "Tuesday 24 February 2026  01:26:07 +0000 (0:00:00.646)       0:00:02.290 ****** ",
                                "\u001b[0;32mok: [testbed-node-0]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-1]\u001b[0m",
                                "\u001b[0;32mok: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [mariadb : Taking full database backup via Mariabackup] *******************",
                                "Tuesday 24 February 2026  01:26:11 +0000 (0:00:03.663)       0:00:05.953 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-node-0]\u001b[0m",
                                "\u001b[1;35m[WARNING]: Could not match supplied host pattern, ignoring: mariadb_restart\u001b[0m",
                                "",
                                "PLAY [Restart mariadb services] ************************************************",
                                "\u001b[0;36mskipping: no hosts matched\u001b[0m",
                                "\u001b[1;35m[WARNING]: Could not match supplied host pattern, ignoring: mariadb_start\u001b[0m",
                                "",
                                "PLAY [Start mariadb services] **************************************************",
                                "\u001b[0;36mskipping: no hosts matched\u001b[0m",
                                "\u001b[1;35m[WARNING]: Could not match supplied host pattern, ignoring:\u001b[0m",
                                "\u001b[1;35mmariadb_bootstrap_restart\u001b[0m",
                                "",
                                "PLAY [Restart bootstrap mariadb service] ***************************************",
                                "\u001b[0;36mskipping: no hosts matched\u001b[0m",
                                "",
                                "PLAY [Apply mariadb post-configuration] ****************************************",
                                "",
                                "TASK [Include mariadb post-deploy.yml] *****************************************",
                                "Tuesday 24 February 2026  01:26:28 +0000 (0:00:17.742)       0:00:23.696 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "TASK [Include mariadb post-upgrade.yml] ****************************************",
                                "Tuesday 24 February 2026  01:26:29 +0000 (0:00:00.356)       0:00:24.052 ****** ",
                                "\u001b[0;36mskipping: [testbed-node-0]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-1]\u001b[0m",
                                "\u001b[0;36mskipping: [testbed-node-2]\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;33mtestbed-node-0\u001b[0m             : \u001b[0;32mok=6   \u001b[0m \u001b[0;33mchanged=1   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=2   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;32mtestbed-node-1\u001b[0m             : \u001b[0;32mok=4   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=3   \u001b[0m rescued=0    ignored=0   ",
                                "\u001b[0;32mtestbed-node-2\u001b[0m             : \u001b[0;32mok=4   \u001b[0m changed=0    unreachable=0    failed=0    \u001b[0;36mskipped=3   \u001b[0m rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Tuesday 24 February 2026  01:26:29 +0000 (0:00:00.471)       0:00:24.524 ****** ",
                                "=============================================================================== ",
                                "mariadb : Taking full database backup via Mariabackup ------------------ 17.74s",
                                "mariadb : Get MariaDB container facts ----------------------------------- 3.66s",
                                "mariadb : include_tasks ------------------------------------------------- 0.65s",
                                "Group hosts based on enabled services ----------------------------------- 0.63s",
                                "Include mariadb post-upgrade.yml ---------------------------------------- 0.47s",
                                "mariadb : Group MariaDB hosts based on shards --------------------------- 0.44s",
                                "Group hosts based on Kolla action --------------------------------------- 0.36s",
                                "Include mariadb post-deploy.yml ----------------------------------------- 0.36s",
                                "+ sh -c /opt/configuration/scripts/check/300-openstack.sh",
                                "+ set -e",
                                "+ source /opt/configuration/scripts/include.sh",
                                "++ export INTERACTIVE=false",
                                "++ INTERACTIVE=false",
                                "++ export OSISM_APPLY_RETRY=1",
                                "++ OSISM_APPLY_RETRY=1",
                                "+ source /opt/configuration/scripts/manager-version.sh",
                                "+++ awk '-F: ' '/^manager_version:/ { print $2 }' /opt/configuration/environments/manager/configuration.yml",
                                "",
                                "# OpenStack endpoints",
                                "",
                                "++ export MANAGER_VERSION=latest",
                                "++ MANAGER_VERSION=latest",
                                "+ export OS_CLOUD=admin",
                                "+ OS_CLOUD=admin",
                                "+ echo",
                                "+ echo '# OpenStack endpoints'",
                                "+ echo",
                                "+ openstack endpoint list",
                                "+----------------------------------+-----------+--------------+-----------------+---------+-----------+---------------------------------------------------------------------+",
                                "| ID                               | Region    | Service Name | Service Type    | Enabled | Interface | URL                                                                 |",
                                "+----------------------------------+-----------+--------------+-----------------+---------+-----------+---------------------------------------------------------------------+",
                                "| 0f79775df2984381b69a39a3c8f528a6 | RegionOne | neutron      | network         | True    | public    | https://api.testbed.osism.xyz:9696                                  |",
                                "| 11d66025a00741ed894131de48a5dae2 | RegionOne | glance       | image           | True    | public    | https://api.testbed.osism.xyz:9292                                  |",
                                "| 1339ce61b4c044e6b62d23025297cb02 | RegionOne | designate    | dns             | True    | public    | https://api.testbed.osism.xyz:9001                                  |",
                                "| 39efb8ff32f1445a98f7befaa27d645d | RegionOne | octavia      | load-balancer   | True    | public    | https://api.testbed.osism.xyz:9876                                  |",
                                "| 3f7ee67821ba444da6a14c0aa2e4a2db | RegionOne | swift        | object-store    | True    | public    | https://api.testbed.osism.xyz:6780/swift/v1/AUTH_%(project_id)s     |",
                                "| 4c07fb06d154427a9d22d05d70e3c3b3 | RegionOne | cinderv3     | volumev3        | True    | public    | https://api.testbed.osism.xyz:8776/v3/%(tenant_id)s                 |",
                                "| 4f6115d6d2694b208e9151fee09a1055 | RegionOne | octavia      | load-balancer   | True    | internal  | https://api-int.testbed.osism.xyz:9876                              |",
                                "| 5381b2fd05f24977b35e574551cc90e0 | RegionOne | placement    | placement       | True    | internal  | https://api-int.testbed.osism.xyz:8780                              |",
                                "| 7886a9dedb5847ef884e8ab23b142b95 | RegionOne | keystone     | identity        | True    | public    | https://api.testbed.osism.xyz:5000                                  |",
                                "| 820b19b0ce25401eb1312ddf139fc769 | RegionOne | swift        | object-store    | True    | internal  | https://api-int.testbed.osism.xyz:6780/swift/v1/AUTH_%(project_id)s |",
                                "| 8a2ce0d906a14e28af196bca035f3420 | RegionOne | magnum       | container-infra | True    | public    | https://api.testbed.osism.xyz:9511/v1                               |",
                                "| 9747730a38b14f92b3875abe2bd50c77 | RegionOne | magnum       | container-infra | True    | internal  | https://api-int.testbed.osism.xyz:9511/v1                           |",
                                "| 98d65750f75b4f09ac1bfd8bf24814d2 | RegionOne | barbican     | key-manager     | True    | internal  | https://api-int.testbed.osism.xyz:9311                              |",
                                "| 99f9e533d552453390d9520d0d1b791b | RegionOne | barbican     | key-manager     | True    | public    | https://api.testbed.osism.xyz:9311                                  |",
                                "| a1b5fd82ee5145a98ca38241b04e6ab9 | RegionOne | placement    | placement       | True    | public    | https://api.testbed.osism.xyz:8780                                  |",
                                "| ae914f5218ba421aa35104854b40493b | RegionOne | keystone     | identity        | True    | internal  | https://api-int.testbed.osism.xyz:5000                              |",
                                "| bed9f6e75968413a94b053e6321fde45 | RegionOne | neutron      | network         | True    | internal  | https://api-int.testbed.osism.xyz:9696                              |",
                                "| c17ded326a994063b389de5737dceafd | RegionOne | nova         | compute         | True    | public    | https://api.testbed.osism.xyz:8774/v2.1                             |",
                                "| ddd8a701f1854e3fa079a918613cdb47 | RegionOne | designate    | dns             | True    | internal  | https://api-int.testbed.osism.xyz:9001                              |",
                                "| e64331c9909c451581a31b1431037a8d | RegionOne | nova         | compute         | True    | internal  | https://api-int.testbed.osism.xyz:8774/v2.1                         |",
                                "| ef0d0bd767bc4f0fb38578ef589d6c7e | RegionOne | cinderv3     | volumev3        | True    | internal  | https://api-int.testbed.osism.xyz:8776/v3/%(tenant_id)s             |",
                                "| fc6fbb20a8434150a5b23f28ce920d9b | RegionOne | glance       | image           | True    | internal  | https://api-int.testbed.osism.xyz:9292                              |",
                                "+----------------------------------+-----------+--------------+-----------------+---------+-----------+---------------------------------------------------------------------+",
                                "",
                                "# Cinder",
                                "",
                                "+ echo",
                                "+ echo '# Cinder'",
                                "+ echo",
                                "+ openstack volume service list",
                                "+------------------+----------------------------+----------+---------+-------+----------------------------+",
                                "| Binary           | Host                       | Zone     | Status  | State | Updated At                 |",
                                "+------------------+----------------------------+----------+---------+-------+----------------------------+",
                                "| cinder-scheduler | testbed-node-0             | internal | enabled | up    | 2026-02-24T01:26:34.000000 |",
                                "| cinder-scheduler | testbed-node-1             | internal | enabled | up    | 2026-02-24T01:26:33.000000 |",
                                "| cinder-scheduler | testbed-node-2             | internal | enabled | up    | 2026-02-24T01:26:33.000000 |",
                                "| cinder-volume    | testbed-node-0@rbd-volumes | nova     | enabled | up    | 2026-02-24T01:26:33.000000 |",
                                "| cinder-volume    | testbed-node-2@rbd-volumes | nova     | enabled | up    | 2026-02-24T01:26:28.000000 |",
                                "| cinder-volume    | testbed-node-1@rbd-volumes | nova     | enabled | up    | 2026-02-24T01:26:28.000000 |",
                                "| cinder-backup    | testbed-node-0             | nova     | enabled | up    | 2026-02-24T01:26:30.000000 |",
                                "| cinder-backup    | testbed-node-1             | nova     | enabled | up    | 2026-02-24T01:26:32.000000 |",
                                "| cinder-backup    | testbed-node-2             | nova     | enabled | up    | 2026-02-24T01:26:32.000000 |",
                                "+------------------+----------------------------+----------+---------+-------+----------------------------+",
                                "",
                                "# Neutron",
                                "",
                                "+ echo",
                                "+ echo '# Neutron'",
                                "+ echo",
                                "+ openstack network agent list",
                                "+--------------------------------------+------------------------------+----------------+-------------------+-------+-------+----------------------------+",
                                "| ID                                   | Agent Type                   | Host           | Availability Zone | Alive | State | Binary                     |",
                                "+--------------------------------------+------------------------------+----------------+-------------------+-------+-------+----------------------------+",
                                "| testbed-node-3                       | OVN Controller agent         | testbed-node-3 |                   | :-)   | UP    | ovn-controller             |",
                                "| testbed-node-5                       | OVN Controller agent         | testbed-node-5 |                   | :-)   | UP    | ovn-controller             |",
                                "| testbed-node-0                       | OVN Controller Gateway agent | testbed-node-0 | nova              | :-)   | UP    | ovn-controller             |",
                                "| testbed-node-2                       | OVN Controller Gateway agent | testbed-node-2 | nova              | :-)   | UP    | ovn-controller             |",
                                "| testbed-node-4                       | OVN Controller agent         | testbed-node-4 |                   | :-)   | UP    | ovn-controller             |",
                                "| testbed-node-1                       | OVN Controller Gateway agent | testbed-node-1 | nova              | :-)   | UP    | ovn-controller             |",
                                "| e645415a-98f5-5758-8cd1-c47af282b5c0 | OVN Metadata agent           | testbed-node-3 |                   | :-)   | UP    | neutron-ovn-metadata-agent |",
                                "| 4939696e-6092-5a33-bb73-b850064684df | OVN Metadata agent           | testbed-node-4 |                   | :-)   | UP    | neutron-ovn-metadata-agent |",
                                "| 36b9d21c-9928-5c0a-9b27-73ac7a3e770c | OVN Metadata agent           | testbed-node-5 |                   | :-)   | UP    | neutron-ovn-metadata-agent |",
                                "+--------------------------------------+------------------------------+----------------+-------------------+-------+-------+----------------------------+",
                                "+ openstack network service provider list",
                                "+---------------+------+---------+",
                                "| Service Type  | Name | Default |",
                                "+---------------+------+---------+",
                                "| L3_ROUTER_NAT | ovn  | True    |",
                                "+---------------+------+---------+",
                                "",
                                "# Nova",
                                "",
                                "+ echo",
                                "+ echo '# Nova'",
                                "+ echo",
                                "+ openstack compute service list",
                                "+--------------------------------------+----------------+----------------+----------+---------+-------+----------------------------+",
                                "| ID                                   | Binary         | Host           | Zone     | Status  | State | Updated At                 |",
                                "+--------------------------------------+----------------+----------------+----------+---------+-------+----------------------------+",
                                "| 159da094-afa1-4e21-a1fa-329a2c038378 | nova-scheduler | testbed-node-0 | internal | enabled | up    | 2026-02-24T01:26:41.000000 |",
                                "| 45821a88-2046-43d9-9358-eb37605d19f4 | nova-scheduler | testbed-node-2 | internal | enabled | up    | 2026-02-24T01:26:44.000000 |",
                                "| e89363db-9a24-4ca8-9b4d-079af985725c | nova-scheduler | testbed-node-1 | internal | enabled | up    | 2026-02-24T01:26:44.000000 |",
                                "| 3e23bf07-6948-4a8b-9ceb-004e5ce25e33 | nova-conductor | testbed-node-0 | internal | enabled | up    | 2026-02-24T01:26:46.000000 |",
                                "| 9dea0c2d-25fa-4ca8-aa55-a2ea86ce238b | nova-conductor | testbed-node-2 | internal | enabled | up    | 2026-02-24T01:26:38.000000 |",
                                "| f3aeba5e-1840-4fa1-9fd9-71f9ee8ebbfb | nova-conductor | testbed-node-1 | internal | enabled | up    | 2026-02-24T01:26:38.000000 |",
                                "| 91bf9cdb-9bc1-45e9-9c85-614d018e902f | nova-compute   | testbed-node-4 | nova     | enabled | up    | 2026-02-24T01:26:38.000000 |",
                                "| a10b69c5-9163-455e-b4ce-1c530538f267 | nova-compute   | testbed-node-3 | nova     | enabled | up    | 2026-02-24T01:26:38.000000 |",
                                "| 2676ef64-9c2a-401a-8eb3-57e84238424a | nova-compute   | testbed-node-5 | nova     | enabled | up    | 2026-02-24T01:26:38.000000 |",
                                "+--------------------------------------+----------------+----------------+----------+---------+-------+----------------------------+",
                                "+ openstack hypervisor list",
                                "+--------------------------------------+---------------------+-----------------+---------------+-------+",
                                "| ID                                   | Hypervisor Hostname | Hypervisor Type | Host IP       | State |",
                                "+--------------------------------------+---------------------+-----------------+---------------+-------+",
                                "| 507d0ef5-91e2-4f9f-8f0f-2edc85623699 | testbed-node-4      | QEMU            | 192.168.16.14 | up    |",
                                "| 0bcf54ec-2474-42b4-9bd1-39dbdc3f0488 | testbed-node-3      | QEMU            | 192.168.16.13 | up    |",
                                "| 525e42fe-41e6-4fe2-9c2d-83a9aefd0b98 | testbed-node-5      | QEMU            | 192.168.16.15 | up    |",
                                "+--------------------------------------+---------------------+-----------------+---------------+-------+",
                                "",
                                "# Run OpenStack test play",
                                "",
                                "+ echo",
                                "+ echo '# Run OpenStack test play'",
                                "+ echo",
                                "+ osism apply --environment openstack test",
                                "\u001b[32m2026-02-24 01:26:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTrying to run play test in environment openstack\u001b[0m",
                                "\u001b[32m2026-02-24 01:27:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mPrepare task for execution of test.\u001b[0m",
                                "\u001b[32m2026-02-24 01:27:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask bf061e28-2200-486a-a46f-075db86a3739 (test) was prepared for execution.\u001b[0m",
                                "\u001b[32m2026-02-24 01:27:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task bf061e28-2200-486a-a46f-075db86a3739 (test) has been started and output is visible here.\u001b[0m",
                                "",
                                "PLAY [Create test project] *****************************************************",
                                "",
                                "TASK [Create test domain] ******************************************************",
                                "Tuesday 24 February 2026  01:27:08 +0000 (0:00:00.084)       0:00:00.084 ****** ",
                                "\u001b[0;33mchanged: [localhost]\u001b[0m",
                                "",
                                "TASK [Create test-admin user] **************************************************",
                                "Tuesday 24 February 2026  01:27:12 +0000 (0:00:04.304)       0:00:04.388 ****** ",
                                "\u001b[0;33mchanged: [localhost]\u001b[0m",
                                "",
                                "TASK [Add manager role to user test-admin] *************************************",
                                "Tuesday 24 February 2026  01:27:17 +0000 (0:00:04.771)       0:00:09.159 ****** ",
                                "\u001b[0;33mchanged: [localhost]\u001b[0m",
                                "",
                                "TASK [Create test project] *****************************************************",
                                "Tuesday 24 February 2026  01:27:24 +0000 (0:00:07.251)       0:00:16.411 ****** ",
                                "\u001b[0;33mchanged: [localhost]\u001b[0m",
                                "",
                                "TASK [Create test user] ********************************************************",
                                "Tuesday 24 February 2026  01:27:29 +0000 (0:00:04.585)       0:00:20.997 ****** ",
                                "\u001b[0;33mchanged: [localhost]\u001b[0m",
                                "",
                                "TASK [Add member roles to user test] *******************************************",
                                "Tuesday 24 February 2026  01:27:33 +0000 (0:00:04.540)       0:00:25.537 ****** ",
                                "\u001b[0;33mchanged: [localhost] => (item=load-balancer_member)\u001b[0m",
                                "\u001b[0;33mchanged: [localhost] => (item=member)\u001b[0m",
                                "\u001b[0;33mchanged: [localhost] => (item=creator)\u001b[0m",
                                "",
                                "TASK [Create test server group] ************************************************",
                                "Tuesday 24 February 2026  01:27:46 +0000 (0:00:12.578)       0:00:38.116 ****** ",
                                "\u001b[0;33mchanged: [localhost]\u001b[0m",
                                "",
                                "TASK [Create ssh security group] ***********************************************",
                                "Tuesday 24 February 2026  01:27:51 +0000 (0:00:04.802)       0:00:42.919 ****** ",
                                "\u001b[0;33mchanged: [localhost]\u001b[0m",
                                "",
                                "TASK [Add rule to ssh security group] ******************************************",
                                "Tuesday 24 February 2026  01:27:56 +0000 (0:00:05.379)       0:00:48.299 ****** ",
                                "\u001b[0;33mchanged: [localhost]\u001b[0m",
                                "",
                                "TASK [Create icmp security group] **********************************************",
                                "Tuesday 24 February 2026  01:28:01 +0000 (0:00:04.665)       0:00:52.964 ****** ",
                                "\u001b[0;33mchanged: [localhost]\u001b[0m",
                                "",
                                "TASK [Add rule to icmp security group] *****************************************",
                                "Tuesday 24 February 2026  01:28:05 +0000 (0:00:04.176)       0:00:57.141 ****** ",
                                "\u001b[0;33mchanged: [localhost]\u001b[0m",
                                "",
                                "TASK [Create test keypair] *****************************************************",
                                "Tuesday 24 February 2026  01:28:09 +0000 (0:00:04.370)       0:01:01.511 ****** ",
                                "\u001b[0;33mchanged: [localhost]\u001b[0m",
                                "",
                                "TASK [Create test network] *****************************************************",
                                "Tuesday 24 February 2026  01:28:14 +0000 (0:00:04.791)       0:01:06.302 ****** ",
                                "\u001b[0;33mchanged: [localhost]\u001b[0m",
                                "",
                                "TASK [Create test subnet] ******************************************************",
                                "Tuesday 24 February 2026  01:28:19 +0000 (0:00:05.117)       0:01:11.420 ****** ",
                                "\u001b[0;33mchanged: [localhost]\u001b[0m",
                                "",
                                "TASK [Create test router] ******************************************************",
                                "Tuesday 24 February 2026  01:28:25 +0000 (0:00:05.905)       0:01:17.326 ****** ",
                                "\u001b[0;33mchanged: [localhost]\u001b[0m",
                                "",
                                "PLAY [Manage test instances and volumes] ***************************************",
                                "",
                                "TASK [Get test server group] ***************************************************",
                                "Tuesday 24 February 2026  01:28:37 +0000 (0:00:11.931)       0:01:29.258 ****** ",
                                "\u001b[0;32mok: [localhost]\u001b[0m",
                                "",
                                "TASK [Detach test volume] ******************************************************",
                                "Tuesday 24 February 2026  01:28:42 +0000 (0:00:04.430)       0:01:33.688 ****** ",
                                "\u001b[0;36mskipping: [localhost]\u001b[0m",
                                "",
                                "TASK [Delete test volume] ******************************************************",
                                "Tuesday 24 February 2026  01:28:42 +0000 (0:00:00.062)       0:01:33.751 ****** ",
                                "\u001b[0;36mskipping: [localhost]\u001b[0m",
                                "",
                                "TASK [Delete test instances] ***************************************************",
                                "Tuesday 24 February 2026  01:28:42 +0000 (0:00:00.053)       0:01:33.804 ****** ",
                                "\u001b[0;36mskipping: [localhost] => (item=test-4) \u001b[0m",
                                "\u001b[0;36mskipping: [localhost] => (item=test-3) \u001b[0m",
                                "\u001b[0;36mskipping: [localhost] => (item=test-2) \u001b[0m",
                                "\u001b[0;36mskipping: [localhost] => (item=test-1) \u001b[0m",
                                "\u001b[0;36mskipping: [localhost] => (item=test) \u001b[0m",
                                "\u001b[0;36mskipping: [localhost]\u001b[0m",
                                "",
                                "TASK [Wait for instance deletion to complete] **********************************",
                                "Tuesday 24 February 2026  01:28:42 +0000 (0:00:00.191)       0:01:33.996 ****** ",
                                "\u001b[0;36mskipping: [localhost]\u001b[0m",
                                "",
                                "TASK [Create test instances] ***************************************************",
                                "Tuesday 24 February 2026  01:28:42 +0000 (0:00:00.176)       0:01:34.173 ****** ",
                                "\u001b[0;33mchanged: [localhost] => (item=test)\u001b[0m",
                                "\u001b[0;33mchanged: [localhost] => (item=test-1)\u001b[0m",
                                "\u001b[0;33mchanged: [localhost] => (item=test-2)\u001b[0m",
                                "\u001b[0;33mchanged: [localhost] => (item=test-3)\u001b[0m",
                                "\u001b[0;33mchanged: [localhost] => (item=test-4)\u001b[0m",
                                "",
                                "TASK [Wait for instance creation to complete] **********************************",
                                "Tuesday 24 February 2026  01:28:48 +0000 (0:00:05.595)       0:01:39.769 ****** ",
                                "\u001b[1;30mFAILED - RETRYING: [localhost]: Wait for instance creation to complete (60 retries left).\u001b[0m",
                                "\u001b[1;30mFAILED - RETRYING: [localhost]: Wait for instance creation to complete (59 retries left).\u001b[0m",
                                "\u001b[1;30mFAILED - RETRYING: [localhost]: Wait for instance creation to complete (58 retries left).\u001b[0m",
                                "\u001b[1;30mFAILED - RETRYING: [localhost]: Wait for instance creation to complete (57 retries left).\u001b[0m",
                                "\u001b[0;33mchanged: [localhost] => (item={'failed': 0, 'started': 1, 'finished': 0, 'ansible_job_id': 'j138772053794.2657', 'results_file': '/ansible/.ansible_async/j138772053794.2657', 'changed': True, 'item': 'test', 'ansible_loop_var': 'item'})\u001b[0m",
                                "\u001b[1;30mFAILED - RETRYING: [localhost]: Wait for instance creation to complete (60 retries left).\u001b[0m",
                                "\u001b[0;33mchanged: [localhost] => (item={'failed': 0, 'started': 1, 'finished': 0, 'ansible_job_id': 'j906118553359.2682', 'results_file': '/ansible/.ansible_async/j906118553359.2682', 'changed': True, 'item': 'test-1', 'ansible_loop_var': 'item'})\u001b[0m",
                                "\u001b[0;33mchanged: [localhost] => (item={'failed': 0, 'started': 1, 'finished': 0, 'ansible_job_id': 'j356775953011.2707', 'results_file': '/ansible/.ansible_async/j356775953011.2707', 'changed': True, 'item': 'test-2', 'ansible_loop_var': 'item'})\u001b[0m",
                                "\u001b[0;33mchanged: [localhost] => (item={'failed': 0, 'started': 1, 'finished': 0, 'ansible_job_id': 'j955926461069.2739', 'results_file': '/ansible/.ansible_async/j955926461069.2739', 'changed': True, 'item': 'test-3', 'ansible_loop_var': 'item'})\u001b[0m",
                                "\u001b[0;33mchanged: [localhost] => (item={'failed': 0, 'started': 1, 'finished': 0, 'ansible_job_id': 'j269502165248.2764', 'results_file': '/ansible/.ansible_async/j269502165248.2764', 'changed': True, 'item': 'test-4', 'ansible_loop_var': 'item'})\u001b[0m",
                                "",
                                "TASK [Add metadata to instances] ***********************************************",
                                "Tuesday 24 February 2026  01:29:46 +0000 (0:00:58.612)       0:02:38.381 ****** ",
                                "\u001b[0;33mchanged: [localhost] => (item=test)\u001b[0m",
                                "\u001b[0;33mchanged: [localhost] => (item=test-1)\u001b[0m",
                                "\u001b[0;33mchanged: [localhost] => (item=test-2)\u001b[0m",
                                "\u001b[0;33mchanged: [localhost] => (item=test-3)\u001b[0m",
                                "\u001b[0;33mchanged: [localhost] => (item=test-4)\u001b[0m",
                                "",
                                "TASK [Wait for metadata to be added] *******************************************",
                                "Tuesday 24 February 2026  01:29:51 +0000 (0:00:04.868)       0:02:43.249 ****** ",
                                "\u001b[1;30mFAILED - RETRYING: [localhost]: Wait for metadata to be added (30 retries left).\u001b[0m",
                                "\u001b[0;33mchanged: [localhost] => (item={'failed': 0, 'started': 1, 'finished': 0, 'ansible_job_id': 'j239672281548.2875', 'results_file': '/ansible/.ansible_async/j239672281548.2875', 'changed': True, 'item': 'test', 'ansible_loop_var': 'item'})\u001b[0m",
                                "\u001b[0;33mchanged: [localhost] => (item={'failed': 0, 'started': 1, 'finished': 0, 'ansible_job_id': 'j691765774518.2900', 'results_file': '/ansible/.ansible_async/j691765774518.2900', 'changed': True, 'item': 'test-1', 'ansible_loop_var': 'item'})\u001b[0m",
                                "\u001b[0;33mchanged: [localhost] => (item={'failed': 0, 'started': 1, 'finished': 0, 'ansible_job_id': 'j210524382123.2925', 'results_file': '/ansible/.ansible_async/j210524382123.2925', 'changed': True, 'item': 'test-2', 'ansible_loop_var': 'item'})\u001b[0m",
                                "\u001b[0;33mchanged: [localhost] => (item={'failed': 0, 'started': 1, 'finished': 0, 'ansible_job_id': 'j82192225245.2950', 'results_file': '/ansible/.ansible_async/j82192225245.2950', 'changed': True, 'item': 'test-3', 'ansible_loop_var': 'item'})\u001b[0m",
                                "\u001b[0;33mchanged: [localhost] => (item={'failed': 0, 'started': 1, 'finished': 0, 'ansible_job_id': 'j577716052385.2975', 'results_file': '/ansible/.ansible_async/j577716052385.2975', 'changed': True, 'item': 'test-4', 'ansible_loop_var': 'item'})\u001b[0m",
                                "",
                                "TASK [Add tag to instances] ****************************************************",
                                "Tuesday 24 February 2026  01:30:01 +0000 (0:00:09.945)       0:02:53.195 ****** ",
                                "\u001b[0;33mchanged: [localhost] => (item=test)\u001b[0m",
                                "\u001b[0;33mchanged: [localhost] => (item=test-1)\u001b[0m",
                                "\u001b[0;33mchanged: [localhost] => (item=test-2)\u001b[0m",
                                "\u001b[0;33mchanged: [localhost] => (item=test-3)\u001b[0m",
                                "\u001b[0;33mchanged: [localhost] => (item=test-4)\u001b[0m",
                                "",
                                "TASK [Wait for tags to be added] ***********************************************",
                                "Tuesday 24 February 2026  01:30:06 +0000 (0:00:05.186)       0:02:58.381 ****** ",
                                "\u001b[1;30mFAILED - RETRYING: [localhost]: Wait for tags to be added (30 retries left).\u001b[0m",
                                "\u001b[0;33mchanged: [localhost] => (item={'failed': 0, 'started': 1, 'finished': 0, 'ansible_job_id': 'j460463240913.3044', 'results_file': '/ansible/.ansible_async/j460463240913.3044', 'changed': True, 'item': 'test', 'ansible_loop_var': 'item'})\u001b[0m",
                                "\u001b[0;33mchanged: [localhost] => (item={'failed': 0, 'started': 1, 'finished': 0, 'ansible_job_id': 'j7366433596.3069', 'results_file': '/ansible/.ansible_async/j7366433596.3069', 'changed': True, 'item': 'test-1', 'ansible_loop_var': 'item'})\u001b[0m",
                                "\u001b[0;33mchanged: [localhost] => (item={'failed': 0, 'started': 1, 'finished': 0, 'ansible_job_id': 'j721472555207.3095', 'results_file': '/ansible/.ansible_async/j721472555207.3095', 'changed': True, 'item': 'test-2', 'ansible_loop_var': 'item'})\u001b[0m",
                                "\u001b[0;33mchanged: [localhost] => (item={'failed': 0, 'started': 1, 'finished': 0, 'ansible_job_id': 'j842733406981.3121', 'results_file': '/ansible/.ansible_async/j842733406981.3121', 'changed': True, 'item': 'test-3', 'ansible_loop_var': 'item'})\u001b[0m",
                                "\u001b[0;33mchanged: [localhost] => (item={'failed': 0, 'started': 1, 'finished': 0, 'ansible_job_id': 'j622682305399.3147', 'results_file': '/ansible/.ansible_async/j622682305399.3147', 'changed': True, 'item': 'test-4', 'ansible_loop_var': 'item'})\u001b[0m",
                                "",
                                "TASK [Create test volume] ******************************************************",
                                "Tuesday 24 February 2026  01:30:17 +0000 (0:00:10.389)       0:03:08.770 ****** ",
                                "\u001b[0;33mchanged: [localhost]\u001b[0m",
                                "",
                                "TASK [Attach test volume] ******************************************************",
                                "Tuesday 24 February 2026  01:30:23 +0000 (0:00:06.680)       0:03:15.451 ****** ",
                                "\u001b[0;33mchanged: [localhost]\u001b[0m",
                                "",
                                "TASK [Create floating ip address] **********************************************",
                                "Tuesday 24 February 2026  01:30:37 +0000 (0:00:13.569)       0:03:29.021 ****** ",
                                "\u001b[0;32mok: [localhost]\u001b[0m",
                                "",
                                "TASK [Print floating ip address] ***********************************************",
                                "Tuesday 24 February 2026  01:30:42 +0000 (0:00:04.934)       0:03:33.956 ****** ",
                                "\u001b[0;32mok: [localhost] => {\u001b[0m",
                                "\u001b[0;32m    \"msg\": \"192.168.112.184\"\u001b[0m",
                                "\u001b[0;32m}\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;33mlocalhost\u001b[0m                  : \u001b[0;32mok=26  \u001b[0m \u001b[0;33mchanged=23  \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=4   \u001b[0m rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Tuesday 24 February 2026  01:30:42 +0000 (0:00:00.045)       0:03:34.001 ****** ",
                                "=============================================================================== ",
                                "Wait for instance creation to complete --------------------------------- 58.61s",
                                "Attach test volume ----------------------------------------------------- 13.57s",
                                "Add member roles to user test ------------------------------------------ 12.58s",
                                "Create test router ----------------------------------------------------- 11.93s",
                                "Wait for tags to be added ---------------------------------------------- 10.39s",
                                "Wait for metadata to be added ------------------------------------------- 9.94s",
                                "Add manager role to user test-admin ------------------------------------- 7.25s",
                                "Create test volume ------------------------------------------------------ 6.68s",
                                "Create test subnet ------------------------------------------------------ 5.91s",
                                "Create test instances --------------------------------------------------- 5.60s",
                                "Create ssh security group ----------------------------------------------- 5.38s",
                                "Add tag to instances ---------------------------------------------------- 5.19s",
                                "Create test network ----------------------------------------------------- 5.12s",
                                "Create floating ip address ---------------------------------------------- 4.93s",
                                "Add metadata to instances ----------------------------------------------- 4.87s",
                                "Create test server group ------------------------------------------------ 4.80s",
                                "Create test keypair ----------------------------------------------------- 4.79s",
                                "Create test-admin user -------------------------------------------------- 4.77s",
                                "Add rule to ssh security group ------------------------------------------ 4.67s",
                                "Create test project ----------------------------------------------------- 4.59s",
                                "+ server_list",
                                "+ openstack --os-cloud test server list",
                                "+--------------------------------------+--------+--------+---------------------------------------+--------------------------+----------+",
                                "| ID                                   | Name   | Status | Networks                              | Image                    | Flavor   |",
                                "+--------------------------------------+--------+--------+---------------------------------------+--------------------------+----------+",
                                "| 0b9309c0-5fce-46a5-86f1-8d5360f5077a | test-4 | ACTIVE | test=192.168.112.193, 192.168.200.119 | N/A (booted from volume) | SCS-1L-1 |",
                                "| 36006f27-69d5-4fba-95d5-a09ad6e3d2b6 | test-3 | ACTIVE | test=192.168.112.192, 192.168.200.217 | N/A (booted from volume) | SCS-1L-1 |",
                                "| 5ca62725-a636-4f3b-bb5e-7b2c661006b6 | test-1 | ACTIVE | test=192.168.112.182, 192.168.200.206 | N/A (booted from volume) | SCS-1L-1 |",
                                "| 73a9cbe2-a4c7-4bb9-be1d-0f4606a87824 | test   | ACTIVE | test=192.168.112.184, 192.168.200.68  | N/A (booted from volume) | SCS-1L-1 |",
                                "| e237ead4-6261-44c0-95f1-03bd2fd83f6a | test-2 | ACTIVE | test=192.168.112.141, 192.168.200.188 | N/A (booted from volume) | SCS-1L-1 |",
                                "+--------------------------------------+--------+--------+---------------------------------------+--------------------------+----------+",
                                "+ openstack --os-cloud test server show test",
                                "+-------------------------------------+-------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+",
                                "| Field                               | Value                                                                                                                                                                                                                                                                                                                                                                               |",
                                "+-------------------------------------+-------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+",
                                "| OS-DCF:diskConfig                   | MANUAL                                                                                                                                                                                                                                                                                                                                                                              |",
                                "| OS-EXT-AZ:availability_zone         | nova                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| OS-EXT-SRV-ATTR:host                | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| OS-EXT-SRV-ATTR:hostname            | test                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| OS-EXT-SRV-ATTR:hypervisor_hostname | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| OS-EXT-SRV-ATTR:instance_name       | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| OS-EXT-SRV-ATTR:kernel_id           | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| OS-EXT-SRV-ATTR:launch_index        | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| OS-EXT-SRV-ATTR:ramdisk_id          | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| OS-EXT-SRV-ATTR:reservation_id      | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| OS-EXT-SRV-ATTR:root_device_name    | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| OS-EXT-SRV-ATTR:user_data           | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| OS-EXT-STS:power_state              | Running                                                                                                                                                                                                                                                                                                                                                                             |",
                                "| OS-EXT-STS:task_state               | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| OS-EXT-STS:vm_state                 | active                                                                                                                                                                                                                                                                                                                                                                              |",
                                "| OS-SRV-USG:launched_at              | 2026-02-24T01:29:21.000000                                                                                                                                                                                                                                                                                                                                                          |",
                                "| OS-SRV-USG:terminated_at            | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| accessIPv4                          |                                                                                                                                                                                                                                                                                                                                                                                     |",
                                "| accessIPv6                          |                                                                                                                                                                                                                                                                                                                                                                                     |",
                                "| addresses                           | test=192.168.112.184, 192.168.200.68                                                                                                                                                                                                                                                                                                                                                |",
                                "| config_drive                        |                                                                                                                                                                                                                                                                                                                                                                                     |",
                                "| created                             | 2026-02-24T01:28:53Z                                                                                                                                                                                                                                                                                                                                                                |",
                                "| description                         | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| flavor                              | description=, disk='0', ephemeral='0', extra_specs.hw_rng:allowed='true', extra_specs.scs:cpu-type='crowded-core', extra_specs.scs:disk0-type='network', extra_specs.scs:name-v1='SCS-1L:1', extra_specs.scs:name-v2='SCS-1L-1', id='SCS-1L-1', is_disabled=, is_public='True', location=, name='SCS-1L-1', original_name='SCS-1L-1', ram='1024', rxtx_factor=, swap='0', vcpus='1' |",
                                "| hostId                              | 22a9aa40e3d2fc5564193a7c5a62e11757162f1fc18a38281aa3588f                                                                                                                                                                                                                                                                                                                            |",
                                "| host_status                         | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| id                                  | 73a9cbe2-a4c7-4bb9-be1d-0f4606a87824                                                                                                                                                                                                                                                                                                                                                |",
                                "| image                               | N/A (booted from volume)                                                                                                                                                                                                                                                                                                                                                            |",
                                "| key_name                            | test                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| locked                              | False                                                                                                                                                                                                                                                                                                                                                                               |",
                                "| locked_reason                       | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| name                                | test                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| pinned_availability_zone            | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| progress                            | 0                                                                                                                                                                                                                                                                                                                                                                                   |",
                                "| project_id                          | b7fd8521139f4f698ef53541c2215403                                                                                                                                                                                                                                                                                                                                                    |",
                                "| properties                          | hostname='test'                                                                                                                                                                                                                                                                                                                                                                     |",
                                "| security_groups                     | name='icmp'                                                                                                                                                                                                                                                                                                                                                                         |",
                                "|                                     | name='ssh'                                                                                                                                                                                                                                                                                                                                                                          |",
                                "| server_groups                       | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| status                              | ACTIVE                                                                                                                                                                                                                                                                                                                                                                              |",
                                "| tags                                | test                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| trusted_image_certificates          | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| updated                             | 2026-02-24T01:29:52Z                                                                                                                                                                                                                                                                                                                                                                |",
                                "| user_id                             | e7bd38d58eed475ab28679fe129b0889                                                                                                                                                                                                                                                                                                                                                    |",
                                "| volumes_attached                    | delete_on_termination='True', id='fc2483d8-4dc3-4f58-852c-acaadfc5f2d2'                                                                                                                                                                                                                                                                                                             |",
                                "|                                     | delete_on_termination='False', id='1b918256-4d4e-4cf8-9f1c-3a729ddb7086'                                                                                                                                                                                                                                                                                                            |",
                                "+-------------------------------------+-------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+",
                                "+ openstack --os-cloud test server show test-1",
                                "+-------------------------------------+-------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+",
                                "| Field                               | Value                                                                                                                                                                                                                                                                                                                                                                               |",
                                "+-------------------------------------+-------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+",
                                "| OS-DCF:diskConfig                   | MANUAL                                                                                                                                                                                                                                                                                                                                                                              |",
                                "| OS-EXT-AZ:availability_zone         | nova                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| OS-EXT-SRV-ATTR:host                | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| OS-EXT-SRV-ATTR:hostname            | test-1                                                                                                                                                                                                                                                                                                                                                                              |",
                                "| OS-EXT-SRV-ATTR:hypervisor_hostname | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| OS-EXT-SRV-ATTR:instance_name       | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| OS-EXT-SRV-ATTR:kernel_id           | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| OS-EXT-SRV-ATTR:launch_index        | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| OS-EXT-SRV-ATTR:ramdisk_id          | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| OS-EXT-SRV-ATTR:reservation_id      | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| OS-EXT-SRV-ATTR:root_device_name    | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| OS-EXT-SRV-ATTR:user_data           | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| OS-EXT-STS:power_state              | Running                                                                                                                                                                                                                                                                                                                                                                             |",
                                "| OS-EXT-STS:task_state               | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| OS-EXT-STS:vm_state                 | active                                                                                                                                                                                                                                                                                                                                                                              |",
                                "| OS-SRV-USG:launched_at              | 2026-02-24T01:29:23.000000                                                                                                                                                                                                                                                                                                                                                          |",
                                "| OS-SRV-USG:terminated_at            | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| accessIPv4                          |                                                                                                                                                                                                                                                                                                                                                                                     |",
                                "| accessIPv6                          |                                                                                                                                                                                                                                                                                                                                                                                     |",
                                "| addresses                           | test=192.168.112.182, 192.168.200.206                                                                                                                                                                                                                                                                                                                                               |",
                                "| config_drive                        |                                                                                                                                                                                                                                                                                                                                                                                     |",
                                "| created                             | 2026-02-24T01:28:53Z                                                                                                                                                                                                                                                                                                                                                                |",
                                "| description                         | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| flavor                              | description=, disk='0', ephemeral='0', extra_specs.hw_rng:allowed='true', extra_specs.scs:cpu-type='crowded-core', extra_specs.scs:disk0-type='network', extra_specs.scs:name-v1='SCS-1L:1', extra_specs.scs:name-v2='SCS-1L-1', id='SCS-1L-1', is_disabled=, is_public='True', location=, name='SCS-1L-1', original_name='SCS-1L-1', ram='1024', rxtx_factor=, swap='0', vcpus='1' |",
                                "| hostId                              | 22a9aa40e3d2fc5564193a7c5a62e11757162f1fc18a38281aa3588f                                                                                                                                                                                                                                                                                                                            |",
                                "| host_status                         | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| id                                  | 5ca62725-a636-4f3b-bb5e-7b2c661006b6                                                                                                                                                                                                                                                                                                                                                |",
                                "| image                               | N/A (booted from volume)                                                                                                                                                                                                                                                                                                                                                            |",
                                "| key_name                            | test                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| locked                              | False                                                                                                                                                                                                                                                                                                                                                                               |",
                                "| locked_reason                       | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| name                                | test-1                                                                                                                                                                                                                                                                                                                                                                              |",
                                "| pinned_availability_zone            | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| progress                            | 0                                                                                                                                                                                                                                                                                                                                                                                   |",
                                "| project_id                          | b7fd8521139f4f698ef53541c2215403                                                                                                                                                                                                                                                                                                                                                    |",
                                "| properties                          | hostname='test-1'                                                                                                                                                                                                                                                                                                                                                                   |",
                                "| security_groups                     | name='icmp'                                                                                                                                                                                                                                                                                                                                                                         |",
                                "|                                     | name='ssh'                                                                                                                                                                                                                                                                                                                                                                          |",
                                "| server_groups                       | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| status                              | ACTIVE                                                                                                                                                                                                                                                                                                                                                                              |",
                                "| tags                                | test                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| trusted_image_certificates          | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| updated                             | 2026-02-24T01:29:53Z                                                                                                                                                                                                                                                                                                                                                                |",
                                "| user_id                             | e7bd38d58eed475ab28679fe129b0889                                                                                                                                                                                                                                                                                                                                                    |",
                                "| volumes_attached                    | delete_on_termination='True', id='2f4f7b85-91b4-430d-abe3-4e964e5dea87'                                                                                                                                                                                                                                                                                                             |",
                                "+-------------------------------------+-------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+",
                                "+ openstack --os-cloud test server show test-2",
                                "+-------------------------------------+-------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+",
                                "| Field                               | Value                                                                                                                                                                                                                                                                                                                                                                               |",
                                "+-------------------------------------+-------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+",
                                "| OS-DCF:diskConfig                   | MANUAL                                                                                                                                                                                                                                                                                                                                                                              |",
                                "| OS-EXT-AZ:availability_zone         | nova                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| OS-EXT-SRV-ATTR:host                | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| OS-EXT-SRV-ATTR:hostname            | test-2                                                                                                                                                                                                                                                                                                                                                                              |",
                                "| OS-EXT-SRV-ATTR:hypervisor_hostname | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| OS-EXT-SRV-ATTR:instance_name       | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| OS-EXT-SRV-ATTR:kernel_id           | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| OS-EXT-SRV-ATTR:launch_index        | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| OS-EXT-SRV-ATTR:ramdisk_id          | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| OS-EXT-SRV-ATTR:reservation_id      | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| OS-EXT-SRV-ATTR:root_device_name    | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| OS-EXT-SRV-ATTR:user_data           | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| OS-EXT-STS:power_state              | Running                                                                                                                                                                                                                                                                                                                                                                             |",
                                "| OS-EXT-STS:task_state               | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| OS-EXT-STS:vm_state                 | active                                                                                                                                                                                                                                                                                                                                                                              |",
                                "| OS-SRV-USG:launched_at              | 2026-02-24T01:29:21.000000                                                                                                                                                                                                                                                                                                                                                          |",
                                "| OS-SRV-USG:terminated_at            | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| accessIPv4                          |                                                                                                                                                                                                                                                                                                                                                                                     |",
                                "| accessIPv6                          |                                                                                                                                                                                                                                                                                                                                                                                     |",
                                "| addresses                           | test=192.168.112.141, 192.168.200.188                                                                                                                                                                                                                                                                                                                                               |",
                                "| config_drive                        |                                                                                                                                                                                                                                                                                                                                                                                     |",
                                "| created                             | 2026-02-24T01:28:53Z                                                                                                                                                                                                                                                                                                                                                                |",
                                "| description                         | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| flavor                              | description=, disk='0', ephemeral='0', extra_specs.hw_rng:allowed='true', extra_specs.scs:cpu-type='crowded-core', extra_specs.scs:disk0-type='network', extra_specs.scs:name-v1='SCS-1L:1', extra_specs.scs:name-v2='SCS-1L-1', id='SCS-1L-1', is_disabled=, is_public='True', location=, name='SCS-1L-1', original_name='SCS-1L-1', ram='1024', rxtx_factor=, swap='0', vcpus='1' |",
                                "| hostId                              | 660187b7e152ec0335e13743959201b6eb9412e4dde6e4a1f5c8c5a7                                                                                                                                                                                                                                                                                                                            |",
                                "| host_status                         | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| id                                  | e237ead4-6261-44c0-95f1-03bd2fd83f6a                                                                                                                                                                                                                                                                                                                                                |",
                                "| image                               | N/A (booted from volume)                                                                                                                                                                                                                                                                                                                                                            |",
                                "| key_name                            | test                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| locked                              | False                                                                                                                                                                                                                                                                                                                                                                               |",
                                "| locked_reason                       | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| name                                | test-2                                                                                                                                                                                                                                                                                                                                                                              |",
                                "| pinned_availability_zone            | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| progress                            | 0                                                                                                                                                                                                                                                                                                                                                                                   |",
                                "| project_id                          | b7fd8521139f4f698ef53541c2215403                                                                                                                                                                                                                                                                                                                                                    |",
                                "| properties                          | hostname='test-2'                                                                                                                                                                                                                                                                                                                                                                   |",
                                "| security_groups                     | name='icmp'                                                                                                                                                                                                                                                                                                                                                                         |",
                                "|                                     | name='ssh'                                                                                                                                                                                                                                                                                                                                                                          |",
                                "| server_groups                       | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| status                              | ACTIVE                                                                                                                                                                                                                                                                                                                                                                              |",
                                "| tags                                | test                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| trusted_image_certificates          | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| updated                             | 2026-02-24T01:29:54Z                                                                                                                                                                                                                                                                                                                                                                |",
                                "| user_id                             | e7bd38d58eed475ab28679fe129b0889                                                                                                                                                                                                                                                                                                                                                    |",
                                "| volumes_attached                    | delete_on_termination='True', id='461d2ade-dad0-452c-ab59-2b1683c48a38'                                                                                                                                                                                                                                                                                                             |",
                                "+-------------------------------------+-------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+",
                                "+ openstack --os-cloud test server show test-3",
                                "+-------------------------------------+-------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+",
                                "| Field                               | Value                                                                                                                                                                                                                                                                                                                                                                               |",
                                "+-------------------------------------+-------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+",
                                "| OS-DCF:diskConfig                   | MANUAL                                                                                                                                                                                                                                                                                                                                                                              |",
                                "| OS-EXT-AZ:availability_zone         | nova                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| OS-EXT-SRV-ATTR:host                | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| OS-EXT-SRV-ATTR:hostname            | test-3                                                                                                                                                                                                                                                                                                                                                                              |",
                                "| OS-EXT-SRV-ATTR:hypervisor_hostname | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| OS-EXT-SRV-ATTR:instance_name       | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| OS-EXT-SRV-ATTR:kernel_id           | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| OS-EXT-SRV-ATTR:launch_index        | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| OS-EXT-SRV-ATTR:ramdisk_id          | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| OS-EXT-SRV-ATTR:reservation_id      | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| OS-EXT-SRV-ATTR:root_device_name    | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| OS-EXT-SRV-ATTR:user_data           | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| OS-EXT-STS:power_state              | Running                                                                                                                                                                                                                                                                                                                                                                             |",
                                "| OS-EXT-STS:task_state               | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| OS-EXT-STS:vm_state                 | active                                                                                                                                                                                                                                                                                                                                                                              |",
                                "| OS-SRV-USG:launched_at              | 2026-02-24T01:29:23.000000                                                                                                                                                                                                                                                                                                                                                          |",
                                "| OS-SRV-USG:terminated_at            | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| accessIPv4                          |                                                                                                                                                                                                                                                                                                                                                                                     |",
                                "| accessIPv6                          |                                                                                                                                                                                                                                                                                                                                                                                     |",
                                "| addresses                           | test=192.168.112.192, 192.168.200.217                                                                                                                                                                                                                                                                                                                                               |",
                                "| config_drive                        |                                                                                                                                                                                                                                                                                                                                                                                     |",
                                "| created                             | 2026-02-24T01:28:54Z                                                                                                                                                                                                                                                                                                                                                                |",
                                "| description                         | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| flavor                              | description=, disk='0', ephemeral='0', extra_specs.hw_rng:allowed='true', extra_specs.scs:cpu-type='crowded-core', extra_specs.scs:disk0-type='network', extra_specs.scs:name-v1='SCS-1L:1', extra_specs.scs:name-v2='SCS-1L-1', id='SCS-1L-1', is_disabled=, is_public='True', location=, name='SCS-1L-1', original_name='SCS-1L-1', ram='1024', rxtx_factor=, swap='0', vcpus='1' |",
                                "| hostId                              | 660187b7e152ec0335e13743959201b6eb9412e4dde6e4a1f5c8c5a7                                                                                                                                                                                                                                                                                                                            |",
                                "| host_status                         | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| id                                  | 36006f27-69d5-4fba-95d5-a09ad6e3d2b6                                                                                                                                                                                                                                                                                                                                                |",
                                "| image                               | N/A (booted from volume)                                                                                                                                                                                                                                                                                                                                                            |",
                                "| key_name                            | test                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| locked                              | False                                                                                                                                                                                                                                                                                                                                                                               |",
                                "| locked_reason                       | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| name                                | test-3                                                                                                                                                                                                                                                                                                                                                                              |",
                                "| pinned_availability_zone            | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| progress                            | 0                                                                                                                                                                                                                                                                                                                                                                                   |",
                                "| project_id                          | b7fd8521139f4f698ef53541c2215403                                                                                                                                                                                                                                                                                                                                                    |",
                                "| properties                          | hostname='test-3'                                                                                                                                                                                                                                                                                                                                                                   |",
                                "| security_groups                     | name='icmp'                                                                                                                                                                                                                                                                                                                                                                         |",
                                "|                                     | name='ssh'                                                                                                                                                                                                                                                                                                                                                                          |",
                                "| server_groups                       | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| status                              | ACTIVE                                                                                                                                                                                                                                                                                                                                                                              |",
                                "| tags                                | test                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| trusted_image_certificates          | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| updated                             | 2026-02-24T01:29:54Z                                                                                                                                                                                                                                                                                                                                                                |",
                                "| user_id                             | e7bd38d58eed475ab28679fe129b0889                                                                                                                                                                                                                                                                                                                                                    |",
                                "| volumes_attached                    | delete_on_termination='True', id='17957e18-17d5-475f-a5a6-11013bd262d8'                                                                                                                                                                                                                                                                                                             |",
                                "+-------------------------------------+-------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+",
                                "+ openstack --os-cloud test server show test-4",
                                "+-------------------------------------+-------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+",
                                "| Field                               | Value                                                                                                                                                                                                                                                                                                                                                                               |",
                                "+-------------------------------------+-------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+",
                                "| OS-DCF:diskConfig                   | MANUAL                                                                                                                                                                                                                                                                                                                                                                              |",
                                "| OS-EXT-AZ:availability_zone         | nova                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| OS-EXT-SRV-ATTR:host                | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| OS-EXT-SRV-ATTR:hostname            | test-4                                                                                                                                                                                                                                                                                                                                                                              |",
                                "| OS-EXT-SRV-ATTR:hypervisor_hostname | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| OS-EXT-SRV-ATTR:instance_name       | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| OS-EXT-SRV-ATTR:kernel_id           | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| OS-EXT-SRV-ATTR:launch_index        | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| OS-EXT-SRV-ATTR:ramdisk_id          | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| OS-EXT-SRV-ATTR:reservation_id      | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| OS-EXT-SRV-ATTR:root_device_name    | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| OS-EXT-SRV-ATTR:user_data           | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| OS-EXT-STS:power_state              | Running                                                                                                                                                                                                                                                                                                                                                                             |",
                                "| OS-EXT-STS:task_state               | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| OS-EXT-STS:vm_state                 | active                                                                                                                                                                                                                                                                                                                                                                              |",
                                "| OS-SRV-USG:launched_at              | 2026-02-24T01:29:21.000000                                                                                                                                                                                                                                                                                                                                                          |",
                                "| OS-SRV-USG:terminated_at            | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| accessIPv4                          |                                                                                                                                                                                                                                                                                                                                                                                     |",
                                "| accessIPv6                          |                                                                                                                                                                                                                                                                                                                                                                                     |",
                                "| addresses                           | test=192.168.112.193, 192.168.200.119                                                                                                                                                                                                                                                                                                                                               |",
                                "| config_drive                        |                                                                                                                                                                                                                                                                                                                                                                                     |",
                                "| created                             | 2026-02-24T01:28:55Z                                                                                                                                                                                                                                                                                                                                                                |",
                                "| description                         | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| flavor                              | description=, disk='0', ephemeral='0', extra_specs.hw_rng:allowed='true', extra_specs.scs:cpu-type='crowded-core', extra_specs.scs:disk0-type='network', extra_specs.scs:name-v1='SCS-1L:1', extra_specs.scs:name-v2='SCS-1L-1', id='SCS-1L-1', is_disabled=, is_public='True', location=, name='SCS-1L-1', original_name='SCS-1L-1', ram='1024', rxtx_factor=, swap='0', vcpus='1' |",
                                "| hostId                              | 22a9aa40e3d2fc5564193a7c5a62e11757162f1fc18a38281aa3588f                                                                                                                                                                                                                                                                                                                            |",
                                "| host_status                         | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| id                                  | 0b9309c0-5fce-46a5-86f1-8d5360f5077a                                                                                                                                                                                                                                                                                                                                                |",
                                "| image                               | N/A (booted from volume)                                                                                                                                                                                                                                                                                                                                                            |",
                                "| key_name                            | test                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| locked                              | False                                                                                                                                                                                                                                                                                                                                                                               |",
                                "| locked_reason                       | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| name                                | test-4                                                                                                                                                                                                                                                                                                                                                                              |",
                                "| pinned_availability_zone            | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| progress                            | 0                                                                                                                                                                                                                                                                                                                                                                                   |",
                                "| project_id                          | b7fd8521139f4f698ef53541c2215403                                                                                                                                                                                                                                                                                                                                                    |",
                                "| properties                          | hostname='test-4'                                                                                                                                                                                                                                                                                                                                                                   |",
                                "| security_groups                     | name='icmp'                                                                                                                                                                                                                                                                                                                                                                         |",
                                "|                                     | name='ssh'                                                                                                                                                                                                                                                                                                                                                                          |",
                                "| server_groups                       | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| status                              | ACTIVE                                                                                                                                                                                                                                                                                                                                                                              |",
                                "| tags                                | test                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| trusted_image_certificates          | None                                                                                                                                                                                                                                                                                                                                                                                |",
                                "| updated                             | 2026-02-24T01:29:55Z                                                                                                                                                                                                                                                                                                                                                                |",
                                "| user_id                             | e7bd38d58eed475ab28679fe129b0889                                                                                                                                                                                                                                                                                                                                                    |",
                                "| volumes_attached                    | delete_on_termination='True', id='90f82ae0-4c58-479a-837e-190419bc3244'                                                                                                                                                                                                                                                                                                             |",
                                "+-------------------------------------+-------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+",
                                "+ server_ping",
                                "++ openstack --os-cloud test floating ip list --status ACTIVE -f value -c 'Floating IP Address'",
                                "++ tr -d '\\r'",
                                "+ for address in $(openstack --os-cloud test floating ip list --status ACTIVE -f value -c \"Floating IP Address\" | tr -d '\\r')",
                                "+ ping -c3 192.168.112.193",
                                "PING 192.168.112.193 (192.168.112.193) 56(84) bytes of data.",
                                "64 bytes from 192.168.112.193: icmp_seq=1 ttl=63 time=8.78 ms",
                                "64 bytes from 192.168.112.193: icmp_seq=2 ttl=63 time=2.35 ms",
                                "64 bytes from 192.168.112.193: icmp_seq=3 ttl=63 time=2.29 ms",
                                "",
                                "--- 192.168.112.193 ping statistics ---",
                                "3 packets transmitted, 3 received, 0% packet loss, time 2002ms",
                                "rtt min/avg/max/mdev = 2.292/4.474/8.782/3.046 ms",
                                "+ for address in $(openstack --os-cloud test floating ip list --status ACTIVE -f value -c \"Floating IP Address\" | tr -d '\\r')",
                                "+ ping -c3 192.168.112.184",
                                "PING 192.168.112.184 (192.168.112.184) 56(84) bytes of data.",
                                "64 bytes from 192.168.112.184: icmp_seq=1 ttl=63 time=5.67 ms",
                                "64 bytes from 192.168.112.184: icmp_seq=2 ttl=63 time=2.22 ms",
                                "64 bytes from 192.168.112.184: icmp_seq=3 ttl=63 time=1.75 ms",
                                "",
                                "--- 192.168.112.184 ping statistics ---",
                                "3 packets transmitted, 3 received, 0% packet loss, time 2003ms",
                                "rtt min/avg/max/mdev = 1.749/3.210/5.667/1.747 ms",
                                "+ for address in $(openstack --os-cloud test floating ip list --status ACTIVE -f value -c \"Floating IP Address\" | tr -d '\\r')",
                                "+ ping -c3 192.168.112.182",
                                "PING 192.168.112.182 (192.168.112.182) 56(84) bytes of data.",
                                "64 bytes from 192.168.112.182: icmp_seq=1 ttl=63 time=4.92 ms",
                                "64 bytes from 192.168.112.182: icmp_seq=2 ttl=63 time=2.41 ms",
                                "64 bytes from 192.168.112.182: icmp_seq=3 ttl=63 time=1.83 ms",
                                "",
                                "--- 192.168.112.182 ping statistics ---",
                                "3 packets transmitted, 3 received, 0% packet loss, time 2002ms",
                                "rtt min/avg/max/mdev = 1.831/3.054/4.921/1.340 ms",
                                "+ for address in $(openstack --os-cloud test floating ip list --status ACTIVE -f value -c \"Floating IP Address\" | tr -d '\\r')",
                                "+ ping -c3 192.168.112.141",
                                "PING 192.168.112.141 (192.168.112.141) 56(84) bytes of data.",
                                "64 bytes from 192.168.112.141: icmp_seq=1 ttl=63 time=5.78 ms",
                                "64 bytes from 192.168.112.141: icmp_seq=2 ttl=63 time=2.10 ms",
                                "64 bytes from 192.168.112.141: icmp_seq=3 ttl=63 time=2.04 ms",
                                "",
                                "--- 192.168.112.141 ping statistics ---",
                                "3 packets transmitted, 3 received, 0% packet loss, time 2003ms",
                                "rtt min/avg/max/mdev = 2.037/3.305/5.782/1.751 ms",
                                "+ for address in $(openstack --os-cloud test floating ip list --status ACTIVE -f value -c \"Floating IP Address\" | tr -d '\\r')",
                                "+ ping -c3 192.168.112.192",
                                "PING 192.168.112.192 (192.168.112.192) 56(84) bytes of data.",
                                "64 bytes from 192.168.112.192: icmp_seq=1 ttl=63 time=5.25 ms",
                                "64 bytes from 192.168.112.192: icmp_seq=2 ttl=63 time=1.97 ms",
                                "64 bytes from 192.168.112.192: icmp_seq=3 ttl=63 time=1.64 ms",
                                "",
                                "--- 192.168.112.192 ping statistics ---",
                                "3 packets transmitted, 3 received, 0% packet loss, time 2002ms",
                                "rtt min/avg/max/mdev = 1.641/2.954/5.251/1.629 ms",
                                "+ [[ latest == \\l\\a\\t\\e\\s\\t ]]",
                                "+ compute_list",
                                "+ osism manage compute list testbed-node-3",
                                "\u001b[32m2026-02-24 01:31:17\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mUnable to get ansible vault password\u001b[0m",
                                "\u001b[32m2026-02-24 01:31:17\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mUnable to get vault secret: [Errno 2] No such file or directory: '/share/ansible_vault_password.key'\u001b[0m",
                                "\u001b[32m2026-02-24 01:31:17\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mDropping encrypted entries\u001b[0m",
                                "+--------------------------------------+--------+----------+",
                                "| ID                                   | Name   | Status   |",
                                "|--------------------------------------+--------+----------|",
                                "| 0b9309c0-5fce-46a5-86f1-8d5360f5077a | test-4 | ACTIVE   |",
                                "| 5ca62725-a636-4f3b-bb5e-7b2c661006b6 | test-1 | ACTIVE   |",
                                "| 73a9cbe2-a4c7-4bb9-be1d-0f4606a87824 | test   | ACTIVE   |",
                                "+--------------------------------------+--------+----------+",
                                "+ osism manage compute list testbed-node-4",
                                "\u001b[32m2026-02-24 01:31:22\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mUnable to get ansible vault password\u001b[0m",
                                "\u001b[32m2026-02-24 01:31:22\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mUnable to get vault secret: [Errno 2] No such file or directory: '/share/ansible_vault_password.key'\u001b[0m",
                                "\u001b[32m2026-02-24 01:31:22\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mDropping encrypted entries\u001b[0m",
                                "+--------------------------------------+--------+----------+",
                                "| ID                                   | Name   | Status   |",
                                "|--------------------------------------+--------+----------|",
                                "| 36006f27-69d5-4fba-95d5-a09ad6e3d2b6 | test-3 | ACTIVE   |",
                                "| e237ead4-6261-44c0-95f1-03bd2fd83f6a | test-2 | ACTIVE   |",
                                "+--------------------------------------+--------+----------+",
                                "+ osism manage compute list testbed-node-5",
                                "\u001b[32m2026-02-24 01:31:26\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mUnable to get ansible vault password\u001b[0m",
                                "\u001b[32m2026-02-24 01:31:26\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mUnable to get vault secret: [Errno 2] No such file or directory: '/share/ansible_vault_password.key'\u001b[0m",
                                "\u001b[32m2026-02-24 01:31:26\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mDropping encrypted entries\u001b[0m",
                                "+------+--------+----------+",
                                "| ID   | Name   | Status   |",
                                "|------+--------+----------|",
                                "+------+--------+----------+",
                                "+ osism manage compute migrate --yes --target testbed-node-3 testbed-node-4",
                                "\u001b[32m2026-02-24 01:31:29\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mUnable to get ansible vault password\u001b[0m",
                                "\u001b[32m2026-02-24 01:31:29\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mUnable to get vault secret: [Errno 2] No such file or directory: '/share/ansible_vault_password.key'\u001b[0m",
                                "\u001b[32m2026-02-24 01:31:29\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mDropping encrypted entries\u001b[0m",
                                "\u001b[32m2026-02-24 01:31:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migrating server 36006f27-69d5-4fba-95d5-a09ad6e3d2b6\u001b[0m",
                                "\u001b[32m2026-02-24 01:31:43\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 36006f27-69d5-4fba-95d5-a09ad6e3d2b6 (test-3) is still in progress\u001b[0m",
                                "\u001b[32m2026-02-24 01:31:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 36006f27-69d5-4fba-95d5-a09ad6e3d2b6 (test-3) is still in progress\u001b[0m",
                                "\u001b[32m2026-02-24 01:31:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 36006f27-69d5-4fba-95d5-a09ad6e3d2b6 (test-3) is still in progress\u001b[0m",
                                "\u001b[32m2026-02-24 01:31:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 36006f27-69d5-4fba-95d5-a09ad6e3d2b6 (test-3) is still in progress\u001b[0m",
                                "\u001b[32m2026-02-24 01:31:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 36006f27-69d5-4fba-95d5-a09ad6e3d2b6 (test-3) is still in progress\u001b[0m",
                                "\u001b[32m2026-02-24 01:31:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 36006f27-69d5-4fba-95d5-a09ad6e3d2b6 (test-3) is still in progress\u001b[0m",
                                "\u001b[32m2026-02-24 01:31:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 36006f27-69d5-4fba-95d5-a09ad6e3d2b6 (test-3) is still in progress\u001b[0m",
                                "\u001b[32m2026-02-24 01:32:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 36006f27-69d5-4fba-95d5-a09ad6e3d2b6 (test-3) is still in progress\u001b[0m",
                                "\u001b[32m2026-02-24 01:32:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 36006f27-69d5-4fba-95d5-a09ad6e3d2b6 (test-3) completed with status ACTIVE\u001b[0m",
                                "\u001b[32m2026-02-24 01:32:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migrating server e237ead4-6261-44c0-95f1-03bd2fd83f6a\u001b[0m",
                                "\u001b[32m2026-02-24 01:32:13\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of e237ead4-6261-44c0-95f1-03bd2fd83f6a (test-2) is still in progress\u001b[0m",
                                "\u001b[32m2026-02-24 01:32:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of e237ead4-6261-44c0-95f1-03bd2fd83f6a (test-2) is still in progress\u001b[0m",
                                "\u001b[32m2026-02-24 01:32:18\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of e237ead4-6261-44c0-95f1-03bd2fd83f6a (test-2) is still in progress\u001b[0m",
                                "\u001b[32m2026-02-24 01:32:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of e237ead4-6261-44c0-95f1-03bd2fd83f6a (test-2) is still in progress\u001b[0m",
                                "\u001b[32m2026-02-24 01:32:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of e237ead4-6261-44c0-95f1-03bd2fd83f6a (test-2) is still in progress\u001b[0m",
                                "\u001b[32m2026-02-24 01:32:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of e237ead4-6261-44c0-95f1-03bd2fd83f6a (test-2) is still in progress\u001b[0m",
                                "\u001b[32m2026-02-24 01:32:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of e237ead4-6261-44c0-95f1-03bd2fd83f6a (test-2) is still in progress\u001b[0m",
                                "\u001b[32m2026-02-24 01:32:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of e237ead4-6261-44c0-95f1-03bd2fd83f6a (test-2) is still in progress\u001b[0m",
                                "\u001b[32m2026-02-24 01:32:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of e237ead4-6261-44c0-95f1-03bd2fd83f6a (test-2) completed with status ACTIVE\u001b[0m",
                                "+ compute_list",
                                "+ osism manage compute list testbed-node-3",
                                "\u001b[32m2026-02-24 01:32:35\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mUnable to get ansible vault password\u001b[0m",
                                "\u001b[32m2026-02-24 01:32:35\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mUnable to get vault secret: [Errno 2] No such file or directory: '/share/ansible_vault_password.key'\u001b[0m",
                                "\u001b[32m2026-02-24 01:32:35\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mDropping encrypted entries\u001b[0m",
                                "+--------------------------------------+--------+----------+",
                                "| ID                                   | Name   | Status   |",
                                "|--------------------------------------+--------+----------|",
                                "| 0b9309c0-5fce-46a5-86f1-8d5360f5077a | test-4 | ACTIVE   |",
                                "| 36006f27-69d5-4fba-95d5-a09ad6e3d2b6 | test-3 | ACTIVE   |",
                                "| 5ca62725-a636-4f3b-bb5e-7b2c661006b6 | test-1 | ACTIVE   |",
                                "| 73a9cbe2-a4c7-4bb9-be1d-0f4606a87824 | test   | ACTIVE   |",
                                "| e237ead4-6261-44c0-95f1-03bd2fd83f6a | test-2 | ACTIVE   |",
                                "+--------------------------------------+--------+----------+",
                                "+ osism manage compute list testbed-node-4",
                                "\u001b[32m2026-02-24 01:32:39\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mUnable to get ansible vault password\u001b[0m",
                                "\u001b[32m2026-02-24 01:32:39\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mUnable to get vault secret: [Errno 2] No such file or directory: '/share/ansible_vault_password.key'\u001b[0m",
                                "\u001b[32m2026-02-24 01:32:39\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mDropping encrypted entries\u001b[0m",
                                "+------+--------+----------+",
                                "| ID   | Name   | Status   |",
                                "|------+--------+----------|",
                                "+------+--------+----------+",
                                "+ osism manage compute list testbed-node-5",
                                "\u001b[32m2026-02-24 01:32:43\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mUnable to get ansible vault password\u001b[0m",
                                "\u001b[32m2026-02-24 01:32:43\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mUnable to get vault secret: [Errno 2] No such file or directory: '/share/ansible_vault_password.key'\u001b[0m",
                                "\u001b[32m2026-02-24 01:32:43\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mDropping encrypted entries\u001b[0m",
                                "+------+--------+----------+",
                                "| ID   | Name   | Status   |",
                                "|------+--------+----------|",
                                "+------+--------+----------+",
                                "+ server_ping",
                                "++ openstack --os-cloud test floating ip list --status ACTIVE -f value -c 'Floating IP Address'",
                                "++ tr -d '\\r'",
                                "+ for address in $(openstack --os-cloud test floating ip list --status ACTIVE -f value -c \"Floating IP Address\" | tr -d '\\r')",
                                "+ ping -c3 192.168.112.193",
                                "PING 192.168.112.193 (192.168.112.193) 56(84) bytes of data.",
                                "64 bytes from 192.168.112.193: icmp_seq=1 ttl=63 time=8.80 ms",
                                "64 bytes from 192.168.112.193: icmp_seq=2 ttl=63 time=2.40 ms",
                                "64 bytes from 192.168.112.193: icmp_seq=3 ttl=63 time=1.92 ms",
                                "",
                                "--- 192.168.112.193 ping statistics ---",
                                "3 packets transmitted, 3 received, 0% packet loss, time 2003ms",
                                "rtt min/avg/max/mdev = 1.918/4.373/8.802/3.137 ms",
                                "+ for address in $(openstack --os-cloud test floating ip list --status ACTIVE -f value -c \"Floating IP Address\" | tr -d '\\r')",
                                "+ ping -c3 192.168.112.184",
                                "PING 192.168.112.184 (192.168.112.184) 56(84) bytes of data.",
                                "64 bytes from 192.168.112.184: icmp_seq=1 ttl=63 time=5.03 ms",
                                "64 bytes from 192.168.112.184: icmp_seq=2 ttl=63 time=2.81 ms",
                                "64 bytes from 192.168.112.184: icmp_seq=3 ttl=63 time=2.00 ms",
                                "",
                                "--- 192.168.112.184 ping statistics ---",
                                "3 packets transmitted, 3 received, 0% packet loss, time 2003ms",
                                "rtt min/avg/max/mdev = 2.002/3.281/5.033/1.281 ms",
                                "+ for address in $(openstack --os-cloud test floating ip list --status ACTIVE -f value -c \"Floating IP Address\" | tr -d '\\r')",
                                "+ ping -c3 192.168.112.182",
                                "PING 192.168.112.182 (192.168.112.182) 56(84) bytes of data.",
                                "64 bytes from 192.168.112.182: icmp_seq=1 ttl=63 time=6.10 ms",
                                "64 bytes from 192.168.112.182: icmp_seq=2 ttl=63 time=1.99 ms",
                                "64 bytes from 192.168.112.182: icmp_seq=3 ttl=63 time=1.81 ms",
                                "",
                                "--- 192.168.112.182 ping statistics ---",
                                "3 packets transmitted, 3 received, 0% packet loss, time 2002ms",
                                "rtt min/avg/max/mdev = 1.807/3.300/6.104/1.983 ms",
                                "+ for address in $(openstack --os-cloud test floating ip list --status ACTIVE -f value -c \"Floating IP Address\" | tr -d '\\r')",
                                "+ ping -c3 192.168.112.141",
                                "PING 192.168.112.141 (192.168.112.141) 56(84) bytes of data.",
                                "64 bytes from 192.168.112.141: icmp_seq=1 ttl=63 time=5.24 ms",
                                "64 bytes from 192.168.112.141: icmp_seq=2 ttl=63 time=1.88 ms",
                                "64 bytes from 192.168.112.141: icmp_seq=3 ttl=63 time=1.67 ms",
                                "",
                                "--- 192.168.112.141 ping statistics ---",
                                "3 packets transmitted, 3 received, 0% packet loss, time 2003ms",
                                "rtt min/avg/max/mdev = 1.668/2.928/5.236/1.634 ms",
                                "+ for address in $(openstack --os-cloud test floating ip list --status ACTIVE -f value -c \"Floating IP Address\" | tr -d '\\r')",
                                "+ ping -c3 192.168.112.192",
                                "PING 192.168.112.192 (192.168.112.192) 56(84) bytes of data.",
                                "64 bytes from 192.168.112.192: icmp_seq=1 ttl=63 time=5.20 ms",
                                "64 bytes from 192.168.112.192: icmp_seq=2 ttl=63 time=2.93 ms",
                                "64 bytes from 192.168.112.192: icmp_seq=3 ttl=63 time=2.17 ms",
                                "",
                                "--- 192.168.112.192 ping statistics ---",
                                "3 packets transmitted, 3 received, 0% packet loss, time 2003ms",
                                "rtt min/avg/max/mdev = 2.171/3.433/5.200/1.286 ms",
                                "+ osism manage compute migrate --yes --target testbed-node-3 testbed-node-5",
                                "\u001b[32m2026-02-24 01:33:00\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mUnable to get ansible vault password\u001b[0m",
                                "\u001b[32m2026-02-24 01:33:00\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mUnable to get vault secret: [Errno 2] No such file or directory: '/share/ansible_vault_password.key'\u001b[0m",
                                "\u001b[32m2026-02-24 01:33:00\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mDropping encrypted entries\u001b[0m",
                                "\u001b[32m2026-02-24 01:33:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mNo migratable instances found on node testbed-node-5\u001b[0m",
                                "+ compute_list",
                                "+ osism manage compute list testbed-node-3",
                                "\u001b[32m2026-02-24 01:33:04\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mUnable to get ansible vault password\u001b[0m",
                                "\u001b[32m2026-02-24 01:33:04\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mUnable to get vault secret: [Errno 2] No such file or directory: '/share/ansible_vault_password.key'\u001b[0m",
                                "\u001b[32m2026-02-24 01:33:04\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mDropping encrypted entries\u001b[0m",
                                "+--------------------------------------+--------+----------+",
                                "| ID                                   | Name   | Status   |",
                                "|--------------------------------------+--------+----------|",
                                "| 0b9309c0-5fce-46a5-86f1-8d5360f5077a | test-4 | ACTIVE   |",
                                "| 36006f27-69d5-4fba-95d5-a09ad6e3d2b6 | test-3 | ACTIVE   |",
                                "| 5ca62725-a636-4f3b-bb5e-7b2c661006b6 | test-1 | ACTIVE   |",
                                "| 73a9cbe2-a4c7-4bb9-be1d-0f4606a87824 | test   | ACTIVE   |",
                                "| e237ead4-6261-44c0-95f1-03bd2fd83f6a | test-2 | ACTIVE   |",
                                "+--------------------------------------+--------+----------+",
                                "+ osism manage compute list testbed-node-4",
                                "\u001b[32m2026-02-24 01:33:08\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mUnable to get ansible vault password\u001b[0m",
                                "\u001b[32m2026-02-24 01:33:08\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mUnable to get vault secret: [Errno 2] No such file or directory: '/share/ansible_vault_password.key'\u001b[0m",
                                "\u001b[32m2026-02-24 01:33:08\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mDropping encrypted entries\u001b[0m",
                                "+------+--------+----------+",
                                "| ID   | Name   | Status   |",
                                "|------+--------+----------|",
                                "+------+--------+----------+",
                                "+ osism manage compute list testbed-node-5",
                                "\u001b[32m2026-02-24 01:33:12\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mUnable to get ansible vault password\u001b[0m",
                                "\u001b[32m2026-02-24 01:33:12\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mUnable to get vault secret: [Errno 2] No such file or directory: '/share/ansible_vault_password.key'\u001b[0m",
                                "\u001b[32m2026-02-24 01:33:12\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mDropping encrypted entries\u001b[0m",
                                "+------+--------+----------+",
                                "| ID   | Name   | Status   |",
                                "|------+--------+----------|",
                                "+------+--------+----------+",
                                "+ server_ping",
                                "++ openstack --os-cloud test floating ip list --status ACTIVE -f value -c 'Floating IP Address'",
                                "++ tr -d '\\r'",
                                "+ for address in $(openstack --os-cloud test floating ip list --status ACTIVE -f value -c \"Floating IP Address\" | tr -d '\\r')",
                                "+ ping -c3 192.168.112.193",
                                "PING 192.168.112.193 (192.168.112.193) 56(84) bytes of data.",
                                "64 bytes from 192.168.112.193: icmp_seq=1 ttl=63 time=12.5 ms",
                                "64 bytes from 192.168.112.193: icmp_seq=2 ttl=63 time=2.48 ms",
                                "64 bytes from 192.168.112.193: icmp_seq=3 ttl=63 time=2.23 ms",
                                "",
                                "--- 192.168.112.193 ping statistics ---",
                                "3 packets transmitted, 3 received, 0% packet loss, time 2003ms",
                                "rtt min/avg/max/mdev = 2.231/5.726/12.466/4.766 ms",
                                "+ for address in $(openstack --os-cloud test floating ip list --status ACTIVE -f value -c \"Floating IP Address\" | tr -d '\\r')",
                                "+ ping -c3 192.168.112.184",
                                "PING 192.168.112.184 (192.168.112.184) 56(84) bytes of data.",
                                "64 bytes from 192.168.112.184: icmp_seq=1 ttl=63 time=5.77 ms",
                                "64 bytes from 192.168.112.184: icmp_seq=2 ttl=63 time=2.71 ms",
                                "64 bytes from 192.168.112.184: icmp_seq=3 ttl=63 time=2.21 ms",
                                "",
                                "--- 192.168.112.184 ping statistics ---",
                                "3 packets transmitted, 3 received, 0% packet loss, time 2002ms",
                                "rtt min/avg/max/mdev = 2.211/3.564/5.770/1.572 ms",
                                "+ for address in $(openstack --os-cloud test floating ip list --status ACTIVE -f value -c \"Floating IP Address\" | tr -d '\\r')",
                                "+ ping -c3 192.168.112.182",
                                "PING 192.168.112.182 (192.168.112.182) 56(84) bytes of data.",
                                "64 bytes from 192.168.112.182: icmp_seq=1 ttl=63 time=5.63 ms",
                                "64 bytes from 192.168.112.182: icmp_seq=2 ttl=63 time=2.95 ms",
                                "64 bytes from 192.168.112.182: icmp_seq=3 ttl=63 time=1.93 ms",
                                "",
                                "--- 192.168.112.182 ping statistics ---",
                                "3 packets transmitted, 3 received, 0% packet loss, time 2003ms",
                                "rtt min/avg/max/mdev = 1.928/3.500/5.626/1.559 ms",
                                "+ for address in $(openstack --os-cloud test floating ip list --status ACTIVE -f value -c \"Floating IP Address\" | tr -d '\\r')",
                                "+ ping -c3 192.168.112.141",
                                "PING 192.168.112.141 (192.168.112.141) 56(84) bytes of data.",
                                "64 bytes from 192.168.112.141: icmp_seq=1 ttl=63 time=5.45 ms",
                                "64 bytes from 192.168.112.141: icmp_seq=2 ttl=63 time=1.91 ms",
                                "64 bytes from 192.168.112.141: icmp_seq=3 ttl=63 time=2.18 ms",
                                "",
                                "--- 192.168.112.141 ping statistics ---",
                                "3 packets transmitted, 3 received, 0% packet loss, time 2002ms",
                                "rtt min/avg/max/mdev = 1.908/3.182/5.454/1.610 ms",
                                "+ for address in $(openstack --os-cloud test floating ip list --status ACTIVE -f value -c \"Floating IP Address\" | tr -d '\\r')",
                                "+ ping -c3 192.168.112.192",
                                "PING 192.168.112.192 (192.168.112.192) 56(84) bytes of data.",
                                "64 bytes from 192.168.112.192: icmp_seq=1 ttl=63 time=6.51 ms",
                                "64 bytes from 192.168.112.192: icmp_seq=2 ttl=63 time=2.55 ms",
                                "64 bytes from 192.168.112.192: icmp_seq=3 ttl=63 time=2.73 ms",
                                "",
                                "--- 192.168.112.192 ping statistics ---",
                                "3 packets transmitted, 3 received, 0% packet loss, time 2003ms",
                                "rtt min/avg/max/mdev = 2.548/3.928/6.505/1.823 ms",
                                "+ osism manage compute migrate --yes --target testbed-node-4 testbed-node-3",
                                "\u001b[32m2026-02-24 01:33:29\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mUnable to get ansible vault password\u001b[0m",
                                "\u001b[32m2026-02-24 01:33:29\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mUnable to get vault secret: [Errno 2] No such file or directory: '/share/ansible_vault_password.key'\u001b[0m",
                                "\u001b[32m2026-02-24 01:33:29\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mDropping encrypted entries\u001b[0m",
                                "\u001b[32m2026-02-24 01:33:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migrating server 0b9309c0-5fce-46a5-86f1-8d5360f5077a\u001b[0m",
                                "\u001b[32m2026-02-24 01:33:42\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 0b9309c0-5fce-46a5-86f1-8d5360f5077a (test-4) is still in progress\u001b[0m",
                                "\u001b[32m2026-02-24 01:33:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 0b9309c0-5fce-46a5-86f1-8d5360f5077a (test-4) is still in progress\u001b[0m",
                                "\u001b[32m2026-02-24 01:33:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 0b9309c0-5fce-46a5-86f1-8d5360f5077a (test-4) is still in progress\u001b[0m",
                                "\u001b[32m2026-02-24 01:33:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 0b9309c0-5fce-46a5-86f1-8d5360f5077a (test-4) is still in progress\u001b[0m",
                                "\u001b[32m2026-02-24 01:33:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 0b9309c0-5fce-46a5-86f1-8d5360f5077a (test-4) is still in progress\u001b[0m",
                                "\u001b[32m2026-02-24 01:33:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 0b9309c0-5fce-46a5-86f1-8d5360f5077a (test-4) is still in progress\u001b[0m",
                                "\u001b[32m2026-02-24 01:33:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 0b9309c0-5fce-46a5-86f1-8d5360f5077a (test-4) is still in progress\u001b[0m",
                                "\u001b[32m2026-02-24 01:33:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 0b9309c0-5fce-46a5-86f1-8d5360f5077a (test-4) is still in progress\u001b[0m",
                                "\u001b[32m2026-02-24 01:34:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 0b9309c0-5fce-46a5-86f1-8d5360f5077a (test-4) completed with status ACTIVE\u001b[0m",
                                "\u001b[32m2026-02-24 01:34:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migrating server 36006f27-69d5-4fba-95d5-a09ad6e3d2b6\u001b[0m",
                                "\u001b[32m2026-02-24 01:34:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 36006f27-69d5-4fba-95d5-a09ad6e3d2b6 (test-3) is still in progress\u001b[0m",
                                "\u001b[32m2026-02-24 01:34:14\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 36006f27-69d5-4fba-95d5-a09ad6e3d2b6 (test-3) is still in progress\u001b[0m",
                                "\u001b[32m2026-02-24 01:34:16\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 36006f27-69d5-4fba-95d5-a09ad6e3d2b6 (test-3) is still in progress\u001b[0m",
                                "\u001b[32m2026-02-24 01:34:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 36006f27-69d5-4fba-95d5-a09ad6e3d2b6 (test-3) is still in progress\u001b[0m",
                                "\u001b[32m2026-02-24 01:34:21\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 36006f27-69d5-4fba-95d5-a09ad6e3d2b6 (test-3) is still in progress\u001b[0m",
                                "\u001b[32m2026-02-24 01:34:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 36006f27-69d5-4fba-95d5-a09ad6e3d2b6 (test-3) is still in progress\u001b[0m",
                                "\u001b[32m2026-02-24 01:34:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 36006f27-69d5-4fba-95d5-a09ad6e3d2b6 (test-3) is still in progress\u001b[0m",
                                "\u001b[32m2026-02-24 01:34:28\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 36006f27-69d5-4fba-95d5-a09ad6e3d2b6 (test-3) is still in progress\u001b[0m",
                                "\u001b[32m2026-02-24 01:34:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 36006f27-69d5-4fba-95d5-a09ad6e3d2b6 (test-3) completed with status ACTIVE\u001b[0m",
                                "\u001b[32m2026-02-24 01:34:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migrating server 5ca62725-a636-4f3b-bb5e-7b2c661006b6\u001b[0m",
                                "\u001b[32m2026-02-24 01:34:41\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 5ca62725-a636-4f3b-bb5e-7b2c661006b6 (test-1) is still in progress\u001b[0m",
                                "\u001b[32m2026-02-24 01:34:44\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 5ca62725-a636-4f3b-bb5e-7b2c661006b6 (test-1) is still in progress\u001b[0m",
                                "\u001b[32m2026-02-24 01:34:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 5ca62725-a636-4f3b-bb5e-7b2c661006b6 (test-1) is still in progress\u001b[0m",
                                "\u001b[32m2026-02-24 01:34:48\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 5ca62725-a636-4f3b-bb5e-7b2c661006b6 (test-1) is still in progress\u001b[0m",
                                "\u001b[32m2026-02-24 01:34:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 5ca62725-a636-4f3b-bb5e-7b2c661006b6 (test-1) is still in progress\u001b[0m",
                                "\u001b[32m2026-02-24 01:34:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 5ca62725-a636-4f3b-bb5e-7b2c661006b6 (test-1) is still in progress\u001b[0m",
                                "\u001b[32m2026-02-24 01:34:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 5ca62725-a636-4f3b-bb5e-7b2c661006b6 (test-1) is still in progress\u001b[0m",
                                "\u001b[32m2026-02-24 01:34:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 5ca62725-a636-4f3b-bb5e-7b2c661006b6 (test-1) is still in progress\u001b[0m",
                                "\u001b[32m2026-02-24 01:35:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 5ca62725-a636-4f3b-bb5e-7b2c661006b6 (test-1) completed with status ACTIVE\u001b[0m",
                                "\u001b[32m2026-02-24 01:35:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migrating server 73a9cbe2-a4c7-4bb9-be1d-0f4606a87824\u001b[0m",
                                "\u001b[32m2026-02-24 01:35:12\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 73a9cbe2-a4c7-4bb9-be1d-0f4606a87824 (test) is still in progress\u001b[0m",
                                "\u001b[32m2026-02-24 01:35:15\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 73a9cbe2-a4c7-4bb9-be1d-0f4606a87824 (test) is still in progress\u001b[0m",
                                "\u001b[32m2026-02-24 01:35:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 73a9cbe2-a4c7-4bb9-be1d-0f4606a87824 (test) is still in progress\u001b[0m",
                                "\u001b[32m2026-02-24 01:35:19\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 73a9cbe2-a4c7-4bb9-be1d-0f4606a87824 (test) is still in progress\u001b[0m",
                                "\u001b[32m2026-02-24 01:35:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 73a9cbe2-a4c7-4bb9-be1d-0f4606a87824 (test) is still in progress\u001b[0m",
                                "\u001b[32m2026-02-24 01:35:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 73a9cbe2-a4c7-4bb9-be1d-0f4606a87824 (test) is still in progress\u001b[0m",
                                "\u001b[32m2026-02-24 01:35:26\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 73a9cbe2-a4c7-4bb9-be1d-0f4606a87824 (test) is still in progress\u001b[0m",
                                "\u001b[32m2026-02-24 01:35:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 73a9cbe2-a4c7-4bb9-be1d-0f4606a87824 (test) is still in progress\u001b[0m",
                                "\u001b[32m2026-02-24 01:35:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 73a9cbe2-a4c7-4bb9-be1d-0f4606a87824 (test) is still in progress\u001b[0m",
                                "\u001b[32m2026-02-24 01:35:33\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 73a9cbe2-a4c7-4bb9-be1d-0f4606a87824 (test) is still in progress\u001b[0m",
                                "\u001b[32m2026-02-24 01:35:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 73a9cbe2-a4c7-4bb9-be1d-0f4606a87824 (test) completed with status ACTIVE\u001b[0m",
                                "\u001b[32m2026-02-24 01:35:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migrating server e237ead4-6261-44c0-95f1-03bd2fd83f6a\u001b[0m",
                                "\u001b[32m2026-02-24 01:35:46\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of e237ead4-6261-44c0-95f1-03bd2fd83f6a (test-2) is still in progress\u001b[0m",
                                "\u001b[32m2026-02-24 01:35:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of e237ead4-6261-44c0-95f1-03bd2fd83f6a (test-2) is still in progress\u001b[0m",
                                "\u001b[32m2026-02-24 01:35:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of e237ead4-6261-44c0-95f1-03bd2fd83f6a (test-2) is still in progress\u001b[0m",
                                "\u001b[32m2026-02-24 01:35:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of e237ead4-6261-44c0-95f1-03bd2fd83f6a (test-2) is still in progress\u001b[0m",
                                "\u001b[32m2026-02-24 01:35:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of e237ead4-6261-44c0-95f1-03bd2fd83f6a (test-2) is still in progress\u001b[0m",
                                "\u001b[32m2026-02-24 01:35:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of e237ead4-6261-44c0-95f1-03bd2fd83f6a (test-2) is still in progress\u001b[0m",
                                "\u001b[32m2026-02-24 01:36:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of e237ead4-6261-44c0-95f1-03bd2fd83f6a (test-2) is still in progress\u001b[0m",
                                "\u001b[32m2026-02-24 01:36:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of e237ead4-6261-44c0-95f1-03bd2fd83f6a (test-2) is still in progress\u001b[0m",
                                "\u001b[32m2026-02-24 01:36:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of e237ead4-6261-44c0-95f1-03bd2fd83f6a (test-2) completed with status ACTIVE\u001b[0m",
                                "+ compute_list",
                                "+ osism manage compute list testbed-node-3",
                                "\u001b[32m2026-02-24 01:36:07\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mUnable to get ansible vault password\u001b[0m",
                                "\u001b[32m2026-02-24 01:36:07\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mUnable to get vault secret: [Errno 2] No such file or directory: '/share/ansible_vault_password.key'\u001b[0m",
                                "\u001b[32m2026-02-24 01:36:07\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mDropping encrypted entries\u001b[0m",
                                "+------+--------+----------+",
                                "| ID   | Name   | Status   |",
                                "|------+--------+----------|",
                                "+------+--------+----------+",
                                "+ osism manage compute list testbed-node-4",
                                "\u001b[32m2026-02-24 01:36:11\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mUnable to get ansible vault password\u001b[0m",
                                "\u001b[32m2026-02-24 01:36:11\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mUnable to get vault secret: [Errno 2] No such file or directory: '/share/ansible_vault_password.key'\u001b[0m",
                                "\u001b[32m2026-02-24 01:36:11\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mDropping encrypted entries\u001b[0m",
                                "+--------------------------------------+--------+----------+",
                                "| ID                                   | Name   | Status   |",
                                "|--------------------------------------+--------+----------|",
                                "| 0b9309c0-5fce-46a5-86f1-8d5360f5077a | test-4 | ACTIVE   |",
                                "| 36006f27-69d5-4fba-95d5-a09ad6e3d2b6 | test-3 | ACTIVE   |",
                                "| 5ca62725-a636-4f3b-bb5e-7b2c661006b6 | test-1 | ACTIVE   |",
                                "| 73a9cbe2-a4c7-4bb9-be1d-0f4606a87824 | test   | ACTIVE   |",
                                "| e237ead4-6261-44c0-95f1-03bd2fd83f6a | test-2 | ACTIVE   |",
                                "+--------------------------------------+--------+----------+",
                                "+ osism manage compute list testbed-node-5",
                                "\u001b[32m2026-02-24 01:36:14\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mUnable to get ansible vault password\u001b[0m",
                                "\u001b[32m2026-02-24 01:36:14\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mUnable to get vault secret: [Errno 2] No such file or directory: '/share/ansible_vault_password.key'\u001b[0m",
                                "\u001b[32m2026-02-24 01:36:14\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mDropping encrypted entries\u001b[0m",
                                "+------+--------+----------+",
                                "| ID   | Name   | Status   |",
                                "|------+--------+----------|",
                                "+------+--------+----------+",
                                "+ server_ping",
                                "++ tr -d '\\r'",
                                "++ openstack --os-cloud test floating ip list --status ACTIVE -f value -c 'Floating IP Address'",
                                "+ for address in $(openstack --os-cloud test floating ip list --status ACTIVE -f value -c \"Floating IP Address\" | tr -d '\\r')",
                                "+ ping -c3 192.168.112.193",
                                "PING 192.168.112.193 (192.168.112.193) 56(84) bytes of data.",
                                "64 bytes from 192.168.112.193: icmp_seq=1 ttl=63 time=9.30 ms",
                                "64 bytes from 192.168.112.193: icmp_seq=2 ttl=63 time=2.07 ms",
                                "64 bytes from 192.168.112.193: icmp_seq=3 ttl=63 time=1.91 ms",
                                "",
                                "--- 192.168.112.193 ping statistics ---",
                                "3 packets transmitted, 3 received, 0% packet loss, time 2003ms",
                                "rtt min/avg/max/mdev = 1.912/4.427/9.296/3.443 ms",
                                "+ for address in $(openstack --os-cloud test floating ip list --status ACTIVE -f value -c \"Floating IP Address\" | tr -d '\\r')",
                                "+ ping -c3 192.168.112.184",
                                "PING 192.168.112.184 (192.168.112.184) 56(84) bytes of data.",
                                "64 bytes from 192.168.112.184: icmp_seq=1 ttl=63 time=6.00 ms",
                                "64 bytes from 192.168.112.184: icmp_seq=2 ttl=63 time=2.92 ms",
                                "64 bytes from 192.168.112.184: icmp_seq=3 ttl=63 time=1.82 ms",
                                "",
                                "--- 192.168.112.184 ping statistics ---",
                                "3 packets transmitted, 3 received, 0% packet loss, time 2003ms",
                                "rtt min/avg/max/mdev = 1.822/3.582/6.001/1.768 ms",
                                "+ for address in $(openstack --os-cloud test floating ip list --status ACTIVE -f value -c \"Floating IP Address\" | tr -d '\\r')",
                                "+ ping -c3 192.168.112.182",
                                "PING 192.168.112.182 (192.168.112.182) 56(84) bytes of data.",
                                "64 bytes from 192.168.112.182: icmp_seq=1 ttl=63 time=4.97 ms",
                                "64 bytes from 192.168.112.182: icmp_seq=2 ttl=63 time=2.17 ms",
                                "64 bytes from 192.168.112.182: icmp_seq=3 ttl=63 time=1.76 ms",
                                "",
                                "--- 192.168.112.182 ping statistics ---",
                                "3 packets transmitted, 3 received, 0% packet loss, time 2002ms",
                                "rtt min/avg/max/mdev = 1.762/2.968/4.970/1.425 ms",
                                "+ for address in $(openstack --os-cloud test floating ip list --status ACTIVE -f value -c \"Floating IP Address\" | tr -d '\\r')",
                                "+ ping -c3 192.168.112.141",
                                "PING 192.168.112.141 (192.168.112.141) 56(84) bytes of data.",
                                "64 bytes from 192.168.112.141: icmp_seq=1 ttl=63 time=5.99 ms",
                                "64 bytes from 192.168.112.141: icmp_seq=2 ttl=63 time=2.14 ms",
                                "64 bytes from 192.168.112.141: icmp_seq=3 ttl=63 time=1.76 ms",
                                "",
                                "--- 192.168.112.141 ping statistics ---",
                                "3 packets transmitted, 3 received, 0% packet loss, time 2002ms",
                                "rtt min/avg/max/mdev = 1.759/3.298/5.991/1.910 ms",
                                "+ for address in $(openstack --os-cloud test floating ip list --status ACTIVE -f value -c \"Floating IP Address\" | tr -d '\\r')",
                                "+ ping -c3 192.168.112.192",
                                "PING 192.168.112.192 (192.168.112.192) 56(84) bytes of data.",
                                "64 bytes from 192.168.112.192: icmp_seq=1 ttl=63 time=5.50 ms",
                                "64 bytes from 192.168.112.192: icmp_seq=2 ttl=63 time=1.96 ms",
                                "64 bytes from 192.168.112.192: icmp_seq=3 ttl=63 time=1.34 ms",
                                "",
                                "--- 192.168.112.192 ping statistics ---",
                                "3 packets transmitted, 3 received, 0% packet loss, time 2003ms",
                                "rtt min/avg/max/mdev = 1.335/2.933/5.504/1.835 ms",
                                "+ osism manage compute migrate --yes --target testbed-node-5 testbed-node-4",
                                "\u001b[32m2026-02-24 01:36:31\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mUnable to get ansible vault password\u001b[0m",
                                "\u001b[32m2026-02-24 01:36:31\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mUnable to get vault secret: [Errno 2] No such file or directory: '/share/ansible_vault_password.key'\u001b[0m",
                                "\u001b[32m2026-02-24 01:36:31\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mDropping encrypted entries\u001b[0m",
                                "\u001b[32m2026-02-24 01:36:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migrating server 0b9309c0-5fce-46a5-86f1-8d5360f5077a\u001b[0m",
                                "\u001b[32m2026-02-24 01:36:45\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 0b9309c0-5fce-46a5-86f1-8d5360f5077a (test-4) is still in progress\u001b[0m",
                                "\u001b[32m2026-02-24 01:36:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 0b9309c0-5fce-46a5-86f1-8d5360f5077a (test-4) is still in progress\u001b[0m",
                                "\u001b[32m2026-02-24 01:36:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 0b9309c0-5fce-46a5-86f1-8d5360f5077a (test-4) is still in progress\u001b[0m",
                                "\u001b[32m2026-02-24 01:36:52\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 0b9309c0-5fce-46a5-86f1-8d5360f5077a (test-4) is still in progress\u001b[0m",
                                "\u001b[32m2026-02-24 01:36:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 0b9309c0-5fce-46a5-86f1-8d5360f5077a (test-4) is still in progress\u001b[0m",
                                "\u001b[32m2026-02-24 01:36:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 0b9309c0-5fce-46a5-86f1-8d5360f5077a (test-4) is still in progress\u001b[0m",
                                "\u001b[32m2026-02-24 01:36:59\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 0b9309c0-5fce-46a5-86f1-8d5360f5077a (test-4) is still in progress\u001b[0m",
                                "\u001b[32m2026-02-24 01:37:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 0b9309c0-5fce-46a5-86f1-8d5360f5077a (test-4) is still in progress\u001b[0m",
                                "\u001b[32m2026-02-24 01:37:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 0b9309c0-5fce-46a5-86f1-8d5360f5077a (test-4) is still in progress\u001b[0m",
                                "\u001b[32m2026-02-24 01:37:06\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 0b9309c0-5fce-46a5-86f1-8d5360f5077a (test-4) is still in progress\u001b[0m",
                                "\u001b[32m2026-02-24 01:37:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 0b9309c0-5fce-46a5-86f1-8d5360f5077a (test-4) is still in progress\u001b[0m",
                                "\u001b[32m2026-02-24 01:37:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 0b9309c0-5fce-46a5-86f1-8d5360f5077a (test-4) completed with status ACTIVE\u001b[0m",
                                "\u001b[32m2026-02-24 01:37:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migrating server 36006f27-69d5-4fba-95d5-a09ad6e3d2b6\u001b[0m",
                                "\u001b[32m2026-02-24 01:37:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 36006f27-69d5-4fba-95d5-a09ad6e3d2b6 (test-3) is still in progress\u001b[0m",
                                "\u001b[32m2026-02-24 01:37:23\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 36006f27-69d5-4fba-95d5-a09ad6e3d2b6 (test-3) is still in progress\u001b[0m",
                                "\u001b[32m2026-02-24 01:37:25\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 36006f27-69d5-4fba-95d5-a09ad6e3d2b6 (test-3) is still in progress\u001b[0m",
                                "\u001b[32m2026-02-24 01:37:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 36006f27-69d5-4fba-95d5-a09ad6e3d2b6 (test-3) is still in progress\u001b[0m",
                                "\u001b[32m2026-02-24 01:37:30\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 36006f27-69d5-4fba-95d5-a09ad6e3d2b6 (test-3) is still in progress\u001b[0m",
                                "\u001b[32m2026-02-24 01:37:32\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 36006f27-69d5-4fba-95d5-a09ad6e3d2b6 (test-3) is still in progress\u001b[0m",
                                "\u001b[32m2026-02-24 01:37:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 36006f27-69d5-4fba-95d5-a09ad6e3d2b6 (test-3) is still in progress\u001b[0m",
                                "\u001b[32m2026-02-24 01:37:37\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 36006f27-69d5-4fba-95d5-a09ad6e3d2b6 (test-3) is still in progress\u001b[0m",
                                "\u001b[32m2026-02-24 01:37:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 36006f27-69d5-4fba-95d5-a09ad6e3d2b6 (test-3) completed with status ACTIVE\u001b[0m",
                                "\u001b[32m2026-02-24 01:37:39\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migrating server 5ca62725-a636-4f3b-bb5e-7b2c661006b6\u001b[0m",
                                "\u001b[32m2026-02-24 01:37:49\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 5ca62725-a636-4f3b-bb5e-7b2c661006b6 (test-1) is still in progress\u001b[0m",
                                "\u001b[32m2026-02-24 01:37:51\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 5ca62725-a636-4f3b-bb5e-7b2c661006b6 (test-1) is still in progress\u001b[0m",
                                "\u001b[32m2026-02-24 01:37:54\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 5ca62725-a636-4f3b-bb5e-7b2c661006b6 (test-1) is still in progress\u001b[0m",
                                "\u001b[32m2026-02-24 01:37:56\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 5ca62725-a636-4f3b-bb5e-7b2c661006b6 (test-1) is still in progress\u001b[0m",
                                "\u001b[32m2026-02-24 01:37:58\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 5ca62725-a636-4f3b-bb5e-7b2c661006b6 (test-1) is still in progress\u001b[0m",
                                "\u001b[32m2026-02-24 01:38:01\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 5ca62725-a636-4f3b-bb5e-7b2c661006b6 (test-1) is still in progress\u001b[0m",
                                "\u001b[32m2026-02-24 01:38:03\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 5ca62725-a636-4f3b-bb5e-7b2c661006b6 (test-1) is still in progress\u001b[0m",
                                "\u001b[32m2026-02-24 01:38:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 5ca62725-a636-4f3b-bb5e-7b2c661006b6 (test-1) is still in progress\u001b[0m",
                                "\u001b[32m2026-02-24 01:38:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 5ca62725-a636-4f3b-bb5e-7b2c661006b6 (test-1) completed with status ACTIVE\u001b[0m",
                                "\u001b[32m2026-02-24 01:38:08\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migrating server 73a9cbe2-a4c7-4bb9-be1d-0f4606a87824\u001b[0m",
                                "\u001b[32m2026-02-24 01:38:17\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 73a9cbe2-a4c7-4bb9-be1d-0f4606a87824 (test) is still in progress\u001b[0m",
                                "\u001b[32m2026-02-24 01:38:20\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 73a9cbe2-a4c7-4bb9-be1d-0f4606a87824 (test) is still in progress\u001b[0m",
                                "\u001b[32m2026-02-24 01:38:22\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 73a9cbe2-a4c7-4bb9-be1d-0f4606a87824 (test) is still in progress\u001b[0m",
                                "\u001b[32m2026-02-24 01:38:24\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 73a9cbe2-a4c7-4bb9-be1d-0f4606a87824 (test) is still in progress\u001b[0m",
                                "\u001b[32m2026-02-24 01:38:27\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 73a9cbe2-a4c7-4bb9-be1d-0f4606a87824 (test) is still in progress\u001b[0m",
                                "\u001b[32m2026-02-24 01:38:29\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 73a9cbe2-a4c7-4bb9-be1d-0f4606a87824 (test) is still in progress\u001b[0m",
                                "\u001b[32m2026-02-24 01:38:31\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 73a9cbe2-a4c7-4bb9-be1d-0f4606a87824 (test) is still in progress\u001b[0m",
                                "\u001b[32m2026-02-24 01:38:34\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 73a9cbe2-a4c7-4bb9-be1d-0f4606a87824 (test) is still in progress\u001b[0m",
                                "\u001b[32m2026-02-24 01:38:36\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 73a9cbe2-a4c7-4bb9-be1d-0f4606a87824 (test) is still in progress\u001b[0m",
                                "\u001b[32m2026-02-24 01:38:38\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 73a9cbe2-a4c7-4bb9-be1d-0f4606a87824 (test) is still in progress\u001b[0m",
                                "\u001b[32m2026-02-24 01:38:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of 73a9cbe2-a4c7-4bb9-be1d-0f4606a87824 (test) completed with status ACTIVE\u001b[0m",
                                "\u001b[32m2026-02-24 01:38:40\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migrating server e237ead4-6261-44c0-95f1-03bd2fd83f6a\u001b[0m",
                                "\u001b[32m2026-02-24 01:38:50\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of e237ead4-6261-44c0-95f1-03bd2fd83f6a (test-2) is still in progress\u001b[0m",
                                "\u001b[32m2026-02-24 01:38:53\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of e237ead4-6261-44c0-95f1-03bd2fd83f6a (test-2) is still in progress\u001b[0m",
                                "\u001b[32m2026-02-24 01:38:55\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of e237ead4-6261-44c0-95f1-03bd2fd83f6a (test-2) is still in progress\u001b[0m",
                                "\u001b[32m2026-02-24 01:38:57\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of e237ead4-6261-44c0-95f1-03bd2fd83f6a (test-2) is still in progress\u001b[0m",
                                "\u001b[32m2026-02-24 01:39:00\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of e237ead4-6261-44c0-95f1-03bd2fd83f6a (test-2) is still in progress\u001b[0m",
                                "\u001b[32m2026-02-24 01:39:02\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of e237ead4-6261-44c0-95f1-03bd2fd83f6a (test-2) is still in progress\u001b[0m",
                                "\u001b[32m2026-02-24 01:39:05\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of e237ead4-6261-44c0-95f1-03bd2fd83f6a (test-2) is still in progress\u001b[0m",
                                "\u001b[32m2026-02-24 01:39:07\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of e237ead4-6261-44c0-95f1-03bd2fd83f6a (test-2) is still in progress\u001b[0m",
                                "\u001b[32m2026-02-24 01:39:10\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mLive migration of e237ead4-6261-44c0-95f1-03bd2fd83f6a (test-2) completed with status ACTIVE\u001b[0m",
                                "+ compute_list",
                                "+ osism manage compute list testbed-node-3",
                                "\u001b[32m2026-02-24 01:39:12\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mUnable to get ansible vault password\u001b[0m",
                                "\u001b[32m2026-02-24 01:39:12\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mUnable to get vault secret: [Errno 2] No such file or directory: '/share/ansible_vault_password.key'\u001b[0m",
                                "\u001b[32m2026-02-24 01:39:12\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mDropping encrypted entries\u001b[0m",
                                "+------+--------+----------+",
                                "| ID   | Name   | Status   |",
                                "|------+--------+----------|",
                                "+------+--------+----------+",
                                "+ osism manage compute list testbed-node-4",
                                "\u001b[32m2026-02-24 01:39:15\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mUnable to get ansible vault password\u001b[0m",
                                "\u001b[32m2026-02-24 01:39:15\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mUnable to get vault secret: [Errno 2] No such file or directory: '/share/ansible_vault_password.key'\u001b[0m",
                                "\u001b[32m2026-02-24 01:39:15\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mDropping encrypted entries\u001b[0m",
                                "+------+--------+----------+",
                                "| ID   | Name   | Status   |",
                                "|------+--------+----------|",
                                "+------+--------+----------+",
                                "+ osism manage compute list testbed-node-5",
                                "\u001b[32m2026-02-24 01:39:19\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mUnable to get ansible vault password\u001b[0m",
                                "\u001b[32m2026-02-24 01:39:19\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mUnable to get vault secret: [Errno 2] No such file or directory: '/share/ansible_vault_password.key'\u001b[0m",
                                "\u001b[32m2026-02-24 01:39:19\u001b[0m | \u001b[31m\u001b[1mERROR   \u001b[0m | \u001b[31m\u001b[1mDropping encrypted entries\u001b[0m",
                                "+--------------------------------------+--------+----------+",
                                "| ID                                   | Name   | Status   |",
                                "|--------------------------------------+--------+----------|",
                                "| 0b9309c0-5fce-46a5-86f1-8d5360f5077a | test-4 | ACTIVE   |",
                                "| 36006f27-69d5-4fba-95d5-a09ad6e3d2b6 | test-3 | ACTIVE   |",
                                "| 5ca62725-a636-4f3b-bb5e-7b2c661006b6 | test-1 | ACTIVE   |",
                                "| 73a9cbe2-a4c7-4bb9-be1d-0f4606a87824 | test   | ACTIVE   |",
                                "| e237ead4-6261-44c0-95f1-03bd2fd83f6a | test-2 | ACTIVE   |",
                                "+--------------------------------------+--------+----------+",
                                "+ server_ping",
                                "++ openstack --os-cloud test floating ip list --status ACTIVE -f value -c 'Floating IP Address'",
                                "++ tr -d '\\r'",
                                "+ for address in $(openstack --os-cloud test floating ip list --status ACTIVE -f value -c \"Floating IP Address\" | tr -d '\\r')",
                                "+ ping -c3 192.168.112.193",
                                "PING 192.168.112.193 (192.168.112.193) 56(84) bytes of data.",
                                "64 bytes from 192.168.112.193: icmp_seq=1 ttl=63 time=6.05 ms",
                                "64 bytes from 192.168.112.193: icmp_seq=2 ttl=63 time=2.51 ms",
                                "64 bytes from 192.168.112.193: icmp_seq=3 ttl=63 time=1.73 ms",
                                "",
                                "--- 192.168.112.193 ping statistics ---",
                                "3 packets transmitted, 3 received, 0% packet loss, time 2003ms",
                                "rtt min/avg/max/mdev = 1.732/3.428/6.046/1.877 ms",
                                "+ for address in $(openstack --os-cloud test floating ip list --status ACTIVE -f value -c \"Floating IP Address\" | tr -d '\\r')",
                                "+ ping -c3 192.168.112.184",
                                "PING 192.168.112.184 (192.168.112.184) 56(84) bytes of data.",
                                "64 bytes from 192.168.112.184: icmp_seq=1 ttl=63 time=6.48 ms",
                                "64 bytes from 192.168.112.184: icmp_seq=2 ttl=63 time=2.72 ms",
                                "64 bytes from 192.168.112.184: icmp_seq=3 ttl=63 time=1.97 ms",
                                "",
                                "--- 192.168.112.184 ping statistics ---",
                                "3 packets transmitted, 3 received, 0% packet loss, time 2004ms",
                                "rtt min/avg/max/mdev = 1.972/3.721/6.478/1.972 ms",
                                "+ for address in $(openstack --os-cloud test floating ip list --status ACTIVE -f value -c \"Floating IP Address\" | tr -d '\\r')",
                                "+ ping -c3 192.168.112.182",
                                "PING 192.168.112.182 (192.168.112.182) 56(84) bytes of data.",
                                "64 bytes from 192.168.112.182: icmp_seq=1 ttl=63 time=5.39 ms",
                                "64 bytes from 192.168.112.182: icmp_seq=2 ttl=63 time=2.04 ms",
                                "64 bytes from 192.168.112.182: icmp_seq=3 ttl=63 time=1.95 ms",
                                "",
                                "--- 192.168.112.182 ping statistics ---",
                                "3 packets transmitted, 3 received, 0% packet loss, time 2003ms",
                                "rtt min/avg/max/mdev = 1.953/3.127/5.390/1.600 ms",
                                "+ for address in $(openstack --os-cloud test floating ip list --status ACTIVE -f value -c \"Floating IP Address\" | tr -d '\\r')",
                                "+ ping -c3 192.168.112.141",
                                "PING 192.168.112.141 (192.168.112.141) 56(84) bytes of data.",
                                "64 bytes from 192.168.112.141: icmp_seq=1 ttl=63 time=7.32 ms",
                                "64 bytes from 192.168.112.141: icmp_seq=2 ttl=63 time=3.61 ms",
                                "64 bytes from 192.168.112.141: icmp_seq=3 ttl=63 time=2.57 ms",
                                "",
                                "--- 192.168.112.141 ping statistics ---",
                                "3 packets transmitted, 3 received, 0% packet loss, time 2003ms",
                                "rtt min/avg/max/mdev = 2.573/4.503/7.324/2.039 ms",
                                "+ for address in $(openstack --os-cloud test floating ip list --status ACTIVE -f value -c \"Floating IP Address\" | tr -d '\\r')",
                                "+ ping -c3 192.168.112.192",
                                "PING 192.168.112.192 (192.168.112.192) 56(84) bytes of data.",
                                "64 bytes from 192.168.112.192: icmp_seq=1 ttl=63 time=8.24 ms",
                                "64 bytes from 192.168.112.192: icmp_seq=2 ttl=63 time=3.07 ms",
                                "64 bytes from 192.168.112.192: icmp_seq=3 ttl=63 time=1.99 ms",
                                "",
                                "--- 192.168.112.192 ping statistics ---",
                                "3 packets transmitted, 3 received, 0% packet loss, time 2003ms",
                                "rtt min/avg/max/mdev = 1.985/4.433/8.243/2.730 ms"
                            ],
                            "zuul_log_id": "0242ac1f-6606-1b2b-a63a-00000000001e-1-orchestrator"
                        }
                    },
                    "task": {
                        "duration": {
                            "end": "2026-02-24T01:39:34.337571Z",
                            "start": "2026-02-24T01:22:14.082437Z"
                        },
                        "id": "0242ac1f-6606-1b2b-a63a-00000000001e",
                        "name": "Run checks"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "ansible.builtin.command",
                            "changed": true,
                            "cmd": [
                                "ssh",
                                "-i",
                                "/home/zuul-testbed02/src/github.com/osism/testbed/terraform/.id_rsa.ci",
                                "dragon@81.163.192.195",
                                "/opt/configuration/scripts/check/302-openstack-with-tempest.sh"
                            ],
                            "delta": "0:01:59.826585",
                            "end": "2026-02-24 01:41:34.527931",
                            "invocation": {
                                "module_args": {
                                    "_raw_params": "ssh -i /home/zuul-testbed02/src/github.com/osism/testbed/terraform/.id_rsa.ci dragon@81.163.192.195 /opt/configuration/scripts/check/302-openstack-with-tempest.sh",
                                    "_uses_shell": false,
                                    "argv": null,
                                    "chdir": null,
                                    "creates": null,
                                    "executable": null,
                                    "expand_argument_vars": true,
                                    "removes": null,
                                    "stdin": null,
                                    "stdin_add_newline": true,
                                    "strip_empty_ends": true,
                                    "zuul_ansible_split_streams": false,
                                    "zuul_log_id": "0242ac1f-6606-1b2b-a63a-00000000001f-1-orchestrator",
                                    "zuul_output_max_bytes": 1073741824
                                }
                            },
                            "msg": "",
                            "rc": 0,
                            "start": "2026-02-24 01:39:34.701346",
                            "stderr": "",
                            "stderr_lines": [],
                            "stdout": "+ set -e\n+ source /opt/manager-vars.sh\n++ export NUMBER_OF_NODES=6\n++ NUMBER_OF_NODES=6\n++ export CEPH_VERSION=reef\n++ CEPH_VERSION=reef\n++ export CONFIGURATION_VERSION=main\n++ CONFIGURATION_VERSION=main\n++ export MANAGER_VERSION=latest\n++ MANAGER_VERSION=latest\n++ export OPENSTACK_VERSION=2024.2\n++ OPENSTACK_VERSION=2024.2\n++ export ARA=false\n++ ARA=false\n++ export DEPLOY_MODE=manager\n++ DEPLOY_MODE=manager\n++ export TEMPEST=true\n++ TEMPEST=true\n++ export IS_ZUUL=true\n++ IS_ZUUL=true\n++ export MANAGER_PUBLIC_IP_ADDRESS=81.163.192.195\n++ MANAGER_PUBLIC_IP_ADDRESS=81.163.192.195\n++ export EXTERNAL_API=false\n++ EXTERNAL_API=false\n++ export IMAGE_USER=ubuntu\n++ IMAGE_USER=ubuntu\n++ export IMAGE_NODE_USER=ubuntu\n\n# Tempest\n\n++ IMAGE_NODE_USER=ubuntu\n++ export CEPH_STACK=ceph-ansible\n++ CEPH_STACK=ceph-ansible\n+ echo\n+ echo '# Tempest'\n+ echo\n+ [[ ! -e /opt/tempest ]]\n+ osism apply tempest --skip-tags run-tempest\n\u001b[32m2026-02-24 01:39:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mPrepare task for execution of tempest.\u001b[0m\n\u001b[32m2026-02-24 01:39:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4b5266cf-3125-4217-841b-bd12a113c414 (tempest) was prepared for execution.\u001b[0m\n\u001b[32m2026-02-24 01:39:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task 4b5266cf-3125-4217-841b-bd12a113c414 (tempest) has been started and output is visible here.\u001b[0m\n\nPLAY [Run tempest] *************************************************************\n\nTASK [osism.validations.tempest : Create tempest workdir] **********************\nTuesday 24 February 2026  01:39:52 +0000 (0:00:00.249)       0:00:00.250 ****** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nTASK [osism.validations.tempest : Copy tempest wrapper script] *****************\nTuesday 24 February 2026  01:39:52 +0000 (0:00:00.740)       0:00:00.990 ****** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nTASK [osism.validations.tempest : Check for existing tempest initialisation] ***\nTuesday 24 February 2026  01:39:54 +0000 (0:00:01.278)       0:00:02.268 ****** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\nTASK [osism.validations.tempest : Init tempest] ********************************\nTuesday 24 February 2026  01:39:54 +0000 (0:00:00.413)       0:00:02.682 ****** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nTASK [osism.validations.tempest : Resolve image IDs] ***************************\nTuesday 24 February 2026  01:40:18 +0000 (0:00:24.048)       0:00:26.730 ****** \n\u001b[0;32mok: [testbed-manager -> localhost] => (item=Cirros 0.6.3)\u001b[0m\n\u001b[0;32mok: [testbed-manager -> localhost] => (item=Cirros 0.6.2)\u001b[0m\n\nTASK [osism.validations.tempest : Assert images have been resolved] ************\nTuesday 24 February 2026  01:40:26 +0000 (0:00:08.392)       0:00:35.123 ****** \n\u001b[0;32mok: [testbed-manager] => {\u001b[0m\n\u001b[0;32m    \"changed\": false,\u001b[0m\n\u001b[0;32m    \"msg\": \"All assertions passed\"\u001b[0m\n\u001b[0;32m}\u001b[0m\n\nTASK [osism.validations.tempest : Get auth token] ******************************\nTuesday 24 February 2026  01:40:27 +0000 (0:00:00.147)       0:00:35.270 ****** \n\u001b[0;32mok: [testbed-manager -> localhost]\u001b[0m\n\nTASK [osism.validations.tempest : Get endpoint catalog] ************************\nTuesday 24 February 2026  01:40:30 +0000 (0:00:03.326)       0:00:38.596 ****** \n\u001b[0;32mok: [testbed-manager -> localhost]\u001b[0m\n\nTASK [osism.validations.tempest : Get service catalog] *************************\nTuesday 24 February 2026  01:40:31 +0000 (0:00:01.572)       0:00:40.169 ****** \n\u001b[0;32mok: [testbed-manager -> localhost]\u001b[0m\n\nTASK [osism.validations.tempest : Register img_file name] **********************\nTuesday 24 February 2026  01:40:35 +0000 (0:00:03.439)       0:00:43.609 ****** \n\u001b[0;32mok: [testbed-manager -> localhost]\u001b[0m\n\nTASK [osism.validations.tempest : Download img_file from image_ref] ************\nTuesday 24 February 2026  01:40:35 +0000 (0:00:00.202)       0:00:43.811 ****** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nTASK [osism.validations.tempest : Install qemu-utils package] ******************\nTuesday 24 February 2026  01:40:38 +0000 (0:00:02.966)       0:00:46.778 ****** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nTASK [osism.validations.tempest : Convert img_file to qcow2 format] ************\nTuesday 24 February 2026  01:40:48 +0000 (0:00:09.477)       0:00:56.255 ****** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nTASK [osism.validations.tempest : Get network API extensions] ******************\nTuesday 24 February 2026  01:40:48 +0000 (0:00:00.761)       0:00:57.017 ****** \n\u001b[0;32mok: [testbed-manager -> localhost]\u001b[0m\n\nTASK [osism.validations.tempest : Revoke token] ********************************\nTuesday 24 February 2026  01:40:50 +0000 (0:00:01.417)       0:00:58.434 ****** \n\u001b[0;32mok: [testbed-manager -> localhost]\u001b[0m\n\nTASK [osism.validations.tempest : Set fact for config option api_extensions] ***\nTuesday 24 February 2026  01:40:51 +0000 (0:00:01.443)       0:00:59.878 ****** \n\u001b[0;32mok: [testbed-manager -> localhost]\u001b[0m\n\nTASK [osism.validations.tempest : Set fact for config option img_file] *********\nTuesday 24 February 2026  01:40:51 +0000 (0:00:00.197)       0:01:00.075 ****** \n\u001b[0;32mok: [testbed-manager -> localhost]\u001b[0m\n\nTASK [osism.validations.tempest : Resolve floating network ID] *****************\nTuesday 24 February 2026  01:40:52 +0000 (0:00:00.168)       0:01:00.244 ****** \n\u001b[0;32mok: [testbed-manager -> localhost]\u001b[0m\n\nTASK [osism.validations.tempest : Assert floating network id has been resolved] ***\nTuesday 24 February 2026  01:40:55 +0000 (0:00:03.545)       0:01:03.789 ****** \n\u001b[0;32mok: [testbed-manager -> localhost] => {\u001b[0m\n\u001b[0;32m    \"changed\": false,\u001b[0m\n\u001b[0;32m    \"msg\": \"All assertions passed\"\u001b[0m\n\u001b[0;32m}\u001b[0m\n\nTASK [osism.validations.tempest : Resolve flavor IDs] **************************\nTuesday 24 February 2026  01:40:55 +0000 (0:00:00.184)       0:01:03.974 ****** \n\u001b[0;36mskipping: [testbed-manager] => (item={'name': 'tempest-1', 'vcpus': 1, 'ram': 1024, 'disk': 1}) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager] => (item={'name': 'tempest-2', 'vcpus': 2, 'ram': 2048, 'disk': 2}) \u001b[0m\n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\nTASK [osism.validations.tempest : Assert flavors have been resolved] ***********\nTuesday 24 February 2026  01:40:56 +0000 (0:00:00.325)       0:01:04.300 ****** \n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\nTASK [osism.validations.tempest : Get stats of exclude list] *******************\nTuesday 24 February 2026  01:40:56 +0000 (0:00:00.153)       0:01:04.453 ****** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\nTASK [osism.validations.tempest : Copy exclude list] ***************************\nTuesday 24 February 2026  01:40:56 +0000 (0:00:00.470)       0:01:04.923 ****** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nTASK [osism.validations.tempest : Get stats of include list] *******************\nTuesday 24 February 2026  01:40:57 +0000 (0:00:00.861)       0:01:05.784 ****** \n\u001b[0;32mok: [testbed-manager]\u001b[0m\n\nTASK [osism.validations.tempest : Copy include list] ***************************\nTuesday 24 February 2026  01:40:57 +0000 (0:00:00.391)       0:01:06.176 ****** \n\u001b[0;36mskipping: [testbed-manager]\u001b[0m\n\nTASK [osism.validations.tempest : Create tempest flavors] **********************\nTuesday 24 February 2026  01:40:58 +0000 (0:00:00.142)       0:01:06.318 ****** \n\u001b[0;33mchanged: [testbed-manager -> localhost] => (item={'name': 'tempest-1', 'vcpus': 1, 'ram': 1024, 'disk': 1})\u001b[0m\n\u001b[0;33mchanged: [testbed-manager -> localhost] => (item={'name': 'tempest-2', 'vcpus': 2, 'ram': 2048, 'disk': 2})\u001b[0m\n\nTASK [osism.validations.tempest : Copy tempest.conf file] **********************\nTuesday 24 February 2026  01:41:05 +0000 (0:00:07.331)       0:01:13.650 ****** \n\u001b[0;33mchanged: [testbed-manager]\u001b[0m\n\nPLAY RECAP *********************************************************************\n\u001b[0;33mtestbed-manager\u001b[0m            : \u001b[0;32mok=24  \u001b[0m \u001b[0;33mchanged=9   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=3   \u001b[0m rescued=0    ignored=0   \n\n\nTASKS RECAP ********************************************************************\nTuesday 24 February 2026  01:41:06 +0000 (0:00:00.903)       0:01:14.553 ****** \n=============================================================================== \nosism.validations.tempest : Init tempest ------------------------------- 24.05s\nosism.validations.tempest : Install qemu-utils package ------------------ 9.48s\nosism.validations.tempest : Resolve image IDs --------------------------- 8.39s\nosism.validations.tempest : Create tempest flavors ---------------------- 7.33s\nosism.validations.tempest : Resolve floating network ID ----------------- 3.55s\nosism.validations.tempest : Get service catalog ------------------------- 3.44s\nosism.validations.tempest : Get auth token ------------------------------ 3.33s\nosism.validations.tempest : Download img_file from image_ref ------------ 2.97s\nosism.validations.tempest : Get endpoint catalog ------------------------ 1.57s\nosism.validations.tempest : Revoke token -------------------------------- 1.44s\nosism.validations.tempest : Get network API extensions ------------------ 1.42s\nosism.validations.tempest : Copy tempest wrapper script ----------------- 1.28s\nosism.validations.tempest : Copy tempest.conf file ---------------------- 0.90s\nosism.validations.tempest : Copy exclude list --------------------------- 0.86s\nosism.validations.tempest : Convert img_file to qcow2 format ------------ 0.76s\nosism.validations.tempest : Create tempest workdir ---------------------- 0.74s\nosism.validations.tempest : Get stats of exclude list ------------------- 0.47s\nosism.validations.tempest : Check for existing tempest initialisation --- 0.41s\nosism.validations.tempest : Get stats of include list ------------------- 0.39s\nosism.validations.tempest : Resolve flavor IDs -------------------------- 0.33s\n+ sed -i '/log_dir =/d' /opt/tempest/etc/tempest.conf\n+ sed -i '/log_file =/d' /opt/tempest/etc/tempest.conf\n\n## IDENTITY (API)\n\n+ [[ false == \\t\\r\\u\\e ]]\n+ echo\n+ echo '## IDENTITY (API)'\n+ echo\n+ _tempest tempest.api.identity.v3\n+ local regex=tempest.api.identity.v3\n+ docker run --rm -v /opt/tempest:/tempest -v /etc/ssl/certs:/etc/ssl/certs:ro -e PYTHONWARNINGS=ignore::SyntaxWarning --network host --name tempest registry.osism.tech/osism/tempest:latest run --workspace-path /tempest/workspace.yaml --workspace tempest --exclude-list /tempest/exclude.lst --regex tempest.api.identity.v3 --concurrency 16\n++ date +%Y%m%d-%H%M\n+ tee -a /opt/tempest/20260224-0141.log\ntempest: 'run --workspace-path /tempest/workspace.yaml --workspace tempest --exclude-list /tempest/exclude.lst --regex tempest.api.identity.v3 --concurrency 16' is not a tempest command. See 'tempest --help'.\nDid you mean one of these?\n  help\n  init\n\n## IMAGE (API)\n\n+ echo\n+ echo '## IMAGE (API)'\n+ echo\n+ _tempest tempest.api.image.v2\n+ local regex=tempest.api.image.v2\n+ docker run --rm -v /opt/tempest:/tempest -v /etc/ssl/certs:/etc/ssl/certs:ro -e PYTHONWARNINGS=ignore::SyntaxWarning --network host --name tempest registry.osism.tech/osism/tempest:latest run --workspace-path /tempest/workspace.yaml --workspace tempest --exclude-list /tempest/exclude.lst --regex tempest.api.image.v2 --concurrency 16\n++ date +%Y%m%d-%H%M\n+ tee -a /opt/tempest/20260224-0141.log\ntempest: 'run --workspace-path /tempest/workspace.yaml --workspace tempest --exclude-list /tempest/exclude.lst --regex tempest.api.image.v2 --concurrency 16' is not a tempest command. See 'tempest --help'.\nDid you mean one of these?\n  help\n  init\n\n## NETWORK (API)\n\n+ echo\n+ echo '## NETWORK (API)'\n+ echo\n+ _tempest tempest.api.network\n+ local regex=tempest.api.network\n+ docker run --rm -v /opt/tempest:/tempest -v /etc/ssl/certs:/etc/ssl/certs:ro -e PYTHONWARNINGS=ignore::SyntaxWarning --network host --name tempest registry.osism.tech/osism/tempest:latest run --workspace-path /tempest/workspace.yaml --workspace tempest --exclude-list /tempest/exclude.lst --regex tempest.api.network --concurrency 16\n++ date +%Y%m%d-%H%M\n+ tee -a /opt/tempest/20260224-0141.log\ntempest: 'run --workspace-path /tempest/workspace.yaml --workspace tempest --exclude-list /tempest/exclude.lst --regex tempest.api.network --concurrency 16' is not a tempest command. See 'tempest --help'.\nDid you mean one of these?\n  help\n  init\n\n## VOLUME (API)\n\n+ echo\n+ echo '## VOLUME (API)'\n+ echo\n+ _tempest tempest.api.volume\n+ local regex=tempest.api.volume\n+ docker run --rm -v /opt/tempest:/tempest -v /etc/ssl/certs:/etc/ssl/certs:ro -e PYTHONWARNINGS=ignore::SyntaxWarning --network host --name tempest registry.osism.tech/osism/tempest:latest run --workspace-path /tempest/workspace.yaml --workspace tempest --exclude-list /tempest/exclude.lst --regex tempest.api.volume --concurrency 16\n++ date +%Y%m%d-%H%M\n+ tee -a /opt/tempest/20260224-0141.log\ntempest: 'run --workspace-path /tempest/workspace.yaml --workspace tempest --exclude-list /tempest/exclude.lst --regex tempest.api.volume --concurrency 16' is not a tempest command. See 'tempest --help'.\nDid you mean one of these?\n  help\n  init\n\n## COMPUTE (API)\n\n+ echo\n+ echo '## COMPUTE (API)'\n+ echo\n+ _tempest tempest.api.compute\n+ local regex=tempest.api.compute\n+ docker run --rm -v /opt/tempest:/tempest -v /etc/ssl/certs:/etc/ssl/certs:ro -e PYTHONWARNINGS=ignore::SyntaxWarning --network host --name tempest registry.osism.tech/osism/tempest:latest run --workspace-path /tempest/workspace.yaml --workspace tempest --exclude-list /tempest/exclude.lst --regex tempest.api.compute --concurrency 16\n++ date +%Y%m%d-%H%M\n+ tee -a /opt/tempest/20260224-0141.log\ntempest: 'run --workspace-path /tempest/workspace.yaml --workspace tempest --exclude-list /tempest/exclude.lst --regex tempest.api.compute --concurrency 16' is not a tempest command. See 'tempest --help'.\nDid you mean one of these?\n  help\n  init\n\n## DNS (API)\n+ echo\n+ echo '## DNS (API)'\n+ echo\n\n+ _tempest designate_tempest_plugin.tests.api.v2\n+ local regex=designate_tempest_plugin.tests.api.v2\n+ docker run --rm -v /opt/tempest:/tempest -v /etc/ssl/certs:/etc/ssl/certs:ro -e PYTHONWARNINGS=ignore::SyntaxWarning --network host --name tempest registry.osism.tech/osism/tempest:latest run --workspace-path /tempest/workspace.yaml --workspace tempest --exclude-list /tempest/exclude.lst --regex designate_tempest_plugin.tests.api.v2 --concurrency 16\n++ date +%Y%m%d-%H%M\n+ tee -a /opt/tempest/20260224-0141.log\ntempest: 'run --workspace-path /tempest/workspace.yaml --workspace tempest --exclude-list /tempest/exclude.lst --regex designate_tempest_plugin.tests.api.v2 --concurrency 16' is not a tempest command. See 'tempest --help'.\nDid you mean one of these?\n  help\n  init\n\n## OBJECT-STORE (API)\n\n+ echo\n+ echo '## OBJECT-STORE (API)'\n+ echo\n+ _tempest tempest.api.object_storage\n+ local regex=tempest.api.object_storage\n+ docker run --rm -v /opt/tempest:/tempest -v /etc/ssl/certs:/etc/ssl/certs:ro -e PYTHONWARNINGS=ignore::SyntaxWarning --network host --name tempest registry.osism.tech/osism/tempest:latest run --workspace-path /tempest/workspace.yaml --workspace tempest --exclude-list /tempest/exclude.lst --regex tempest.api.object_storage --concurrency 16\n++ date +%Y%m%d-%H%M\n+ tee -a /opt/tempest/20260224-0141.log\ntempest: 'run --workspace-path /tempest/workspace.yaml --workspace tempest --exclude-list /tempest/exclude.lst --regex tempest.api.object_storage --concurrency 16' is not a tempest command. See 'tempest --help'.\nDid you mean one of these?\n  help\n  init",
                            "stdout_lines": [
                                "+ set -e",
                                "+ source /opt/manager-vars.sh",
                                "++ export NUMBER_OF_NODES=6",
                                "++ NUMBER_OF_NODES=6",
                                "++ export CEPH_VERSION=reef",
                                "++ CEPH_VERSION=reef",
                                "++ export CONFIGURATION_VERSION=main",
                                "++ CONFIGURATION_VERSION=main",
                                "++ export MANAGER_VERSION=latest",
                                "++ MANAGER_VERSION=latest",
                                "++ export OPENSTACK_VERSION=2024.2",
                                "++ OPENSTACK_VERSION=2024.2",
                                "++ export ARA=false",
                                "++ ARA=false",
                                "++ export DEPLOY_MODE=manager",
                                "++ DEPLOY_MODE=manager",
                                "++ export TEMPEST=true",
                                "++ TEMPEST=true",
                                "++ export IS_ZUUL=true",
                                "++ IS_ZUUL=true",
                                "++ export MANAGER_PUBLIC_IP_ADDRESS=81.163.192.195",
                                "++ MANAGER_PUBLIC_IP_ADDRESS=81.163.192.195",
                                "++ export EXTERNAL_API=false",
                                "++ EXTERNAL_API=false",
                                "++ export IMAGE_USER=ubuntu",
                                "++ IMAGE_USER=ubuntu",
                                "++ export IMAGE_NODE_USER=ubuntu",
                                "",
                                "# Tempest",
                                "",
                                "++ IMAGE_NODE_USER=ubuntu",
                                "++ export CEPH_STACK=ceph-ansible",
                                "++ CEPH_STACK=ceph-ansible",
                                "+ echo",
                                "+ echo '# Tempest'",
                                "+ echo",
                                "+ [[ ! -e /opt/tempest ]]",
                                "+ osism apply tempest --skip-tags run-tempest",
                                "\u001b[32m2026-02-24 01:39:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mPrepare task for execution of tempest.\u001b[0m",
                                "\u001b[32m2026-02-24 01:39:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mTask 4b5266cf-3125-4217-841b-bd12a113c414 (tempest) was prepared for execution.\u001b[0m",
                                "\u001b[32m2026-02-24 01:39:47\u001b[0m | \u001b[1mINFO    \u001b[0m | \u001b[1mIt takes a moment until task 4b5266cf-3125-4217-841b-bd12a113c414 (tempest) has been started and output is visible here.\u001b[0m",
                                "",
                                "PLAY [Run tempest] *************************************************************",
                                "",
                                "TASK [osism.validations.tempest : Create tempest workdir] **********************",
                                "Tuesday 24 February 2026  01:39:52 +0000 (0:00:00.249)       0:00:00.250 ****** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.validations.tempest : Copy tempest wrapper script] *****************",
                                "Tuesday 24 February 2026  01:39:52 +0000 (0:00:00.740)       0:00:00.990 ****** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.validations.tempest : Check for existing tempest initialisation] ***",
                                "Tuesday 24 February 2026  01:39:54 +0000 (0:00:01.278)       0:00:02.268 ****** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.validations.tempest : Init tempest] ********************************",
                                "Tuesday 24 February 2026  01:39:54 +0000 (0:00:00.413)       0:00:02.682 ****** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.validations.tempest : Resolve image IDs] ***************************",
                                "Tuesday 24 February 2026  01:40:18 +0000 (0:00:24.048)       0:00:26.730 ****** ",
                                "\u001b[0;32mok: [testbed-manager -> localhost] => (item=Cirros 0.6.3)\u001b[0m",
                                "\u001b[0;32mok: [testbed-manager -> localhost] => (item=Cirros 0.6.2)\u001b[0m",
                                "",
                                "TASK [osism.validations.tempest : Assert images have been resolved] ************",
                                "Tuesday 24 February 2026  01:40:26 +0000 (0:00:08.392)       0:00:35.123 ****** ",
                                "\u001b[0;32mok: [testbed-manager] => {\u001b[0m",
                                "\u001b[0;32m    \"changed\": false,\u001b[0m",
                                "\u001b[0;32m    \"msg\": \"All assertions passed\"\u001b[0m",
                                "\u001b[0;32m}\u001b[0m",
                                "",
                                "TASK [osism.validations.tempest : Get auth token] ******************************",
                                "Tuesday 24 February 2026  01:40:27 +0000 (0:00:00.147)       0:00:35.270 ****** ",
                                "\u001b[0;32mok: [testbed-manager -> localhost]\u001b[0m",
                                "",
                                "TASK [osism.validations.tempest : Get endpoint catalog] ************************",
                                "Tuesday 24 February 2026  01:40:30 +0000 (0:00:03.326)       0:00:38.596 ****** ",
                                "\u001b[0;32mok: [testbed-manager -> localhost]\u001b[0m",
                                "",
                                "TASK [osism.validations.tempest : Get service catalog] *************************",
                                "Tuesday 24 February 2026  01:40:31 +0000 (0:00:01.572)       0:00:40.169 ****** ",
                                "\u001b[0;32mok: [testbed-manager -> localhost]\u001b[0m",
                                "",
                                "TASK [osism.validations.tempest : Register img_file name] **********************",
                                "Tuesday 24 February 2026  01:40:35 +0000 (0:00:03.439)       0:00:43.609 ****** ",
                                "\u001b[0;32mok: [testbed-manager -> localhost]\u001b[0m",
                                "",
                                "TASK [osism.validations.tempest : Download img_file from image_ref] ************",
                                "Tuesday 24 February 2026  01:40:35 +0000 (0:00:00.202)       0:00:43.811 ****** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.validations.tempest : Install qemu-utils package] ******************",
                                "Tuesday 24 February 2026  01:40:38 +0000 (0:00:02.966)       0:00:46.778 ****** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.validations.tempest : Convert img_file to qcow2 format] ************",
                                "Tuesday 24 February 2026  01:40:48 +0000 (0:00:09.477)       0:00:56.255 ****** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.validations.tempest : Get network API extensions] ******************",
                                "Tuesday 24 February 2026  01:40:48 +0000 (0:00:00.761)       0:00:57.017 ****** ",
                                "\u001b[0;32mok: [testbed-manager -> localhost]\u001b[0m",
                                "",
                                "TASK [osism.validations.tempest : Revoke token] ********************************",
                                "Tuesday 24 February 2026  01:40:50 +0000 (0:00:01.417)       0:00:58.434 ****** ",
                                "\u001b[0;32mok: [testbed-manager -> localhost]\u001b[0m",
                                "",
                                "TASK [osism.validations.tempest : Set fact for config option api_extensions] ***",
                                "Tuesday 24 February 2026  01:40:51 +0000 (0:00:01.443)       0:00:59.878 ****** ",
                                "\u001b[0;32mok: [testbed-manager -> localhost]\u001b[0m",
                                "",
                                "TASK [osism.validations.tempest : Set fact for config option img_file] *********",
                                "Tuesday 24 February 2026  01:40:51 +0000 (0:00:00.197)       0:01:00.075 ****** ",
                                "\u001b[0;32mok: [testbed-manager -> localhost]\u001b[0m",
                                "",
                                "TASK [osism.validations.tempest : Resolve floating network ID] *****************",
                                "Tuesday 24 February 2026  01:40:52 +0000 (0:00:00.168)       0:01:00.244 ****** ",
                                "\u001b[0;32mok: [testbed-manager -> localhost]\u001b[0m",
                                "",
                                "TASK [osism.validations.tempest : Assert floating network id has been resolved] ***",
                                "Tuesday 24 February 2026  01:40:55 +0000 (0:00:03.545)       0:01:03.789 ****** ",
                                "\u001b[0;32mok: [testbed-manager -> localhost] => {\u001b[0m",
                                "\u001b[0;32m    \"changed\": false,\u001b[0m",
                                "\u001b[0;32m    \"msg\": \"All assertions passed\"\u001b[0m",
                                "\u001b[0;32m}\u001b[0m",
                                "",
                                "TASK [osism.validations.tempest : Resolve flavor IDs] **************************",
                                "Tuesday 24 February 2026  01:40:55 +0000 (0:00:00.184)       0:01:03.974 ****** ",
                                "\u001b[0;36mskipping: [testbed-manager] => (item={'name': 'tempest-1', 'vcpus': 1, 'ram': 1024, 'disk': 1}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager] => (item={'name': 'tempest-2', 'vcpus': 2, 'ram': 2048, 'disk': 2}) \u001b[0m",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.validations.tempest : Assert flavors have been resolved] ***********",
                                "Tuesday 24 February 2026  01:40:56 +0000 (0:00:00.325)       0:01:04.300 ****** ",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.validations.tempest : Get stats of exclude list] *******************",
                                "Tuesday 24 February 2026  01:40:56 +0000 (0:00:00.153)       0:01:04.453 ****** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.validations.tempest : Copy exclude list] ***************************",
                                "Tuesday 24 February 2026  01:40:56 +0000 (0:00:00.470)       0:01:04.923 ****** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.validations.tempest : Get stats of include list] *******************",
                                "Tuesday 24 February 2026  01:40:57 +0000 (0:00:00.861)       0:01:05.784 ****** ",
                                "\u001b[0;32mok: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.validations.tempest : Copy include list] ***************************",
                                "Tuesday 24 February 2026  01:40:57 +0000 (0:00:00.391)       0:01:06.176 ****** ",
                                "\u001b[0;36mskipping: [testbed-manager]\u001b[0m",
                                "",
                                "TASK [osism.validations.tempest : Create tempest flavors] **********************",
                                "Tuesday 24 February 2026  01:40:58 +0000 (0:00:00.142)       0:01:06.318 ****** ",
                                "\u001b[0;33mchanged: [testbed-manager -> localhost] => (item={'name': 'tempest-1', 'vcpus': 1, 'ram': 1024, 'disk': 1})\u001b[0m",
                                "\u001b[0;33mchanged: [testbed-manager -> localhost] => (item={'name': 'tempest-2', 'vcpus': 2, 'ram': 2048, 'disk': 2})\u001b[0m",
                                "",
                                "TASK [osism.validations.tempest : Copy tempest.conf file] **********************",
                                "Tuesday 24 February 2026  01:41:05 +0000 (0:00:07.331)       0:01:13.650 ****** ",
                                "\u001b[0;33mchanged: [testbed-manager]\u001b[0m",
                                "",
                                "PLAY RECAP *********************************************************************",
                                "\u001b[0;33mtestbed-manager\u001b[0m            : \u001b[0;32mok=24  \u001b[0m \u001b[0;33mchanged=9   \u001b[0m unreachable=0    failed=0    \u001b[0;36mskipped=3   \u001b[0m rescued=0    ignored=0   ",
                                "",
                                "",
                                "TASKS RECAP ********************************************************************",
                                "Tuesday 24 February 2026  01:41:06 +0000 (0:00:00.903)       0:01:14.553 ****** ",
                                "=============================================================================== ",
                                "osism.validations.tempest : Init tempest ------------------------------- 24.05s",
                                "osism.validations.tempest : Install qemu-utils package ------------------ 9.48s",
                                "osism.validations.tempest : Resolve image IDs --------------------------- 8.39s",
                                "osism.validations.tempest : Create tempest flavors ---------------------- 7.33s",
                                "osism.validations.tempest : Resolve floating network ID ----------------- 3.55s",
                                "osism.validations.tempest : Get service catalog ------------------------- 3.44s",
                                "osism.validations.tempest : Get auth token ------------------------------ 3.33s",
                                "osism.validations.tempest : Download img_file from image_ref ------------ 2.97s",
                                "osism.validations.tempest : Get endpoint catalog ------------------------ 1.57s",
                                "osism.validations.tempest : Revoke token -------------------------------- 1.44s",
                                "osism.validations.tempest : Get network API extensions ------------------ 1.42s",
                                "osism.validations.tempest : Copy tempest wrapper script ----------------- 1.28s",
                                "osism.validations.tempest : Copy tempest.conf file ---------------------- 0.90s",
                                "osism.validations.tempest : Copy exclude list --------------------------- 0.86s",
                                "osism.validations.tempest : Convert img_file to qcow2 format ------------ 0.76s",
                                "osism.validations.tempest : Create tempest workdir ---------------------- 0.74s",
                                "osism.validations.tempest : Get stats of exclude list ------------------- 0.47s",
                                "osism.validations.tempest : Check for existing tempest initialisation --- 0.41s",
                                "osism.validations.tempest : Get stats of include list ------------------- 0.39s",
                                "osism.validations.tempest : Resolve flavor IDs -------------------------- 0.33s",
                                "+ sed -i '/log_dir =/d' /opt/tempest/etc/tempest.conf",
                                "+ sed -i '/log_file =/d' /opt/tempest/etc/tempest.conf",
                                "",
                                "## IDENTITY (API)",
                                "",
                                "+ [[ false == \\t\\r\\u\\e ]]",
                                "+ echo",
                                "+ echo '## IDENTITY (API)'",
                                "+ echo",
                                "+ _tempest tempest.api.identity.v3",
                                "+ local regex=tempest.api.identity.v3",
                                "+ docker run --rm -v /opt/tempest:/tempest -v /etc/ssl/certs:/etc/ssl/certs:ro -e PYTHONWARNINGS=ignore::SyntaxWarning --network host --name tempest registry.osism.tech/osism/tempest:latest run --workspace-path /tempest/workspace.yaml --workspace tempest --exclude-list /tempest/exclude.lst --regex tempest.api.identity.v3 --concurrency 16",
                                "++ date +%Y%m%d-%H%M",
                                "+ tee -a /opt/tempest/20260224-0141.log",
                                "tempest: 'run --workspace-path /tempest/workspace.yaml --workspace tempest --exclude-list /tempest/exclude.lst --regex tempest.api.identity.v3 --concurrency 16' is not a tempest command. See 'tempest --help'.",
                                "Did you mean one of these?",
                                "  help",
                                "  init",
                                "",
                                "## IMAGE (API)",
                                "",
                                "+ echo",
                                "+ echo '## IMAGE (API)'",
                                "+ echo",
                                "+ _tempest tempest.api.image.v2",
                                "+ local regex=tempest.api.image.v2",
                                "+ docker run --rm -v /opt/tempest:/tempest -v /etc/ssl/certs:/etc/ssl/certs:ro -e PYTHONWARNINGS=ignore::SyntaxWarning --network host --name tempest registry.osism.tech/osism/tempest:latest run --workspace-path /tempest/workspace.yaml --workspace tempest --exclude-list /tempest/exclude.lst --regex tempest.api.image.v2 --concurrency 16",
                                "++ date +%Y%m%d-%H%M",
                                "+ tee -a /opt/tempest/20260224-0141.log",
                                "tempest: 'run --workspace-path /tempest/workspace.yaml --workspace tempest --exclude-list /tempest/exclude.lst --regex tempest.api.image.v2 --concurrency 16' is not a tempest command. See 'tempest --help'.",
                                "Did you mean one of these?",
                                "  help",
                                "  init",
                                "",
                                "## NETWORK (API)",
                                "",
                                "+ echo",
                                "+ echo '## NETWORK (API)'",
                                "+ echo",
                                "+ _tempest tempest.api.network",
                                "+ local regex=tempest.api.network",
                                "+ docker run --rm -v /opt/tempest:/tempest -v /etc/ssl/certs:/etc/ssl/certs:ro -e PYTHONWARNINGS=ignore::SyntaxWarning --network host --name tempest registry.osism.tech/osism/tempest:latest run --workspace-path /tempest/workspace.yaml --workspace tempest --exclude-list /tempest/exclude.lst --regex tempest.api.network --concurrency 16",
                                "++ date +%Y%m%d-%H%M",
                                "+ tee -a /opt/tempest/20260224-0141.log",
                                "tempest: 'run --workspace-path /tempest/workspace.yaml --workspace tempest --exclude-list /tempest/exclude.lst --regex tempest.api.network --concurrency 16' is not a tempest command. See 'tempest --help'.",
                                "Did you mean one of these?",
                                "  help",
                                "  init",
                                "",
                                "## VOLUME (API)",
                                "",
                                "+ echo",
                                "+ echo '## VOLUME (API)'",
                                "+ echo",
                                "+ _tempest tempest.api.volume",
                                "+ local regex=tempest.api.volume",
                                "+ docker run --rm -v /opt/tempest:/tempest -v /etc/ssl/certs:/etc/ssl/certs:ro -e PYTHONWARNINGS=ignore::SyntaxWarning --network host --name tempest registry.osism.tech/osism/tempest:latest run --workspace-path /tempest/workspace.yaml --workspace tempest --exclude-list /tempest/exclude.lst --regex tempest.api.volume --concurrency 16",
                                "++ date +%Y%m%d-%H%M",
                                "+ tee -a /opt/tempest/20260224-0141.log",
                                "tempest: 'run --workspace-path /tempest/workspace.yaml --workspace tempest --exclude-list /tempest/exclude.lst --regex tempest.api.volume --concurrency 16' is not a tempest command. See 'tempest --help'.",
                                "Did you mean one of these?",
                                "  help",
                                "  init",
                                "",
                                "## COMPUTE (API)",
                                "",
                                "+ echo",
                                "+ echo '## COMPUTE (API)'",
                                "+ echo",
                                "+ _tempest tempest.api.compute",
                                "+ local regex=tempest.api.compute",
                                "+ docker run --rm -v /opt/tempest:/tempest -v /etc/ssl/certs:/etc/ssl/certs:ro -e PYTHONWARNINGS=ignore::SyntaxWarning --network host --name tempest registry.osism.tech/osism/tempest:latest run --workspace-path /tempest/workspace.yaml --workspace tempest --exclude-list /tempest/exclude.lst --regex tempest.api.compute --concurrency 16",
                                "++ date +%Y%m%d-%H%M",
                                "+ tee -a /opt/tempest/20260224-0141.log",
                                "tempest: 'run --workspace-path /tempest/workspace.yaml --workspace tempest --exclude-list /tempest/exclude.lst --regex tempest.api.compute --concurrency 16' is not a tempest command. See 'tempest --help'.",
                                "Did you mean one of these?",
                                "  help",
                                "  init",
                                "",
                                "## DNS (API)",
                                "+ echo",
                                "+ echo '## DNS (API)'",
                                "+ echo",
                                "",
                                "+ _tempest designate_tempest_plugin.tests.api.v2",
                                "+ local regex=designate_tempest_plugin.tests.api.v2",
                                "+ docker run --rm -v /opt/tempest:/tempest -v /etc/ssl/certs:/etc/ssl/certs:ro -e PYTHONWARNINGS=ignore::SyntaxWarning --network host --name tempest registry.osism.tech/osism/tempest:latest run --workspace-path /tempest/workspace.yaml --workspace tempest --exclude-list /tempest/exclude.lst --regex designate_tempest_plugin.tests.api.v2 --concurrency 16",
                                "++ date +%Y%m%d-%H%M",
                                "+ tee -a /opt/tempest/20260224-0141.log",
                                "tempest: 'run --workspace-path /tempest/workspace.yaml --workspace tempest --exclude-list /tempest/exclude.lst --regex designate_tempest_plugin.tests.api.v2 --concurrency 16' is not a tempest command. See 'tempest --help'.",
                                "Did you mean one of these?",
                                "  help",
                                "  init",
                                "",
                                "## OBJECT-STORE (API)",
                                "",
                                "+ echo",
                                "+ echo '## OBJECT-STORE (API)'",
                                "+ echo",
                                "+ _tempest tempest.api.object_storage",
                                "+ local regex=tempest.api.object_storage",
                                "+ docker run --rm -v /opt/tempest:/tempest -v /etc/ssl/certs:/etc/ssl/certs:ro -e PYTHONWARNINGS=ignore::SyntaxWarning --network host --name tempest registry.osism.tech/osism/tempest:latest run --workspace-path /tempest/workspace.yaml --workspace tempest --exclude-list /tempest/exclude.lst --regex tempest.api.object_storage --concurrency 16",
                                "++ date +%Y%m%d-%H%M",
                                "+ tee -a /opt/tempest/20260224-0141.log",
                                "tempest: 'run --workspace-path /tempest/workspace.yaml --workspace tempest --exclude-list /tempest/exclude.lst --regex tempest.api.object_storage --concurrency 16' is not a tempest command. See 'tempest --help'.",
                                "Did you mean one of these?",
                                "  help",
                                "  init"
                            ],
                            "zuul_log_id": "0242ac1f-6606-1b2b-a63a-00000000001f-1-orchestrator"
                        }
                    },
                    "task": {
                        "duration": {
                            "end": "2026-02-24T01:41:35.043850Z",
                            "start": "2026-02-24T01:39:34.391809Z"
                        },
                        "id": "0242ac1f-6606-1b2b-a63a-00000000001f",
                        "name": "Run tempest"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "ansible.builtin.command",
                            "changed": false,
                            "false_condition": "_prometheus_alert_status | bool",
                            "skip_reason": "Conditional result was False",
                            "skipped": true
                        }
                    },
                    "task": {
                        "duration": {
                            "end": "2026-02-24T01:41:35.591931Z",
                            "start": "2026-02-24T01:41:35.080610Z"
                        },
                        "id": "0242ac1f-6606-1b2b-a63a-000000000020",
                        "name": "Check prometheus alert status"
                    }
                }
            ]
        }
    ],
    "stats": {
        "orchestrator": {
            "changed": 12,
            "failures": 0,
            "ignored": 0,
            "ok": 25,
            "rescued": 0,
            "skipped": 4,
            "unreachable": 0
        }
    },
    "trusted": false
},
{
    "branch": "main",
    "index": "0",
    "phase": "post",
    "playbook": "github.com/osism/testbed/playbooks/post.yml",
    "plays": [
        {
            "play": {
                "duration": {
                    "end": "2026-02-24T01:41:39.505718Z",
                    "start": "2026-02-24T01:41:36.602421Z"
                },
                "id": "0242ac1f-6606-91b5-2621-000000000002",
                "name": "Post output play"
            },
            "tasks": [
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "stat",
                            "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result",
                            "changed": false,
                            "skipped": true
                        }
                    },
                    "role": {
                        "id": "0242ac1f-6606-91b5-2621-000000000005",
                        "name": "stage-output",
                        "path": "/var/lib/zuul/builds/7573dac089bc49b5b355588f880c7711/untrusted/project_2/github.com/osism/openinfra-zuul-jobs/roles/stage-output"
                    },
                    "task": {
                        "duration": {
                            "end": "2026-02-24T01:41:36.697058Z",
                            "start": "2026-02-24T01:41:36.621027Z"
                        },
                        "id": "0242ac1f-6606-91b5-2621-000000000007",
                        "name": "Register sources"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "command",
                            "changed": true,
                            "cmd": [
                                "sudo",
                                "-n",
                                "true"
                            ],
                            "delta": "0:00:00.010748",
                            "end": "2026-02-24 01:41:37.636703",
                            "failed_when_result": false,
                            "invocation": {
                                "module_args": {
                                    "_raw_params": "sudo -n true",
                                    "_uses_shell": false,
                                    "argv": null,
                                    "chdir": null,
                                    "creates": null,
                                    "executable": null,
                                    "expand_argument_vars": true,
                                    "removes": null,
                                    "stdin": null,
                                    "stdin_add_newline": true,
                                    "strip_empty_ends": true,
                                    "zuul_ansible_split_streams": false,
                                    "zuul_log_id": "0242ac1f-6606-91b5-2621-000000000009-1-orchestrator",
                                    "zuul_output_max_bytes": 1073741824
                                }
                            },
                            "msg": "non-zero return code",
                            "rc": 1,
                            "start": "2026-02-24 01:41:37.625955",
                            "stderr": "",
                            "stderr_lines": [],
                            "stdout": "sudo: a password is required",
                            "stdout_lines": [
                                "sudo: a password is required"
                            ],
                            "zuul_log_id": "0242ac1f-6606-91b5-2621-000000000009-1-orchestrator"
                        }
                    },
                    "role": {
                        "id": "0242ac1f-6606-91b5-2621-000000000005",
                        "name": "stage-output",
                        "path": "/var/lib/zuul/builds/7573dac089bc49b5b355588f880c7711/untrusted/project_2/github.com/osism/openinfra-zuul-jobs/roles/stage-output"
                    },
                    "task": {
                        "duration": {
                            "end": "2026-02-24T01:41:37.752240Z",
                            "start": "2026-02-24T01:41:36.739200Z"
                        },
                        "id": "0242ac1f-6606-91b5-2621-000000000009",
                        "name": "Check sudo"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "set_fact",
                            "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result",
                            "changed": false,
                            "skipped": true
                        }
                    },
                    "role": {
                        "id": "0242ac1f-6606-91b5-2621-000000000005",
                        "name": "stage-output",
                        "path": "/var/lib/zuul/builds/7573dac089bc49b5b355588f880c7711/untrusted/project_2/github.com/osism/openinfra-zuul-jobs/roles/stage-output"
                    },
                    "task": {
                        "duration": {
                            "end": "2026-02-24T01:41:37.793765Z",
                            "start": "2026-02-24T01:41:37.767449Z"
                        },
                        "id": "0242ac1f-6606-91b5-2621-00000000000a",
                        "name": "Set source and destination for files and folders"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "set_fact",
                            "ansible_facts": {
                                "all_sources": []
                            },
                            "changed": false
                        }
                    },
                    "role": {
                        "id": "0242ac1f-6606-91b5-2621-000000000005",
                        "name": "stage-output",
                        "path": "/var/lib/zuul/builds/7573dac089bc49b5b355588f880c7711/untrusted/project_2/github.com/osism/openinfra-zuul-jobs/roles/stage-output"
                    },
                    "task": {
                        "duration": {
                            "end": "2026-02-24T01:41:37.880565Z",
                            "start": "2026-02-24T01:41:37.808494Z"
                        },
                        "id": "0242ac1f-6606-91b5-2621-00000000000c",
                        "name": "Build a list of source, dest dictionaries"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "file",
                            "changed": false,
                            "msg": "All items completed",
                            "results": [
                                {
                                    "ansible_loop_var": "zj_output_dirs",
                                    "changed": false,
                                    "diff": {
                                        "after": {
                                            "path": "/home/zuul-testbed02/zuul-output/docs"
                                        },
                                        "before": {
                                            "path": "/home/zuul-testbed02/zuul-output/docs"
                                        }
                                    },
                                    "failed": false,
                                    "gid": 1003,
                                    "group": "zuul-testbed02",
                                    "invocation": {
                                        "module_args": {
                                            "_diff_peek": null,
                                            "_original_basename": null,
                                            "access_time": null,
                                            "access_time_format": "%Y%m%d%H%M.%S",
                                            "attributes": null,
                                            "follow": true,
                                            "force": false,
                                            "group": null,
                                            "mode": 493,
                                            "modification_time": null,
                                            "modification_time_format": "%Y%m%d%H%M.%S",
                                            "owner": "zuul-testbed02",
                                            "path": "/home/zuul-testbed02/zuul-output/docs",
                                            "recurse": false,
                                            "selevel": null,
                                            "serole": null,
                                            "setype": null,
                                            "seuser": null,
                                            "src": null,
                                            "state": "directory",
                                            "unsafe_writes": false
                                        }
                                    },
                                    "mode": "0755",
                                    "owner": "zuul-testbed02",
                                    "path": "/home/zuul-testbed02/zuul-output/docs",
                                    "size": 4096,
                                    "state": "directory",
                                    "uid": 1002,
                                    "zj_output_dirs": "docs"
                                },
                                {
                                    "ansible_loop_var": "zj_output_dirs",
                                    "changed": false,
                                    "diff": {
                                        "after": {
                                            "path": "/home/zuul-testbed02/zuul-output/artifacts"
                                        },
                                        "before": {
                                            "path": "/home/zuul-testbed02/zuul-output/artifacts"
                                        }
                                    },
                                    "failed": false,
                                    "gid": 1003,
                                    "group": "zuul-testbed02",
                                    "invocation": {
                                        "module_args": {
                                            "_diff_peek": null,
                                            "_original_basename": null,
                                            "access_time": null,
                                            "access_time_format": "%Y%m%d%H%M.%S",
                                            "attributes": null,
                                            "follow": true,
                                            "force": false,
                                            "group": null,
                                            "mode": 493,
                                            "modification_time": null,
                                            "modification_time_format": "%Y%m%d%H%M.%S",
                                            "owner": "zuul-testbed02",
                                            "path": "/home/zuul-testbed02/zuul-output/artifacts",
                                            "recurse": false,
                                            "selevel": null,
                                            "serole": null,
                                            "setype": null,
                                            "seuser": null,
                                            "src": null,
                                            "state": "directory",
                                            "unsafe_writes": false
                                        }
                                    },
                                    "mode": "0755",
                                    "owner": "zuul-testbed02",
                                    "path": "/home/zuul-testbed02/zuul-output/artifacts",
                                    "size": 4096,
                                    "state": "directory",
                                    "uid": 1002,
                                    "zj_output_dirs": "artifacts"
                                },
                                {
                                    "ansible_loop_var": "zj_output_dirs",
                                    "changed": false,
                                    "diff": {
                                        "after": {
                                            "path": "/home/zuul-testbed02/zuul-output/logs"
                                        },
                                        "before": {
                                            "path": "/home/zuul-testbed02/zuul-output/logs"
                                        }
                                    },
                                    "failed": false,
                                    "gid": 1003,
                                    "group": "zuul-testbed02",
                                    "invocation": {
                                        "module_args": {
                                            "_diff_peek": null,
                                            "_original_basename": null,
                                            "access_time": null,
                                            "access_time_format": "%Y%m%d%H%M.%S",
                                            "attributes": null,
                                            "follow": true,
                                            "force": false,
                                            "group": null,
                                            "mode": 493,
                                            "modification_time": null,
                                            "modification_time_format": "%Y%m%d%H%M.%S",
                                            "owner": "zuul-testbed02",
                                            "path": "/home/zuul-testbed02/zuul-output/logs",
                                            "recurse": false,
                                            "selevel": null,
                                            "serole": null,
                                            "setype": null,
                                            "seuser": null,
                                            "src": null,
                                            "state": "directory",
                                            "unsafe_writes": false
                                        }
                                    },
                                    "mode": "0755",
                                    "owner": "zuul-testbed02",
                                    "path": "/home/zuul-testbed02/zuul-output/logs",
                                    "size": 4096,
                                    "state": "directory",
                                    "uid": 1002,
                                    "zj_output_dirs": "logs"
                                }
                            ]
                        }
                    },
                    "role": {
                        "id": "0242ac1f-6606-91b5-2621-000000000005",
                        "name": "stage-output",
                        "path": "/var/lib/zuul/builds/7573dac089bc49b5b355588f880c7711/untrusted/project_2/github.com/osism/openinfra-zuul-jobs/roles/stage-output"
                    },
                    "task": {
                        "duration": {
                            "end": "2026-02-24T01:41:39.005227Z",
                            "start": "2026-02-24T01:41:37.890436Z"
                        },
                        "id": "0242ac1f-6606-91b5-2621-00000000000d",
                        "name": "Ensure target folders exist"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "command",
                            "changed": false,
                            "results": [],
                            "skipped": true,
                            "skipped_reason": "No items in the list"
                        }
                    },
                    "role": {
                        "id": "0242ac1f-6606-91b5-2621-000000000005",
                        "name": "stage-output",
                        "path": "/var/lib/zuul/builds/7573dac089bc49b5b355588f880c7711/untrusted/project_2/github.com/osism/openinfra-zuul-jobs/roles/stage-output"
                    },
                    "task": {
                        "duration": {
                            "end": "2026-02-24T01:41:39.040495Z",
                            "start": "2026-02-24T01:41:39.015272Z"
                        },
                        "id": "0242ac1f-6606-91b5-2621-00000000000f",
                        "name": "Copy files and folders to staging folder"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "file",
                            "changed": false,
                            "diff": {
                                "after": {
                                    "path": "/home/zuul-testbed02/zuul-output/logs"
                                },
                                "before": {
                                    "path": "/home/zuul-testbed02/zuul-output/logs"
                                }
                            },
                            "gid": 1003,
                            "group": "zuul-testbed02",
                            "invocation": {
                                "module_args": {
                                    "_diff_peek": null,
                                    "_original_basename": null,
                                    "access_time": null,
                                    "access_time_format": "%Y%m%d%H%M.%S",
                                    "attributes": null,
                                    "dest": "/home/zuul-testbed02/zuul-output/logs",
                                    "follow": true,
                                    "force": false,
                                    "group": null,
                                    "mode": "u=rwX,g=rX,o=rX",
                                    "modification_time": null,
                                    "modification_time_format": "%Y%m%d%H%M.%S",
                                    "owner": null,
                                    "path": "/home/zuul-testbed02/zuul-output/logs",
                                    "recurse": true,
                                    "selevel": null,
                                    "serole": null,
                                    "setype": null,
                                    "seuser": null,
                                    "src": null,
                                    "state": "directory",
                                    "unsafe_writes": false
                                }
                            },
                            "mode": "0755",
                            "owner": "zuul-testbed02",
                            "path": "/home/zuul-testbed02/zuul-output/logs",
                            "size": 4096,
                            "state": "directory",
                            "uid": 1002
                        }
                    },
                    "role": {
                        "id": "0242ac1f-6606-91b5-2621-000000000005",
                        "name": "stage-output",
                        "path": "/var/lib/zuul/builds/7573dac089bc49b5b355588f880c7711/untrusted/project_2/github.com/osism/openinfra-zuul-jobs/roles/stage-output"
                    },
                    "task": {
                        "duration": {
                            "end": "2026-02-24T01:41:39.377773Z",
                            "start": "2026-02-24T01:41:39.055082Z"
                        },
                        "id": "0242ac1f-6606-91b5-2621-000000000011",
                        "name": "Make all log files readable"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "stage_output_renames",
                            "changed": false,
                            "false_condition": "extensions_to_txt is defined and extensions_to_txt is not none",
                            "skip_reason": "Conditional result was False",
                            "skipped": true
                        }
                    },
                    "role": {
                        "id": "0242ac1f-6606-91b5-2621-000000000005",
                        "name": "stage-output",
                        "path": "/var/lib/zuul/builds/7573dac089bc49b5b355588f880c7711/untrusted/project_2/github.com/osism/openinfra-zuul-jobs/roles/stage-output"
                    },
                    "task": {
                        "duration": {
                            "end": "2026-02-24T01:41:39.428822Z",
                            "start": "2026-02-24T01:41:39.392858Z"
                        },
                        "id": "0242ac1f-6606-91b5-2621-000000000012",
                        "name": "Rename log files that match extensions_to_txt"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "find",
                            "changed": false,
                            "false_condition": "stage_compress_logs",
                            "skip_reason": "Conditional result was False",
                            "skipped": true
                        }
                    },
                    "role": {
                        "id": "0242ac1f-6606-91b5-2621-000000000005",
                        "name": "stage-output",
                        "path": "/var/lib/zuul/builds/7573dac089bc49b5b355588f880c7711/untrusted/project_2/github.com/osism/openinfra-zuul-jobs/roles/stage-output"
                    },
                    "task": {
                        "duration": {
                            "end": "2026-02-24T01:41:39.469626Z",
                            "start": "2026-02-24T01:41:39.444344Z"
                        },
                        "id": "0242ac1f-6606-91b5-2621-000000000014",
                        "name": "Discover log files for compression"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "shell",
                            "changed": false,
                            "false_condition": "stage_compress_logs",
                            "skip_reason": "Conditional result was False",
                            "skipped": true
                        }
                    },
                    "role": {
                        "id": "0242ac1f-6606-91b5-2621-000000000005",
                        "name": "stage-output",
                        "path": "/var/lib/zuul/builds/7573dac089bc49b5b355588f880c7711/untrusted/project_2/github.com/osism/openinfra-zuul-jobs/roles/stage-output"
                    },
                    "task": {
                        "duration": {
                            "end": "2026-02-24T01:41:39.505718Z",
                            "start": "2026-02-24T01:41:39.481272Z"
                        },
                        "id": "0242ac1f-6606-91b5-2621-000000000015",
                        "name": "Archive everything from logs"
                    }
                }
            ]
        },
        {
            "play": {
                "duration": {
                    "end": "2026-02-24T01:42:00.720940Z",
                    "start": "2026-02-24T01:41:39.521699Z"
                },
                "id": "0242ac1f-6606-91b5-2621-000000000017",
                "name": "Post cleanup play"
            },
            "tasks": [
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "ansible.builtin.set_fact",
                            "ansible_facts": {
                                "cloud": "ci"
                            },
                            "changed": false
                        }
                    },
                    "task": {
                        "duration": {
                            "end": "2026-02-24T01:41:39.595519Z",
                            "start": "2026-02-24T01:41:39.535875Z"
                        },
                        "id": "0242ac1f-6606-91b5-2621-000000000019",
                        "name": "Set cloud fact (Zuul deployment)"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "ansible.builtin.set_fact",
                            "changed": false,
                            "false_condition": "'nodepool' not in hostvars[groups['all'][0]]",
                            "skip_reason": "Conditional result was False",
                            "skipped": true
                        }
                    },
                    "task": {
                        "duration": {
                            "end": "2026-02-24T01:41:39.642459Z",
                            "start": "2026-02-24T01:41:39.607077Z"
                        },
                        "id": "0242ac1f-6606-91b5-2621-00000000001a",
                        "name": "Set cloud fact (local deployment)"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "ansible.builtin.shell",
                            "changed": true,
                            "cmd": "OS_CLOUD=ci ~/venv/bin/python3 cleanup.py\n",
                            "delta": "0:00:20.434915",
                            "end": "2026-02-24 01:42:00.408681",
                            "failed_when_result": false,
                            "invocation": {
                                "module_args": {
                                    "_raw_params": "OS_CLOUD=ci ~/venv/bin/python3 cleanup.py\n",
                                    "_uses_shell": true,
                                    "argv": null,
                                    "chdir": "/home/zuul-testbed02/src/github.com/osism/testbed/terraform/scripts",
                                    "creates": null,
                                    "executable": null,
                                    "expand_argument_vars": true,
                                    "removes": null,
                                    "stdin": null,
                                    "stdin_add_newline": true,
                                    "strip_empty_ends": true,
                                    "zuul_ansible_split_streams": false,
                                    "zuul_log_id": "0242ac1f-6606-91b5-2621-00000000001b-1-orchestrator",
                                    "zuul_output_max_bytes": 1073741824
                                }
                            },
                            "msg": "",
                            "rc": 0,
                            "start": "2026-02-24 01:41:39.973766",
                            "stderr": "",
                            "stderr_lines": [],
                            "stdout": "2026-02-24 01:41:40 - clean up servers\n2026-02-24 01:41:41 - testbed-manager\n2026-02-24 01:41:41 - testbed-node-3\n2026-02-24 01:41:41 - testbed-node-1\n2026-02-24 01:41:41 - testbed-node-2\n2026-02-24 01:41:41 - testbed-node-0\n2026-02-24 01:41:41 - testbed-node-5\n2026-02-24 01:41:41 - testbed-node-4\n2026-02-24 01:41:41 - clean up keypairs\n2026-02-24 01:41:41 - testbed\n2026-02-24 01:41:41 - wait for servers to be gone\n2026-02-24 01:41:52 - clean up ports\n2026-02-24 01:41:52 - 3efcac55-5642-4046-a9a6-94f0970dacfb\n2026-02-24 01:41:53 - 4e0612ba-f5fc-449f-bdf0-52cf3f9e0ff4\n2026-02-24 01:41:53 - 52a531cd-5463-4412-a687-49e3322e1d3a\n2026-02-24 01:41:53 - 6419139a-cf37-4920-a5ce-6fbaf77a4267\n2026-02-24 01:41:54 - b86f1b60-f237-4c8b-a846-4b44402e937f\n2026-02-24 01:41:54 - b963b7df-f3b4-4efc-a2f3-fc1a7ac620e0\n2026-02-24 01:41:54 - fd0bab54-e5d7-4115-8ac6-6d5e24c663da\n2026-02-24 01:41:54 - clean up volumes\n2026-02-24 01:41:54 - testbed-volume-3-node-base\n2026-02-24 01:41:54 - testbed-volume-5-node-base\n2026-02-24 01:41:54 - testbed-volume-1-node-base\n2026-02-24 01:41:54 - testbed-volume-4-node-base\n2026-02-24 01:41:54 - testbed-volume-2-node-base\n2026-02-24 01:41:54 - testbed-volume-0-node-base\n2026-02-24 01:41:55 - testbed-volume-manager-base\n2026-02-24 01:41:55 - testbed-volume-5-node-5\n2026-02-24 01:41:55 - testbed-volume-4-node-4\n2026-02-24 01:41:55 - testbed-volume-7-node-4\n2026-02-24 01:41:55 - testbed-volume-0-node-3\n2026-02-24 01:41:55 - testbed-volume-2-node-5\n2026-02-24 01:41:55 - testbed-volume-3-node-3\n2026-02-24 01:41:55 - testbed-volume-8-node-5\n2026-02-24 01:41:55 - testbed-volume-1-node-4\n2026-02-24 01:41:55 - testbed-volume-6-node-3\n2026-02-24 01:41:55 - disconnect routers\n2026-02-24 01:41:55 - testbed\n2026-02-24 01:41:57 - clean up subnets\n2026-02-24 01:41:57 - subnet-testbed-management\n2026-02-24 01:41:57 - clean up networks\n2026-02-24 01:41:57 - net-testbed-management\n2026-02-24 01:41:57 - clean up security groups\n2026-02-24 01:41:57 - testbed-node\n2026-02-24 01:41:58 - testbed-management\n2026-02-24 01:41:58 - clean up floating ips\n2026-02-24 01:41:58 - 81.163.192.195\n2026-02-24 01:41:58 - clean up routers\n2026-02-24 01:41:58 - testbed",
                            "stdout_lines": [
                                "2026-02-24 01:41:40 - clean up servers",
                                "2026-02-24 01:41:41 - testbed-manager",
                                "2026-02-24 01:41:41 - testbed-node-3",
                                "2026-02-24 01:41:41 - testbed-node-1",
                                "2026-02-24 01:41:41 - testbed-node-2",
                                "2026-02-24 01:41:41 - testbed-node-0",
                                "2026-02-24 01:41:41 - testbed-node-5",
                                "2026-02-24 01:41:41 - testbed-node-4",
                                "2026-02-24 01:41:41 - clean up keypairs",
                                "2026-02-24 01:41:41 - testbed",
                                "2026-02-24 01:41:41 - wait for servers to be gone",
                                "2026-02-24 01:41:52 - clean up ports",
                                "2026-02-24 01:41:52 - 3efcac55-5642-4046-a9a6-94f0970dacfb",
                                "2026-02-24 01:41:53 - 4e0612ba-f5fc-449f-bdf0-52cf3f9e0ff4",
                                "2026-02-24 01:41:53 - 52a531cd-5463-4412-a687-49e3322e1d3a",
                                "2026-02-24 01:41:53 - 6419139a-cf37-4920-a5ce-6fbaf77a4267",
                                "2026-02-24 01:41:54 - b86f1b60-f237-4c8b-a846-4b44402e937f",
                                "2026-02-24 01:41:54 - b963b7df-f3b4-4efc-a2f3-fc1a7ac620e0",
                                "2026-02-24 01:41:54 - fd0bab54-e5d7-4115-8ac6-6d5e24c663da",
                                "2026-02-24 01:41:54 - clean up volumes",
                                "2026-02-24 01:41:54 - testbed-volume-3-node-base",
                                "2026-02-24 01:41:54 - testbed-volume-5-node-base",
                                "2026-02-24 01:41:54 - testbed-volume-1-node-base",
                                "2026-02-24 01:41:54 - testbed-volume-4-node-base",
                                "2026-02-24 01:41:54 - testbed-volume-2-node-base",
                                "2026-02-24 01:41:54 - testbed-volume-0-node-base",
                                "2026-02-24 01:41:55 - testbed-volume-manager-base",
                                "2026-02-24 01:41:55 - testbed-volume-5-node-5",
                                "2026-02-24 01:41:55 - testbed-volume-4-node-4",
                                "2026-02-24 01:41:55 - testbed-volume-7-node-4",
                                "2026-02-24 01:41:55 - testbed-volume-0-node-3",
                                "2026-02-24 01:41:55 - testbed-volume-2-node-5",
                                "2026-02-24 01:41:55 - testbed-volume-3-node-3",
                                "2026-02-24 01:41:55 - testbed-volume-8-node-5",
                                "2026-02-24 01:41:55 - testbed-volume-1-node-4",
                                "2026-02-24 01:41:55 - testbed-volume-6-node-3",
                                "2026-02-24 01:41:55 - disconnect routers",
                                "2026-02-24 01:41:55 - testbed",
                                "2026-02-24 01:41:57 - clean up subnets",
                                "2026-02-24 01:41:57 - subnet-testbed-management",
                                "2026-02-24 01:41:57 - clean up networks",
                                "2026-02-24 01:41:57 - net-testbed-management",
                                "2026-02-24 01:41:57 - clean up security groups",
                                "2026-02-24 01:41:57 - testbed-node",
                                "2026-02-24 01:41:58 - testbed-management",
                                "2026-02-24 01:41:58 - clean up floating ips",
                                "2026-02-24 01:41:58 - 81.163.192.195",
                                "2026-02-24 01:41:58 - clean up routers",
                                "2026-02-24 01:41:58 - testbed"
                            ],
                            "zuul_log_id": "0242ac1f-6606-91b5-2621-00000000001b-1-orchestrator"
                        }
                    },
                    "task": {
                        "duration": {
                            "end": "2026-02-24T01:42:00.720940Z",
                            "start": "2026-02-24T01:41:39.686610Z"
                        },
                        "id": "0242ac1f-6606-91b5-2621-00000000001b",
                        "name": "Clean the cloud environment"
                    }
                }
            ]
        }
    ],
    "stats": {
        "orchestrator": {
            "changed": 2,
            "failures": 0,
            "ignored": 0,
            "ok": 6,
            "rescued": 0,
            "skipped": 7,
            "unreachable": 0
        }
    },
    "trusted": false
},
{
    "branch": "main",
    "index": "1",
    "phase": "post",
    "playbook": "github.com/osism/testbed/playbooks/cleanup.yml",
    "plays": [
        {
            "play": {
                "duration": {
                    "end": "2026-02-24T01:42:04.305850Z",
                    "start": "2026-02-24T01:42:01.624690Z"
                },
                "id": "0242ac1f-6606-7974-633e-000000000002",
                "name": "Cleanup play"
            },
            "tasks": [
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "ansible.builtin.set_fact",
                            "ansible_facts": {
                                "cloud": "ci"
                            },
                            "changed": false
                        }
                    },
                    "task": {
                        "duration": {
                            "end": "2026-02-24T01:42:01.700893Z",
                            "start": "2026-02-24T01:42:01.641634Z"
                        },
                        "id": "0242ac1f-6606-7974-633e-000000000004",
                        "name": "Set cloud fact (Zuul deployment)"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "ansible.builtin.set_fact",
                            "changed": false,
                            "false_condition": "'nodepool' not in hostvars[groups['all'][0]]",
                            "skip_reason": "Conditional result was False",
                            "skipped": true
                        }
                    },
                    "task": {
                        "duration": {
                            "end": "2026-02-24T01:42:01.747076Z",
                            "start": "2026-02-24T01:42:01.711200Z"
                        },
                        "id": "0242ac1f-6606-7974-633e-000000000005",
                        "name": "Set cloud fact (local deployment)"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "ansible.builtin.shell",
                            "changed": true,
                            "cmd": "OS_CLOUD=ci ~/venv/bin/python3 cleanup.py\n",
                            "delta": "0:00:01.368925",
                            "end": "2026-02-24 01:42:03.993175",
                            "failed_when_result": false,
                            "invocation": {
                                "module_args": {
                                    "_raw_params": "OS_CLOUD=ci ~/venv/bin/python3 cleanup.py\n",
                                    "_uses_shell": true,
                                    "argv": null,
                                    "chdir": "/home/zuul-testbed02/src/github.com/osism/testbed/terraform/scripts",
                                    "creates": null,
                                    "executable": null,
                                    "expand_argument_vars": true,
                                    "removes": null,
                                    "stdin": null,
                                    "stdin_add_newline": true,
                                    "strip_empty_ends": true,
                                    "zuul_ansible_split_streams": false,
                                    "zuul_log_id": "0242ac1f-6606-7974-633e-000000000006-1-orchestrator",
                                    "zuul_output_max_bytes": 1073741824
                                }
                            },
                            "msg": "",
                            "rc": 0,
                            "start": "2026-02-24 01:42:02.624250",
                            "stderr": "",
                            "stderr_lines": [],
                            "stdout": "2026-02-24 01:42:02 - clean up servers\n2026-02-24 01:42:03 - clean up keypairs\n2026-02-24 01:42:03 - wait for servers to be gone\n2026-02-24 01:42:03 - clean up ports\n2026-02-24 01:42:03 - clean up volumes\n2026-02-24 01:42:03 - disconnect routers\n2026-02-24 01:42:03 - clean up subnets\n2026-02-24 01:42:03 - clean up networks\n2026-02-24 01:42:03 - clean up security groups\n2026-02-24 01:42:03 - clean up floating ips\n2026-02-24 01:42:03 - clean up routers",
                            "stdout_lines": [
                                "2026-02-24 01:42:02 - clean up servers",
                                "2026-02-24 01:42:03 - clean up keypairs",
                                "2026-02-24 01:42:03 - wait for servers to be gone",
                                "2026-02-24 01:42:03 - clean up ports",
                                "2026-02-24 01:42:03 - clean up volumes",
                                "2026-02-24 01:42:03 - disconnect routers",
                                "2026-02-24 01:42:03 - clean up subnets",
                                "2026-02-24 01:42:03 - clean up networks",
                                "2026-02-24 01:42:03 - clean up security groups",
                                "2026-02-24 01:42:03 - clean up floating ips",
                                "2026-02-24 01:42:03 - clean up routers"
                            ],
                            "zuul_log_id": "0242ac1f-6606-7974-633e-000000000006-1-orchestrator"
                        }
                    },
                    "task": {
                        "duration": {
                            "end": "2026-02-24T01:42:04.305850Z",
                            "start": "2026-02-24T01:42:01.789249Z"
                        },
                        "id": "0242ac1f-6606-7974-633e-000000000006",
                        "name": "Clean the cloud environment"
                    }
                }
            ]
        }
    ],
    "stats": {
        "orchestrator": {
            "changed": 1,
            "failures": 0,
            "ignored": 0,
            "ok": 2,
            "rescued": 0,
            "skipped": 1,
            "unreachable": 0
        }
    },
    "trusted": false
},
{
    "branch": "main",
    "index": "2",
    "phase": "post",
    "playbook": "github.com/osism/zuul-config/playbooks/base/post-fetch.yaml",
    "plays": [
        {
            "play": {
                "duration": {
                    "end": "2026-02-24T01:42:08.917898Z",
                    "start": "2026-02-24T01:42:05.265491Z"
                },
                "id": "0242ac1f-6606-530f-7072-000000000002",
                "name": "Base post-fetch"
            },
            "tasks": [
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "set_fact",
                            "changed": false,
                            "false_condition": "groups['all'] | length > 1",
                            "skip_reason": "Conditional result was False",
                            "skipped": true
                        }
                    },
                    "role": {
                        "id": "0242ac1f-6606-530f-7072-000000000006",
                        "name": "fetch-output",
                        "path": "/var/lib/zuul/builds/7573dac089bc49b5b355588f880c7711/trusted/project_1/github.com/osism/openinfra-zuul-jobs/roles/fetch-output"
                    },
                    "task": {
                        "duration": {
                            "end": "2026-02-24T01:42:05.336180Z",
                            "start": "2026-02-24T01:42:05.289864Z"
                        },
                        "id": "0242ac1f-6606-530f-7072-000000000008",
                        "name": "Set log path for multiple nodes"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "set_fact",
                            "ansible_facts": {
                                "log_path": "/var/lib/zuul/builds/7573dac089bc49b5b355588f880c7711/work/logs"
                            },
                            "changed": false
                        }
                    },
                    "role": {
                        "id": "0242ac1f-6606-530f-7072-000000000006",
                        "name": "fetch-output",
                        "path": "/var/lib/zuul/builds/7573dac089bc49b5b355588f880c7711/trusted/project_1/github.com/osism/openinfra-zuul-jobs/roles/fetch-output"
                    },
                    "task": {
                        "duration": {
                            "end": "2026-02-24T01:42:05.385002Z",
                            "start": "2026-02-24T01:42:05.350808Z"
                        },
                        "id": "0242ac1f-6606-530f-7072-000000000009",
                        "name": "Set log path for single node"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "file",
                            "changed": true,
                            "msg": "All items completed",
                            "results": [
                                {
                                    "ansible_loop_var": "zj_output_dir",
                                    "changed": false,
                                    "diff": {
                                        "after": {
                                            "path": "/var/lib/zuul/builds/7573dac089bc49b5b355588f880c7711/work/logs"
                                        },
                                        "before": {
                                            "path": "/var/lib/zuul/builds/7573dac089bc49b5b355588f880c7711/work/logs"
                                        }
                                    },
                                    "failed": false,
                                    "gid": 0,
                                    "group": "root",
                                    "invocation": {
                                        "module_args": {
                                            "_diff_peek": null,
                                            "_original_basename": null,
                                            "access_time": null,
                                            "access_time_format": "%Y%m%d%H%M.%S",
                                            "attributes": null,
                                            "follow": true,
                                            "force": false,
                                            "group": null,
                                            "mode": 493,
                                            "modification_time": null,
                                            "modification_time_format": "%Y%m%d%H%M.%S",
                                            "owner": null,
                                            "path": "/var/lib/zuul/builds/7573dac089bc49b5b355588f880c7711/work/logs",
                                            "recurse": false,
                                            "selevel": null,
                                            "serole": null,
                                            "setype": null,
                                            "seuser": null,
                                            "src": null,
                                            "state": "directory",
                                            "unsafe_writes": false
                                        }
                                    },
                                    "mode": "0755",
                                    "owner": "root",
                                    "path": "/var/lib/zuul/builds/7573dac089bc49b5b355588f880c7711/work/logs",
                                    "size": 4096,
                                    "state": "directory",
                                    "uid": 0,
                                    "zj_output_dir": "/var/lib/zuul/builds/7573dac089bc49b5b355588f880c7711/work/logs"
                                },
                                {
                                    "ansible_loop_var": "zj_output_dir",
                                    "changed": true,
                                    "diff": {
                                        "after": {
                                            "path": "/var/lib/zuul/builds/7573dac089bc49b5b355588f880c7711/work/artifacts",
                                            "state": "directory"
                                        },
                                        "before": {
                                            "path": "/var/lib/zuul/builds/7573dac089bc49b5b355588f880c7711/work/artifacts",
                                            "state": "absent"
                                        }
                                    },
                                    "failed": false,
                                    "gid": 0,
                                    "group": "root",
                                    "invocation": {
                                        "module_args": {
                                            "_diff_peek": null,
                                            "_original_basename": null,
                                            "access_time": null,
                                            "access_time_format": "%Y%m%d%H%M.%S",
                                            "attributes": null,
                                            "follow": true,
                                            "force": false,
                                            "group": null,
                                            "mode": 493,
                                            "modification_time": null,
                                            "modification_time_format": "%Y%m%d%H%M.%S",
                                            "owner": null,
                                            "path": "/var/lib/zuul/builds/7573dac089bc49b5b355588f880c7711/work/artifacts",
                                            "recurse": false,
                                            "selevel": null,
                                            "serole": null,
                                            "setype": null,
                                            "seuser": null,
                                            "src": null,
                                            "state": "directory",
                                            "unsafe_writes": false
                                        }
                                    },
                                    "mode": "0755",
                                    "owner": "root",
                                    "path": "/var/lib/zuul/builds/7573dac089bc49b5b355588f880c7711/work/artifacts",
                                    "size": 4096,
                                    "state": "directory",
                                    "uid": 0,
                                    "zj_output_dir": "/var/lib/zuul/builds/7573dac089bc49b5b355588f880c7711/work/artifacts"
                                },
                                {
                                    "ansible_loop_var": "zj_output_dir",
                                    "changed": true,
                                    "diff": {
                                        "after": {
                                            "path": "/var/lib/zuul/builds/7573dac089bc49b5b355588f880c7711/work/docs",
                                            "state": "directory"
                                        },
                                        "before": {
                                            "path": "/var/lib/zuul/builds/7573dac089bc49b5b355588f880c7711/work/docs",
                                            "state": "absent"
                                        }
                                    },
                                    "failed": false,
                                    "gid": 0,
                                    "group": "root",
                                    "invocation": {
                                        "module_args": {
                                            "_diff_peek": null,
                                            "_original_basename": null,
                                            "access_time": null,
                                            "access_time_format": "%Y%m%d%H%M.%S",
                                            "attributes": null,
                                            "follow": true,
                                            "force": false,
                                            "group": null,
                                            "mode": 493,
                                            "modification_time": null,
                                            "modification_time_format": "%Y%m%d%H%M.%S",
                                            "owner": null,
                                            "path": "/var/lib/zuul/builds/7573dac089bc49b5b355588f880c7711/work/docs",
                                            "recurse": false,
                                            "selevel": null,
                                            "serole": null,
                                            "setype": null,
                                            "seuser": null,
                                            "src": null,
                                            "state": "directory",
                                            "unsafe_writes": false
                                        }
                                    },
                                    "mode": "0755",
                                    "owner": "root",
                                    "path": "/var/lib/zuul/builds/7573dac089bc49b5b355588f880c7711/work/docs",
                                    "size": 4096,
                                    "state": "directory",
                                    "uid": 0,
                                    "zj_output_dir": "/var/lib/zuul/builds/7573dac089bc49b5b355588f880c7711/work/docs"
                                }
                            ]
                        }
                    },
                    "role": {
                        "id": "0242ac1f-6606-530f-7072-000000000006",
                        "name": "fetch-output",
                        "path": "/var/lib/zuul/builds/7573dac089bc49b5b355588f880c7711/trusted/project_1/github.com/osism/openinfra-zuul-jobs/roles/fetch-output"
                    },
                    "task": {
                        "duration": {
                            "end": "2026-02-24T01:42:06.427151Z",
                            "start": "2026-02-24T01:42:05.391891Z"
                        },
                        "id": "0242ac1f-6606-530f-7072-00000000000a",
                        "name": "Ensure local output dirs"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "synchronize",
                            "changed": true,
                            "msg": "All items completed",
                            "results": [
                                {
                                    "ansible_loop_var": "zj_output",
                                    "changed": true,
                                    "cmd": "/usr/bin/rsync --delay-updates -F --compress --archive --no-owner --no-group --rsh='/usr/bin/ssh -S none -o Port=22' --out-format='<<CHANGED>>%i %n%L' zuul-testbed02@81.163.193.140:/home/zuul-testbed02/zuul-output/logs/ /var/lib/zuul/builds/7573dac089bc49b5b355588f880c7711/work/logs/",
                                    "failed": false,
                                    "invocation": {
                                        "module_args": {
                                            "_local_rsync_password": null,
                                            "_local_rsync_path": "rsync",
                                            "_ssh_args": null,
                                            "_substitute_controller": false,
                                            "archive": true,
                                            "checksum": false,
                                            "compress": true,
                                            "copy_links": false,
                                            "delay_updates": true,
                                            "delete": false,
                                            "dest": "/var/lib/zuul/builds/7573dac089bc49b5b355588f880c7711/work/logs/",
                                            "dest_port": 22,
                                            "dirs": false,
                                            "existing_only": false,
                                            "group": false,
                                            "link_dest": null,
                                            "links": null,
                                            "mode": "pull",
                                            "owner": false,
                                            "partial": false,
                                            "perms": null,
                                            "private_key": null,
                                            "recursive": null,
                                            "rsync_opts": [],
                                            "rsync_path": null,
                                            "rsync_timeout": 0,
                                            "set_remote_user": true,
                                            "src": "zuul-testbed02@81.163.193.140:/home/zuul-testbed02/zuul-output/logs/",
                                            "ssh_connection_multiplexing": false,
                                            "times": null,
                                            "use_ssh_args": false,
                                            "verify_host": true
                                        }
                                    },
                                    "msg": ".d..t...... ./\n",
                                    "rc": 0,
                                    "stdout_lines": [
                                        ".d..t...... ./"
                                    ],
                                    "zj_output": {
                                        "dest": "/var/lib/zuul/builds/7573dac089bc49b5b355588f880c7711/work/logs",
                                        "src": "logs"
                                    }
                                },
                                {
                                    "ansible_loop_var": "zj_output",
                                    "changed": true,
                                    "cmd": "/usr/bin/rsync --delay-updates -F --compress --archive --no-owner --no-group --rsh='/usr/bin/ssh -S none -o Port=22' --out-format='<<CHANGED>>%i %n%L' zuul-testbed02@81.163.193.140:/home/zuul-testbed02/zuul-output/artifacts/ /var/lib/zuul/builds/7573dac089bc49b5b355588f880c7711/work/artifacts/",
                                    "failed": false,
                                    "invocation": {
                                        "module_args": {
                                            "_local_rsync_password": null,
                                            "_local_rsync_path": "rsync",
                                            "_ssh_args": null,
                                            "_substitute_controller": false,
                                            "archive": true,
                                            "checksum": false,
                                            "compress": true,
                                            "copy_links": false,
                                            "delay_updates": true,
                                            "delete": false,
                                            "dest": "/var/lib/zuul/builds/7573dac089bc49b5b355588f880c7711/work/artifacts/",
                                            "dest_port": 22,
                                            "dirs": false,
                                            "existing_only": false,
                                            "group": false,
                                            "link_dest": null,
                                            "links": null,
                                            "mode": "pull",
                                            "owner": false,
                                            "partial": false,
                                            "perms": null,
                                            "private_key": null,
                                            "recursive": null,
                                            "rsync_opts": [],
                                            "rsync_path": null,
                                            "rsync_timeout": 0,
                                            "set_remote_user": true,
                                            "src": "zuul-testbed02@81.163.193.140:/home/zuul-testbed02/zuul-output/artifacts/",
                                            "ssh_connection_multiplexing": false,
                                            "times": null,
                                            "use_ssh_args": false,
                                            "verify_host": true
                                        }
                                    },
                                    "msg": ".d..t...... ./\n",
                                    "rc": 0,
                                    "stdout_lines": [
                                        ".d..t...... ./"
                                    ],
                                    "zj_output": {
                                        "dest": "/var/lib/zuul/builds/7573dac089bc49b5b355588f880c7711/work/artifacts",
                                        "src": "artifacts"
                                    }
                                },
                                {
                                    "ansible_loop_var": "zj_output",
                                    "changed": true,
                                    "cmd": "/usr/bin/rsync --delay-updates -F --compress --archive --no-owner --no-group --rsh='/usr/bin/ssh -S none -o Port=22' --out-format='<<CHANGED>>%i %n%L' zuul-testbed02@81.163.193.140:/home/zuul-testbed02/zuul-output/docs/ /var/lib/zuul/builds/7573dac089bc49b5b355588f880c7711/work/docs/",
                                    "failed": false,
                                    "invocation": {
                                        "module_args": {
                                            "_local_rsync_password": null,
                                            "_local_rsync_path": "rsync",
                                            "_ssh_args": null,
                                            "_substitute_controller": false,
                                            "archive": true,
                                            "checksum": false,
                                            "compress": true,
                                            "copy_links": false,
                                            "delay_updates": true,
                                            "delete": false,
                                            "dest": "/var/lib/zuul/builds/7573dac089bc49b5b355588f880c7711/work/docs/",
                                            "dest_port": 22,
                                            "dirs": false,
                                            "existing_only": false,
                                            "group": false,
                                            "link_dest": null,
                                            "links": null,
                                            "mode": "pull",
                                            "owner": false,
                                            "partial": false,
                                            "perms": null,
                                            "private_key": null,
                                            "recursive": null,
                                            "rsync_opts": [],
                                            "rsync_path": null,
                                            "rsync_timeout": 0,
                                            "set_remote_user": true,
                                            "src": "zuul-testbed02@81.163.193.140:/home/zuul-testbed02/zuul-output/docs/",
                                            "ssh_connection_multiplexing": false,
                                            "times": null,
                                            "use_ssh_args": false,
                                            "verify_host": true
                                        }
                                    },
                                    "msg": ".d..t...... ./\n",
                                    "rc": 0,
                                    "stdout_lines": [
                                        ".d..t...... ./"
                                    ],
                                    "zj_output": {
                                        "dest": "/var/lib/zuul/builds/7573dac089bc49b5b355588f880c7711/work/docs",
                                        "src": "docs"
                                    }
                                }
                            ]
                        }
                    },
                    "role": {
                        "id": "0242ac1f-6606-530f-7072-000000000006",
                        "name": "fetch-output",
                        "path": "/var/lib/zuul/builds/7573dac089bc49b5b355588f880c7711/trusted/project_1/github.com/osism/openinfra-zuul-jobs/roles/fetch-output"
                    },
                    "task": {
                        "duration": {
                            "end": "2026-02-24T01:42:08.864776Z",
                            "start": "2026-02-24T01:42:06.443260Z"
                        },
                        "id": "0242ac1f-6606-530f-7072-00000000000c",
                        "name": "Collect logs, artifacts and docs"
                    }
                },
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "shell",
                            "changed": false,
                            "msg": "All items skipped",
                            "results": [
                                {
                                    "ansible_loop_var": "zj_item",
                                    "changed": false,
                                    "false_condition": "zuul.change is defined",
                                    "skip_reason": "Conditional result was False",
                                    "skipped": true,
                                    "zj_item": "artifacts"
                                },
                                {
                                    "ansible_loop_var": "zj_item",
                                    "changed": false,
                                    "false_condition": "zuul.change is defined",
                                    "skip_reason": "Conditional result was False",
                                    "skipped": true,
                                    "zj_item": "docs"
                                }
                            ],
                            "skipped": true
                        }
                    },
                    "role": {
                        "id": "0242ac1f-6606-530f-7072-00000000000f",
                        "name": "merge-output-to-logs",
                        "path": "/var/lib/zuul/builds/7573dac089bc49b5b355588f880c7711/trusted/project_1/github.com/osism/openinfra-zuul-jobs/roles/merge-output-to-logs"
                    },
                    "task": {
                        "duration": {
                            "end": "2026-02-24T01:42:08.917898Z",
                            "start": "2026-02-24T01:42:08.883029Z"
                        },
                        "id": "0242ac1f-6606-530f-7072-000000000011",
                        "name": "Move artifacts and docs to logs dir"
                    }
                }
            ]
        }
    ],
    "stats": {
        "orchestrator": {
            "changed": 2,
            "failures": 0,
            "ignored": 0,
            "ok": 3,
            "rescued": 0,
            "skipped": 2,
            "unreachable": 0
        }
    },
    "trusted": true
},
{
    "branch": "main",
    "index": "3",
    "phase": "post",
    "playbook": "github.com/osism/zuul-config/playbooks/base/post.yaml",
    "plays": [
        {
            "play": {
                "duration": {
                    "end": "2026-02-24T01:42:10.917728Z",
                    "start": "2026-02-24T01:42:09.816193Z"
                },
                "id": "0242ac1f-6606-0da0-f794-000000000002",
                "name": "Base post"
            },
            "tasks": [
                {
                    "hosts": {
                        "orchestrator": {
                            "action": "authorized_key",
                            "changed": true,
                            "comment": null,
                            "exclusive": false,
                            "follow": false,
                            "invocation": {
                                "module_args": {
                                    "changed": true,
                                    "comment": null,
                                    "exclusive": false,
                                    "follow": false,
                                    "key": "ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABgQDJ+0S2SO1gCkr7fCftNF31nAe7WVxUgGUxJF48Xyi6ePW6/DHF3iqRVS/363OM7JwGtrYwwa6lUw9TvKCfs6kERvEV/scmx6xhtS6HR3g2+dcUCzdp4MhDCTkEn/MRojg/TkmQ1P1rrJ/bwYBKuS6hf0TD7JdvpVuHIO1KA4v4V3/jLLeUw66yYIs+xouIj54PwovhrVz3OGSTsytjZVfmgNtu8VQ3bRxxThZ+hkEJia3JJ/TiwbXkL0k7PVDEZFvTnFHpjN8T1a3Kkv2sJ9sDQpbeeeTZFmxon++pifXN8/AawZnaIXWdN5lEP5BmHMzSTM2KVor591v9kFRpNeBuQ/Gr3jE5ZIggAdzRINLXV1dT3dTMRe1TKtQm6P6XXnCjXZV/OAZmB145nAfNREpRJWx5UQRXRks9S3kS07N5MRWP3XhMVBB6gA6m76tbQq+5cbF9zSlZ7ABi6Wy3/LIUnHt8TYjWgKY23KPHyqMqy+nxd4mHcKScBp29NZClNtU= zuul-build-sshkey",
                                    "key_options": null,
                                    "keyfile": "/home/zuul-testbed02/.ssh/authorized_keys",
                                    "manage_dir": true,
                                    "path": null,
                                    "state": "absent",
                                    "user": "zuul-testbed02",
                                    "validate_certs": true
                                }
                            },
                            "key": "ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABgQDJ+0S2SO1gCkr7fCftNF31nAe7WVxUgGUxJF48Xyi6ePW6/DHF3iqRVS/363OM7JwGtrYwwa6lUw9TvKCfs6kERvEV/scmx6xhtS6HR3g2+dcUCzdp4MhDCTkEn/MRojg/TkmQ1P1rrJ/bwYBKuS6hf0TD7JdvpVuHIO1KA4v4V3/jLLeUw66yYIs+xouIj54PwovhrVz3OGSTsytjZVfmgNtu8VQ3bRxxThZ+hkEJia3JJ/TiwbXkL0k7PVDEZFvTnFHpjN8T1a3Kkv2sJ9sDQpbeeeTZFmxon++pifXN8/AawZnaIXWdN5lEP5BmHMzSTM2KVor591v9kFRpNeBuQ/Gr3jE5ZIggAdzRINLXV1dT3dTMRe1TKtQm6P6XXnCjXZV/OAZmB145nAfNREpRJWx5UQRXRks9S3kS07N5MRWP3XhMVBB6gA6m76tbQq+5cbF9zSlZ7ABi6Wy3/LIUnHt8TYjWgKY23KPHyqMqy+nxd4mHcKScBp29NZClNtU= zuul-build-sshkey",
                            "key_options": null,
                            "keyfile": "/home/zuul-testbed02/.ssh/authorized_keys",
                            "manage_dir": true,
                            "path": null,
                            "state": "absent",
                            "user": "zuul-testbed02",
                            "validate_certs": true
                        }
                    },
                    "role": {
                        "id": "0242ac1f-6606-0da0-f794-000000000005",
                        "name": "remove-build-sshkey",
                        "path": "/var/lib/zuul/builds/7573dac089bc49b5b355588f880c7711/trusted/project_1/github.com/osism/openinfra-zuul-jobs/roles/remove-build-sshkey"
                    },
                    "task": {
                        "duration": {
                            "end": "2026-02-24T01:42:10.917728Z",
                            "start": "2026-02-24T01:42:09.831539Z"
                        },
                        "id": "0242ac1f-6606-0da0-f794-000000000007",
                        "name": "Remove the build SSH key from all nodes"
                    }
                }
            ]
        }
    ],
    "stats": {
        "orchestrator": {
            "changed": 1,
            "failures": 0,
            "ignored": 0,
            "ok": 1,
            "rescued": 0,
            "skipped": 0,
            "unreachable": 0
        }
    },
    "trusted": true
}
]
